{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1665, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "ce_loss": 0.0001950513105839491, "cls_loss": 0.04638671875, "epoch": 0, "mask_bce_loss": 0.827459990978241, "mask_dice_loss": 0.07344185560941696, "mask_loss": 0.9009018540382385, "step": 0 }, { "epoch": 0.0006006006006006006, "grad_norm": 38.726375579833984, "learning_rate": 0.0, "loss": 0.8971301913261414, "step": 1 }, { "ce_loss": 3.771077172132209e-05, "cls_loss": 0.09423828125, "epoch": 0.0006006006006006006, "mask_bce_loss": 0.6213144659996033, "mask_dice_loss": 0.06328874081373215, "mask_loss": 0.684603214263916, "step": 1 }, { "epoch": 0.0012012012012012011, "grad_norm": 39.08122253417969, "learning_rate": 4e-08, "loss": 1.0324949026107788, "step": 2 }, { "ce_loss": 2.4739900254644454e-05, "cls_loss": 0.04736328125, "epoch": 0.0012012012012012011, "mask_bce_loss": 1.0017527341842651, "mask_dice_loss": 0.0672466829419136, "mask_loss": 1.0689994096755981, "step": 2 }, { "epoch": 0.0018018018018018018, "grad_norm": 19.775165557861328, "learning_rate": 8e-08, "loss": 0.813344419002533, "step": 3 }, { "ce_loss": 4.5407319703372195e-05, "cls_loss": 0.0302734375, "epoch": 0.0018018018018018018, "mask_bce_loss": 0.22313737869262695, "mask_dice_loss": 0.02978905476629734, "mask_loss": 0.25292643904685974, "step": 3 }, { "epoch": 0.0024024024024024023, "grad_norm": 31.632966995239258, "learning_rate": 1.2e-07, "loss": 1.1208300590515137, "step": 4 }, { "ce_loss": 3.150554039166309e-05, "cls_loss": 0.0238037109375, "epoch": 0.0024024024024024023, "mask_bce_loss": 0.39603158831596375, "mask_dice_loss": 0.03820445016026497, "mask_loss": 0.4342360496520996, "step": 4 }, { "epoch": 0.003003003003003003, "grad_norm": 25.61577606201172, "learning_rate": 1.6e-07, "loss": 0.83367520570755, "step": 5 }, { "ce_loss": 0.00027966639026999474, "cls_loss": 0.03857421875, "epoch": 0.003003003003003003, "mask_bce_loss": 0.4730179011821747, "mask_dice_loss": 0.05590325593948364, "mask_loss": 0.5289211273193359, "step": 5 }, { "epoch": 0.0036036036036036037, "grad_norm": 26.30783462524414, "learning_rate": 2e-07, "loss": 0.8880058526992798, "step": 6 }, { "ce_loss": 5.2694242185680196e-05, "cls_loss": 0.049072265625, "epoch": 0.0036036036036036037, "mask_bce_loss": 0.4146251380443573, "mask_dice_loss": 0.0507015585899353, "mask_loss": 0.4653266966342926, "step": 6 }, { "epoch": 0.004204204204204204, "grad_norm": 24.11886215209961, "learning_rate": 2.4e-07, "loss": 0.9241217374801636, "step": 7 }, { "ce_loss": 0.00015790296311024576, "cls_loss": 0.06298828125, "epoch": 0.004204204204204204, "mask_bce_loss": 0.7550400495529175, "mask_dice_loss": 0.09193099290132523, "mask_loss": 0.8469710350036621, "step": 7 }, { "epoch": 0.004804804804804805, "grad_norm": 99.65950775146484, "learning_rate": 2.8e-07, "loss": 0.9876590967178345, "step": 8 }, { "ce_loss": 0.00011772006837418303, "cls_loss": 0.059814453125, "epoch": 0.004804804804804805, "mask_bce_loss": 1.1678857803344727, "mask_dice_loss": 0.1364162266254425, "mask_loss": 1.3043019771575928, "step": 8 }, { "epoch": 0.005405405405405406, "grad_norm": 34.30335235595703, "learning_rate": 3.2e-07, "loss": 1.0358705520629883, "step": 9 }, { "ce_loss": 6.087202928029001e-05, "cls_loss": 0.046875, "epoch": 0.005405405405405406, "mask_bce_loss": 0.9002136588096619, "mask_dice_loss": 0.05483070760965347, "mask_loss": 0.9550443887710571, "step": 9 }, { "epoch": 0.006006006006006006, "grad_norm": 40.01549530029297, "learning_rate": 3.6e-07, "loss": 0.8243277668952942, "step": 10 }, { "ce_loss": 0.0017144251614809036, "cls_loss": 0.031982421875, "epoch": 0.006006006006006006, "mask_bce_loss": 0.4049767553806305, "mask_dice_loss": 0.025017166510224342, "mask_loss": 0.4299939274787903, "step": 10 }, { "epoch": 0.006606606606606606, "grad_norm": 17.669097900390625, "learning_rate": 4e-07, "loss": 1.0076779127120972, "step": 11 }, { "ce_loss": 0.0008432778413407505, "cls_loss": 0.037841796875, "epoch": 0.006606606606606606, "mask_bce_loss": 0.4568677544593811, "mask_dice_loss": 0.04306888207793236, "mask_loss": 0.49993664026260376, "step": 11 }, { "epoch": 0.007207207207207207, "grad_norm": 18.534191131591797, "learning_rate": 4.3999999999999997e-07, "loss": 0.8444398641586304, "step": 12 }, { "ce_loss": 3.8197289541130885e-05, "cls_loss": 0.059814453125, "epoch": 0.007207207207207207, "mask_bce_loss": 0.5635537505149841, "mask_dice_loss": 0.09957046806812286, "mask_loss": 0.6631242036819458, "step": 12 }, { "epoch": 0.007807807807807808, "grad_norm": 20.84868621826172, "learning_rate": 4.8e-07, "loss": 0.8819471597671509, "step": 13 }, { "ce_loss": 3.022121381945908e-05, "cls_loss": 0.05224609375, "epoch": 0.007807807807807808, "mask_bce_loss": 1.1392101049423218, "mask_dice_loss": 0.05204949527978897, "mask_loss": 1.1912596225738525, "step": 13 }, { "epoch": 0.008408408408408409, "grad_norm": 27.0803279876709, "learning_rate": 5.2e-07, "loss": 1.0044770240783691, "step": 14 }, { "ce_loss": 5.389935176935978e-05, "cls_loss": 0.0546875, "epoch": 0.008408408408408409, "mask_bce_loss": 0.8658686876296997, "mask_dice_loss": 0.11360470205545425, "mask_loss": 0.9794734120368958, "step": 14 }, { "epoch": 0.009009009009009009, "grad_norm": 23.01841926574707, "learning_rate": 5.6e-07, "loss": 1.0702931880950928, "step": 15 }, { "ce_loss": 8.010596502572298e-05, "cls_loss": 0.039794921875, "epoch": 0.009009009009009009, "mask_bce_loss": 0.541164755821228, "mask_dice_loss": 0.21942806243896484, "mask_loss": 0.7605928182601929, "step": 15 }, { "epoch": 0.00960960960960961, "grad_norm": 35.92723083496094, "learning_rate": 6e-07, "loss": 0.8679159879684448, "step": 16 }, { "ce_loss": 0.0638352632522583, "cls_loss": 0.05224609375, "epoch": 0.00960960960960961, "mask_bce_loss": 0.993442177772522, "mask_dice_loss": 0.09628858417272568, "mask_loss": 1.0897307395935059, "step": 16 }, { "epoch": 0.01021021021021021, "grad_norm": 44.996009826660156, "learning_rate": 6.4e-07, "loss": 1.0068349838256836, "step": 17 }, { "ce_loss": 4.294735845178366e-05, "cls_loss": 0.025390625, "epoch": 0.01021021021021021, "mask_bce_loss": 0.12397237867116928, "mask_dice_loss": 0.017591726034879684, "mask_loss": 0.14156410098075867, "step": 17 }, { "epoch": 0.010810810810810811, "grad_norm": 24.620159149169922, "learning_rate": 6.800000000000001e-07, "loss": 0.9991017580032349, "step": 18 }, { "ce_loss": 0.1729658991098404, "cls_loss": 0.05078125, "epoch": 0.010810810810810811, "mask_bce_loss": 0.8894389271736145, "mask_dice_loss": 0.0727667585015297, "mask_loss": 0.962205708026886, "step": 18 }, { "epoch": 0.011411411411411412, "grad_norm": 18.477928161621094, "learning_rate": 7.2e-07, "loss": 0.7885504961013794, "step": 19 }, { "ce_loss": 0.0006899026921018958, "cls_loss": 0.0625, "epoch": 0.011411411411411412, "mask_bce_loss": 0.9015578627586365, "mask_dice_loss": 0.09502631425857544, "mask_loss": 0.9965841770172119, "step": 19 }, { "epoch": 0.012012012012012012, "grad_norm": 28.108856201171875, "learning_rate": 7.599999999999999e-07, "loss": 0.8327940702438354, "step": 20 }, { "ce_loss": 3.513430419843644e-05, "cls_loss": 0.046630859375, "epoch": 0.012012012012012012, "mask_bce_loss": 0.8735513687133789, "mask_dice_loss": 0.07663969695568085, "mask_loss": 0.950191080570221, "step": 20 }, { "epoch": 0.012612612612612612, "grad_norm": 31.9722957611084, "learning_rate": 8e-07, "loss": 1.0417447090148926, "step": 21 }, { "ce_loss": 2.5251891202060506e-05, "cls_loss": 0.039794921875, "epoch": 0.012612612612612612, "mask_bce_loss": 0.5481845736503601, "mask_dice_loss": 0.054270267486572266, "mask_loss": 0.6024548411369324, "step": 21 }, { "epoch": 0.013213213213213212, "grad_norm": 42.219417572021484, "learning_rate": 8.399999999999999e-07, "loss": 1.1107808351516724, "step": 22 }, { "ce_loss": 5.421639434644021e-05, "cls_loss": 0.03466796875, "epoch": 0.013213213213213212, "mask_bce_loss": 0.3797394037246704, "mask_dice_loss": 0.044442735612392426, "mask_loss": 0.42418214678764343, "step": 22 }, { "epoch": 0.013813813813813814, "grad_norm": 23.584108352661133, "learning_rate": 8.799999999999999e-07, "loss": 1.0571240186691284, "step": 23 }, { "ce_loss": 6.031292650732212e-05, "cls_loss": 0.04931640625, "epoch": 0.013813813813813814, "mask_bce_loss": 0.9640222787857056, "mask_dice_loss": 0.11298079788684845, "mask_loss": 1.0770031213760376, "step": 23 }, { "epoch": 0.014414414414414415, "grad_norm": 22.04401206970215, "learning_rate": 9.2e-07, "loss": 0.9899495840072632, "step": 24 }, { "ce_loss": 0.0002670586109161377, "cls_loss": 0.032958984375, "epoch": 0.014414414414414415, "mask_bce_loss": 0.30124953389167786, "mask_dice_loss": 0.029740897938609123, "mask_loss": 0.33099043369293213, "step": 24 }, { "epoch": 0.015015015015015015, "grad_norm": 16.3223934173584, "learning_rate": 9.6e-07, "loss": 0.7710822224617004, "step": 25 }, { "ce_loss": 4.04021811846178e-05, "cls_loss": 0.049560546875, "epoch": 0.015015015015015015, "mask_bce_loss": 0.7024247050285339, "mask_dice_loss": 0.045827534049749374, "mask_loss": 0.7482522130012512, "step": 25 }, { "epoch": 0.015615615615615615, "grad_norm": 30.721508026123047, "learning_rate": 1e-06, "loss": 0.8861181735992432, "step": 26 }, { "ce_loss": 3.2477859349455684e-05, "cls_loss": 0.0634765625, "epoch": 0.015615615615615615, "mask_bce_loss": 1.102895975112915, "mask_dice_loss": 0.1229197159409523, "mask_loss": 1.2258156538009644, "step": 26 }, { "epoch": 0.016216216216216217, "grad_norm": 18.2114200592041, "learning_rate": 1.04e-06, "loss": 0.8854597806930542, "step": 27 }, { "ce_loss": 0.0003752975317183882, "cls_loss": 0.0634765625, "epoch": 0.016216216216216217, "mask_bce_loss": 0.2360563725233078, "mask_dice_loss": 0.16914057731628418, "mask_loss": 0.4051969647407532, "step": 27 }, { "epoch": 0.016816816816816817, "grad_norm": 45.132591247558594, "learning_rate": 1.08e-06, "loss": 0.9073913097381592, "step": 28 }, { "ce_loss": 2.974842573166825e-05, "cls_loss": 0.04345703125, "epoch": 0.016816816816816817, "mask_bce_loss": 0.5517237186431885, "mask_dice_loss": 0.05340861901640892, "mask_loss": 0.6051323413848877, "step": 28 }, { "epoch": 0.017417417417417418, "grad_norm": 33.347007751464844, "learning_rate": 1.12e-06, "loss": 1.0176351070404053, "step": 29 }, { "ce_loss": 5.639245864585973e-05, "cls_loss": 0.03955078125, "epoch": 0.017417417417417418, "mask_bce_loss": 0.39388179779052734, "mask_dice_loss": 0.0345223993062973, "mask_loss": 0.42840421199798584, "step": 29 }, { "epoch": 0.018018018018018018, "grad_norm": 25.136804580688477, "learning_rate": 1.16e-06, "loss": 0.8998695015907288, "step": 30 }, { "ce_loss": 0.0005346655962057412, "cls_loss": 0.04248046875, "epoch": 0.018018018018018018, "mask_bce_loss": 0.40979573130607605, "mask_dice_loss": 0.05846478417515755, "mask_loss": 0.4682605266571045, "step": 30 }, { "epoch": 0.018618618618618618, "grad_norm": 21.5479736328125, "learning_rate": 1.2e-06, "loss": 0.9575955867767334, "step": 31 }, { "ce_loss": 2.600159314170014e-05, "cls_loss": 0.05322265625, "epoch": 0.018618618618618618, "mask_bce_loss": 0.6451166868209839, "mask_dice_loss": 0.08823170512914658, "mask_loss": 0.7333483695983887, "step": 31 }, { "epoch": 0.01921921921921922, "grad_norm": 12.14796257019043, "learning_rate": 1.24e-06, "loss": 0.8752981424331665, "step": 32 }, { "ce_loss": 5.9037498431280255e-05, "cls_loss": 0.05615234375, "epoch": 0.01921921921921922, "mask_bce_loss": 0.7701612114906311, "mask_dice_loss": 0.1273989975452423, "mask_loss": 0.8975602388381958, "step": 32 }, { "epoch": 0.01981981981981982, "grad_norm": 20.392297744750977, "learning_rate": 1.28e-06, "loss": 0.8792307376861572, "step": 33 }, { "ce_loss": 4.0730363252805546e-05, "cls_loss": 0.055908203125, "epoch": 0.01981981981981982, "mask_bce_loss": 0.4866255819797516, "mask_dice_loss": 0.07490810751914978, "mask_loss": 0.5615336894989014, "step": 33 }, { "epoch": 0.02042042042042042, "grad_norm": 24.64166259765625, "learning_rate": 1.32e-06, "loss": 0.8872016072273254, "step": 34 }, { "ce_loss": 9.299322846345603e-05, "cls_loss": 0.027587890625, "epoch": 0.02042042042042042, "mask_bce_loss": 0.378669798374176, "mask_dice_loss": 0.020052386447787285, "mask_loss": 0.39872217178344727, "step": 34 }, { "epoch": 0.021021021021021023, "grad_norm": 26.910572052001953, "learning_rate": 1.3600000000000001e-06, "loss": 0.7661287188529968, "step": 35 }, { "ce_loss": 6.58505450701341e-05, "cls_loss": 0.04638671875, "epoch": 0.021021021021021023, "mask_bce_loss": 0.7942559123039246, "mask_dice_loss": 0.10919521003961563, "mask_loss": 0.903451144695282, "step": 35 }, { "epoch": 0.021621621621621623, "grad_norm": 20.409210205078125, "learning_rate": 1.4e-06, "loss": 0.8798038959503174, "step": 36 }, { "ce_loss": 5.4910939070396125e-05, "cls_loss": 0.04150390625, "epoch": 0.021621621621621623, "mask_bce_loss": 0.661954939365387, "mask_dice_loss": 0.03997492045164108, "mask_loss": 0.7019298672676086, "step": 36 }, { "epoch": 0.022222222222222223, "grad_norm": 28.3300838470459, "learning_rate": 1.44e-06, "loss": 0.8424479961395264, "step": 37 }, { "ce_loss": 0.000174602959305048, "cls_loss": 0.05810546875, "epoch": 0.022222222222222223, "mask_bce_loss": 0.6921066641807556, "mask_dice_loss": 0.10609755665063858, "mask_loss": 0.798204243183136, "step": 37 }, { "epoch": 0.022822822822822823, "grad_norm": 21.242090225219727, "learning_rate": 1.48e-06, "loss": 0.9962613582611084, "step": 38 }, { "ce_loss": 0.0007712126825936139, "cls_loss": 0.03955078125, "epoch": 0.022822822822822823, "mask_bce_loss": 0.4524882733821869, "mask_dice_loss": 0.05076346918940544, "mask_loss": 0.5032517313957214, "step": 38 }, { "epoch": 0.023423423423423424, "grad_norm": 17.265104293823242, "learning_rate": 1.5199999999999998e-06, "loss": 0.7724307179450989, "step": 39 }, { "ce_loss": 4.7900604840833694e-05, "cls_loss": 0.04638671875, "epoch": 0.023423423423423424, "mask_bce_loss": 0.6889616847038269, "mask_dice_loss": 0.07600312680006027, "mask_loss": 0.7649648189544678, "step": 39 }, { "epoch": 0.024024024024024024, "grad_norm": 31.949127197265625, "learning_rate": 1.5599999999999999e-06, "loss": 1.0516979694366455, "step": 40 }, { "ce_loss": 6.268602010095492e-05, "cls_loss": 0.06494140625, "epoch": 0.024024024024024024, "mask_bce_loss": 1.5325088500976562, "mask_dice_loss": 0.1753079742193222, "mask_loss": 1.7078168392181396, "step": 40 }, { "epoch": 0.024624624624624624, "grad_norm": 24.759033203125, "learning_rate": 1.6e-06, "loss": 0.9196768999099731, "step": 41 }, { "ce_loss": 7.853478018660098e-05, "cls_loss": 0.0654296875, "epoch": 0.024624624624624624, "mask_bce_loss": 1.1261519193649292, "mask_dice_loss": 0.15838444232940674, "mask_loss": 1.284536361694336, "step": 41 }, { "epoch": 0.025225225225225224, "grad_norm": 27.044315338134766, "learning_rate": 1.6399999999999998e-06, "loss": 0.8531088829040527, "step": 42 }, { "ce_loss": 3.6666573578258976e-05, "cls_loss": 0.031982421875, "epoch": 0.025225225225225224, "mask_bce_loss": 0.29672226309776306, "mask_dice_loss": 0.019302848726511, "mask_loss": 0.31602510809898376, "step": 42 }, { "epoch": 0.025825825825825825, "grad_norm": 58.472312927246094, "learning_rate": 1.6799999999999998e-06, "loss": 1.105468511581421, "step": 43 }, { "ce_loss": 4.6303553972393274e-05, "cls_loss": 0.0439453125, "epoch": 0.025825825825825825, "mask_bce_loss": 0.5885554552078247, "mask_dice_loss": 0.049448441714048386, "mask_loss": 0.6380038857460022, "step": 43 }, { "epoch": 0.026426426426426425, "grad_norm": 16.537572860717773, "learning_rate": 1.7199999999999998e-06, "loss": 0.797466516494751, "step": 44 }, { "ce_loss": 9.399391274200752e-05, "cls_loss": 0.056640625, "epoch": 0.026426426426426425, "mask_bce_loss": 0.6371244192123413, "mask_dice_loss": 0.11429309099912643, "mask_loss": 0.7514175176620483, "step": 44 }, { "epoch": 0.02702702702702703, "grad_norm": 12.031601905822754, "learning_rate": 1.7599999999999999e-06, "loss": 0.7745881676673889, "step": 45 }, { "ce_loss": 6.0210430092411116e-05, "cls_loss": 0.03564453125, "epoch": 0.02702702702702703, "mask_bce_loss": 0.2503889203071594, "mask_dice_loss": 0.04854744300246239, "mask_loss": 0.2989363670349121, "step": 45 }, { "epoch": 0.02762762762762763, "grad_norm": 17.237585067749023, "learning_rate": 1.8e-06, "loss": 0.9934065341949463, "step": 46 }, { "ce_loss": 4.553304461296648e-05, "cls_loss": 0.021484375, "epoch": 0.02762762762762763, "mask_bce_loss": 0.26533985137939453, "mask_dice_loss": 0.03659762814640999, "mask_loss": 0.3019374907016754, "step": 46 }, { "epoch": 0.02822822822822823, "grad_norm": 19.409591674804688, "learning_rate": 1.84e-06, "loss": 0.9714037179946899, "step": 47 }, { "ce_loss": 0.0002202423638664186, "cls_loss": 0.046630859375, "epoch": 0.02822822822822823, "mask_bce_loss": 0.6096853613853455, "mask_dice_loss": 0.07129650563001633, "mask_loss": 0.6809818744659424, "step": 47 }, { "epoch": 0.02882882882882883, "grad_norm": 25.23453712463379, "learning_rate": 1.8799999999999998e-06, "loss": 0.9917163848876953, "step": 48 }, { "ce_loss": 0.0007484110537916422, "cls_loss": 0.0732421875, "epoch": 0.02882882882882883, "mask_bce_loss": 1.4511293172836304, "mask_dice_loss": 0.1195482611656189, "mask_loss": 1.5706775188446045, "step": 48 }, { "epoch": 0.02942942942942943, "grad_norm": 26.78529167175293, "learning_rate": 1.92e-06, "loss": 0.998130202293396, "step": 49 }, { "ce_loss": 0.0001611205516383052, "cls_loss": 0.048583984375, "epoch": 0.02942942942942943, "mask_bce_loss": 1.1883331537246704, "mask_dice_loss": 0.09713847935199738, "mask_loss": 1.2854716777801514, "step": 49 }, { "epoch": 0.03003003003003003, "grad_norm": 45.17092514038086, "learning_rate": 1.96e-06, "loss": 0.9844520688056946, "step": 50 }, { "ce_loss": 4.2264346120646223e-05, "cls_loss": 0.039306640625, "epoch": 0.03003003003003003, "mask_bce_loss": 0.43732449412345886, "mask_dice_loss": 0.04241565987467766, "mask_loss": 0.4797401428222656, "step": 50 }, { "epoch": 0.03063063063063063, "grad_norm": 18.622940063476562, "learning_rate": 2e-06, "loss": 0.7431746125221252, "step": 51 }, { "ce_loss": 4.524159157881513e-05, "cls_loss": 0.05224609375, "epoch": 0.03063063063063063, "mask_bce_loss": 0.6972489356994629, "mask_dice_loss": 0.06939223408699036, "mask_loss": 0.7666411399841309, "step": 51 }, { "epoch": 0.03123123123123123, "grad_norm": 23.081253051757812, "learning_rate": 1.9999981079850683e-06, "loss": 0.7102840542793274, "step": 52 }, { "ce_loss": 2.2744892703485675e-05, "cls_loss": 0.056640625, "epoch": 0.03123123123123123, "mask_bce_loss": 0.7263869643211365, "mask_dice_loss": 0.11512754112482071, "mask_loss": 0.841514527797699, "step": 52 }, { "epoch": 0.03183183183183183, "grad_norm": 17.701669692993164, "learning_rate": 1.999992431947434e-06, "loss": 0.8800511360168457, "step": 53 }, { "ce_loss": 3.658418427221477e-05, "cls_loss": 0.031005859375, "epoch": 0.03183183183183183, "mask_bce_loss": 0.47442278265953064, "mask_dice_loss": 0.023290280252695084, "mask_loss": 0.4977130591869354, "step": 53 }, { "epoch": 0.032432432432432434, "grad_norm": 16.636831283569336, "learning_rate": 1.999982971908574e-06, "loss": 0.950873851776123, "step": 54 }, { "ce_loss": 4.0989951230585575e-05, "cls_loss": 0.0546875, "epoch": 0.032432432432432434, "mask_bce_loss": 1.1000216007232666, "mask_dice_loss": 0.06958293169736862, "mask_loss": 1.1696045398712158, "step": 54 }, { "epoch": 0.03303303303303303, "grad_norm": 25.972688674926758, "learning_rate": 1.9999697279042863e-06, "loss": 0.9910076856613159, "step": 55 }, { "ce_loss": 0.0010029032127931714, "cls_loss": 0.05517578125, "epoch": 0.03303303303303303, "mask_bce_loss": 0.5584525465965271, "mask_dice_loss": 0.09112341701984406, "mask_loss": 0.64957594871521, "step": 55 }, { "epoch": 0.033633633633633635, "grad_norm": 25.385629653930664, "learning_rate": 1.9999526999846863e-06, "loss": 0.9112304449081421, "step": 56 }, { "ce_loss": 8.466960571240634e-05, "cls_loss": 0.05859375, "epoch": 0.033633633633633635, "mask_bce_loss": 1.2131633758544922, "mask_dice_loss": 0.10453623533248901, "mask_loss": 1.317699670791626, "step": 56 }, { "epoch": 0.03423423423423423, "grad_norm": 18.51480484008789, "learning_rate": 1.999931888214208e-06, "loss": 0.8271615505218506, "step": 57 }, { "ce_loss": 5.422481626737863e-05, "cls_loss": 0.0341796875, "epoch": 0.03423423423423423, "mask_bce_loss": 0.3000411093235016, "mask_dice_loss": 0.0262825395911932, "mask_loss": 0.32632365822792053, "step": 57 }, { "epoch": 0.034834834834834835, "grad_norm": 16.970102310180664, "learning_rate": 1.9999072926716043e-06, "loss": 0.7999513149261475, "step": 58 }, { "ce_loss": 0.00014173495583236217, "cls_loss": 0.0308837890625, "epoch": 0.034834834834834835, "mask_bce_loss": 0.34747347235679626, "mask_dice_loss": 0.028853435069322586, "mask_loss": 0.37632691860198975, "step": 58 }, { "epoch": 0.03543543543543543, "grad_norm": 28.039594650268555, "learning_rate": 1.999878913449945e-06, "loss": 0.9334844350814819, "step": 59 }, { "ce_loss": 0.0034262582194060087, "cls_loss": 0.044921875, "epoch": 0.03543543543543543, "mask_bce_loss": 0.6357197165489197, "mask_dice_loss": 0.0530390702188015, "mask_loss": 0.6887587904930115, "step": 59 }, { "epoch": 0.036036036036036036, "grad_norm": 30.741708755493164, "learning_rate": 1.999846750656618e-06, "loss": 0.8090206384658813, "step": 60 }, { "ce_loss": 9.973296982934698e-05, "cls_loss": 0.050048828125, "epoch": 0.036036036036036036, "mask_bce_loss": 0.5246767401695251, "mask_dice_loss": 0.048053134232759476, "mask_loss": 0.5727298855781555, "step": 60 }, { "epoch": 0.03663663663663664, "grad_norm": 18.749467849731445, "learning_rate": 1.999810804413328e-06, "loss": 0.9542373418807983, "step": 61 }, { "ce_loss": 0.00020475017663557082, "cls_loss": 0.051513671875, "epoch": 0.03663663663663664, "mask_bce_loss": 0.6127321720123291, "mask_dice_loss": 0.08857019990682602, "mask_loss": 0.7013023495674133, "step": 61 }, { "epoch": 0.037237237237237236, "grad_norm": 29.079055786132812, "learning_rate": 1.9997710748560977e-06, "loss": 1.067537546157837, "step": 62 }, { "ce_loss": 0.00011027840810129419, "cls_loss": 0.0179443359375, "epoch": 0.037237237237237236, "mask_bce_loss": 0.2358662337064743, "mask_dice_loss": 0.009749501012265682, "mask_loss": 0.24561573565006256, "step": 62 }, { "epoch": 0.03783783783783784, "grad_norm": 23.854373931884766, "learning_rate": 1.9997275621352635e-06, "loss": 0.9035201668739319, "step": 63 }, { "ce_loss": 5.425048220786266e-05, "cls_loss": 0.038330078125, "epoch": 0.03783783783783784, "mask_bce_loss": 0.7272576689720154, "mask_dice_loss": 0.051964033395051956, "mask_loss": 0.7792217135429382, "step": 63 }, { "epoch": 0.03843843843843844, "grad_norm": 18.042118072509766, "learning_rate": 1.99968026641548e-06, "loss": 0.8922308683395386, "step": 64 }, { "ce_loss": 2.662168481037952e-05, "cls_loss": 0.03271484375, "epoch": 0.03843843843843844, "mask_bce_loss": 0.5677309632301331, "mask_dice_loss": 0.022507717832922935, "mask_loss": 0.5902386903762817, "step": 64 }, { "epoch": 0.03903903903903904, "grad_norm": 37.09247970581055, "learning_rate": 1.999629187875715e-06, "loss": 0.8827009201049805, "step": 65 }, { "ce_loss": 8.443989645456895e-05, "cls_loss": 0.05517578125, "epoch": 0.03903903903903904, "mask_bce_loss": 0.44265681505203247, "mask_dice_loss": 0.08160125464200974, "mask_loss": 0.5242580771446228, "step": 65 }, { "epoch": 0.03963963963963964, "grad_norm": 30.386507034301758, "learning_rate": 1.999574326709251e-06, "loss": 0.8929530382156372, "step": 66 }, { "ce_loss": 0.0017055505886673927, "cls_loss": 0.040283203125, "epoch": 0.03963963963963964, "mask_bce_loss": 0.3534471094608307, "mask_dice_loss": 0.03739039972424507, "mask_loss": 0.39083752036094666, "step": 66 }, { "epoch": 0.04024024024024024, "grad_norm": 24.865018844604492, "learning_rate": 1.9995156831236853e-06, "loss": 0.9024133682250977, "step": 67 }, { "ce_loss": 8.646048081573099e-05, "cls_loss": 0.04443359375, "epoch": 0.04024024024024024, "mask_bce_loss": 0.3700248599052429, "mask_dice_loss": 0.03976249694824219, "mask_loss": 0.4097873568534851, "step": 67 }, { "epoch": 0.04084084084084084, "grad_norm": 38.54929733276367, "learning_rate": 1.999453257340926e-06, "loss": 1.1116657257080078, "step": 68 }, { "ce_loss": 1.8078129869536497e-05, "cls_loss": 0.04296875, "epoch": 0.04084084084084084, "mask_bce_loss": 0.35837674140930176, "mask_dice_loss": 0.036850880831480026, "mask_loss": 0.3952276110649109, "step": 68 }, { "epoch": 0.04144144144144144, "grad_norm": 44.5231819152832, "learning_rate": 1.9993870495971944e-06, "loss": 1.0469083786010742, "step": 69 }, { "ce_loss": 6.634095188928768e-05, "cls_loss": 0.055419921875, "epoch": 0.04144144144144144, "mask_bce_loss": 1.3647772073745728, "mask_dice_loss": 0.11531170457601547, "mask_loss": 1.4800889492034912, "step": 69 }, { "epoch": 0.042042042042042045, "grad_norm": 20.110790252685547, "learning_rate": 1.999317060143023e-06, "loss": 0.8425664305686951, "step": 70 }, { "ce_loss": 0.0014936239458620548, "cls_loss": 0.05029296875, "epoch": 0.042042042042042045, "mask_bce_loss": 0.8923651576042175, "mask_dice_loss": 0.09944449365139008, "mask_loss": 0.9918096661567688, "step": 70 }, { "epoch": 0.04264264264264264, "grad_norm": 23.441234588623047, "learning_rate": 1.999243289243253e-06, "loss": 0.7741584777832031, "step": 71 }, { "ce_loss": 2.098174809361808e-05, "cls_loss": 0.03369140625, "epoch": 0.04264264264264264, "mask_bce_loss": 0.5061402320861816, "mask_dice_loss": 0.027661150321364403, "mask_loss": 0.5338013768196106, "step": 71 }, { "epoch": 0.043243243243243246, "grad_norm": 20.94329261779785, "learning_rate": 1.999165737177037e-06, "loss": 0.9619312286376953, "step": 72 }, { "ce_loss": 3.071580795221962e-05, "cls_loss": 0.057373046875, "epoch": 0.043243243243243246, "mask_bce_loss": 0.36447399854660034, "mask_dice_loss": 0.11872144043445587, "mask_loss": 0.483195424079895, "step": 72 }, { "epoch": 0.04384384384384384, "grad_norm": 47.0284423828125, "learning_rate": 1.999084404237833e-06, "loss": 0.9900135397911072, "step": 73 }, { "ce_loss": 4.8267655074596405e-05, "cls_loss": 0.053955078125, "epoch": 0.04384384384384384, "mask_bce_loss": 0.8514385223388672, "mask_dice_loss": 0.09037443995475769, "mask_loss": 0.9418129920959473, "step": 73 }, { "epoch": 0.044444444444444446, "grad_norm": 29.103445053100586, "learning_rate": 1.9989992907334086e-06, "loss": 0.9866974353790283, "step": 74 }, { "ce_loss": 7.30280953575857e-05, "cls_loss": 0.04443359375, "epoch": 0.044444444444444446, "mask_bce_loss": 0.6565348505973816, "mask_dice_loss": 0.0418408177793026, "mask_loss": 0.6983756422996521, "step": 74 }, { "epoch": 0.04504504504504504, "grad_norm": 21.53803253173828, "learning_rate": 1.9989103969858346e-06, "loss": 0.8234096169471741, "step": 75 }, { "ce_loss": 6.853985541965812e-05, "cls_loss": 0.052001953125, "epoch": 0.04504504504504504, "mask_bce_loss": 0.7527948617935181, "mask_dice_loss": 0.05871155858039856, "mask_loss": 0.8115063905715942, "step": 75 }, { "epoch": 0.04564564564564565, "grad_norm": 18.141559600830078, "learning_rate": 1.9988177233314887e-06, "loss": 1.0149801969528198, "step": 76 }, { "ce_loss": 4.2475046939216554e-05, "cls_loss": 0.04833984375, "epoch": 0.04564564564564565, "mask_bce_loss": 0.4965845048427582, "mask_dice_loss": 0.09383152425289154, "mask_loss": 0.5904160141944885, "step": 76 }, { "epoch": 0.04624624624624624, "grad_norm": 27.105131149291992, "learning_rate": 1.9987212701210502e-06, "loss": 0.976067304611206, "step": 77 }, { "ce_loss": 0.00016863938071765006, "cls_loss": 0.041015625, "epoch": 0.04624624624624624, "mask_bce_loss": 0.8071174025535583, "mask_dice_loss": 0.0870320200920105, "mask_loss": 0.8941494226455688, "step": 77 }, { "epoch": 0.04684684684684685, "grad_norm": 33.87481689453125, "learning_rate": 1.9986210377195005e-06, "loss": 0.9381427764892578, "step": 78 }, { "ce_loss": 4.887039176537655e-05, "cls_loss": 0.0654296875, "epoch": 0.04684684684684685, "mask_bce_loss": 0.3854735195636749, "mask_dice_loss": 0.10779014974832535, "mask_loss": 0.4932636618614197, "step": 78 }, { "epoch": 0.04744744744744745, "grad_norm": 22.114103317260742, "learning_rate": 1.9985170265061227e-06, "loss": 0.9297803640365601, "step": 79 }, { "ce_loss": 5.054527355241589e-05, "cls_loss": 0.061279296875, "epoch": 0.04744744744744745, "mask_bce_loss": 0.7543637752532959, "mask_dice_loss": 0.0807674303650856, "mask_loss": 0.8351312279701233, "step": 79 }, { "epoch": 0.04804804804804805, "grad_norm": 20.775358200073242, "learning_rate": 1.9984092368744982e-06, "loss": 0.8922678232192993, "step": 80 }, { "ce_loss": 0.0014791175490245223, "cls_loss": 0.042236328125, "epoch": 0.04804804804804805, "mask_bce_loss": 0.3931247889995575, "mask_dice_loss": 0.06008703634142876, "mask_loss": 0.45321181416511536, "step": 80 }, { "epoch": 0.04864864864864865, "grad_norm": 23.594331741333008, "learning_rate": 1.9982976692325058e-06, "loss": 0.9315401315689087, "step": 81 }, { "ce_loss": 2.417824543954339e-05, "cls_loss": 0.0235595703125, "epoch": 0.04864864864864865, "mask_bce_loss": 0.16753076016902924, "mask_dice_loss": 0.015696654096245766, "mask_loss": 0.18322741985321045, "step": 81 }, { "epoch": 0.04924924924924925, "grad_norm": 23.18486785888672, "learning_rate": 1.9981823240023213e-06, "loss": 0.9902520179748535, "step": 82 }, { "ce_loss": 0.0013127216370776296, "cls_loss": 0.060546875, "epoch": 0.04924924924924925, "mask_bce_loss": 1.5319952964782715, "mask_dice_loss": 0.1290401816368103, "mask_loss": 1.6610355377197266, "step": 82 }, { "epoch": 0.04984984984984985, "grad_norm": 16.940628051757812, "learning_rate": 1.9980632016204146e-06, "loss": 0.7802420854568481, "step": 83 }, { "ce_loss": 4.403320781420916e-05, "cls_loss": 0.04150390625, "epoch": 0.04984984984984985, "mask_bce_loss": 0.5900759100914001, "mask_dice_loss": 0.08737100660800934, "mask_loss": 0.6774469017982483, "step": 83 }, { "epoch": 0.05045045045045045, "grad_norm": 26.203927993774414, "learning_rate": 1.9979403025375473e-06, "loss": 0.9577487111091614, "step": 84 }, { "ce_loss": 0.0004145260027144104, "cls_loss": 0.0654296875, "epoch": 0.05045045045045045, "mask_bce_loss": 0.6390252709388733, "mask_dice_loss": 0.1829749345779419, "mask_loss": 0.8220002055168152, "step": 84 }, { "epoch": 0.05105105105105105, "grad_norm": 22.672513961791992, "learning_rate": 1.9978136272187744e-06, "loss": 1.025286316871643, "step": 85 }, { "ce_loss": 2.713100184337236e-05, "cls_loss": 0.05517578125, "epoch": 0.05105105105105105, "mask_bce_loss": 1.2262829542160034, "mask_dice_loss": 0.07243143022060394, "mask_loss": 1.2987143993377686, "step": 85 }, { "epoch": 0.05165165165165165, "grad_norm": 13.581890106201172, "learning_rate": 1.9976831761434386e-06, "loss": 0.8612722158432007, "step": 86 }, { "ce_loss": 0.00039063821895979345, "cls_loss": 0.04248046875, "epoch": 0.05165165165165165, "mask_bce_loss": 0.1122715026140213, "mask_dice_loss": 0.03583620861172676, "mask_loss": 0.14810770750045776, "step": 86 }, { "epoch": 0.05225225225225225, "grad_norm": 36.57144546508789, "learning_rate": 1.9975489498051704e-06, "loss": 0.9520080089569092, "step": 87 }, { "ce_loss": 4.4023287046002224e-05, "cls_loss": 0.059326171875, "epoch": 0.05225225225225225, "mask_bce_loss": 1.128250002861023, "mask_dice_loss": 0.10573263466358185, "mask_loss": 1.2339826822280884, "step": 87 }, { "epoch": 0.05285285285285285, "grad_norm": 20.786609649658203, "learning_rate": 1.997410948711887e-06, "loss": 0.8213093876838684, "step": 88 }, { "ce_loss": 0.00020382704678922892, "cls_loss": 0.047607421875, "epoch": 0.05285285285285285, "mask_bce_loss": 1.602237343788147, "mask_dice_loss": 0.18456940352916718, "mask_loss": 1.7868067026138306, "step": 88 }, { "epoch": 0.05345345345345345, "grad_norm": 31.220556259155273, "learning_rate": 1.997269173385788e-06, "loss": 0.9944270253181458, "step": 89 }, { "ce_loss": 5.595104812528007e-05, "cls_loss": 0.05419921875, "epoch": 0.05345345345345345, "mask_bce_loss": 0.7117537260055542, "mask_dice_loss": 0.09142611920833588, "mask_loss": 0.8031798601150513, "step": 89 }, { "epoch": 0.05405405405405406, "grad_norm": 26.581884384155273, "learning_rate": 1.9971236243633556e-06, "loss": 0.9789072275161743, "step": 90 }, { "ce_loss": 3.054776243516244e-05, "cls_loss": 0.055908203125, "epoch": 0.05405405405405406, "mask_bce_loss": 0.8667566180229187, "mask_dice_loss": 0.08821996301412582, "mask_loss": 0.9549765586853027, "step": 90 }, { "epoch": 0.054654654654654654, "grad_norm": 23.00945281982422, "learning_rate": 1.996974302195352e-06, "loss": 1.0500373840332031, "step": 91 }, { "ce_loss": 2.945548840216361e-05, "cls_loss": 0.0478515625, "epoch": 0.054654654654654654, "mask_bce_loss": 0.6279355883598328, "mask_dice_loss": 0.06382519006729126, "mask_loss": 0.691760778427124, "step": 91 }, { "epoch": 0.05525525525525526, "grad_norm": 38.46200180053711, "learning_rate": 1.996821207446816e-06, "loss": 1.1137824058532715, "step": 92 }, { "ce_loss": 4.4216099922778085e-05, "cls_loss": 0.04248046875, "epoch": 0.05525525525525526, "mask_bce_loss": 0.37655892968177795, "mask_dice_loss": 0.1168370246887207, "mask_loss": 0.49339595437049866, "step": 92 }, { "epoch": 0.055855855855855854, "grad_norm": 23.60446548461914, "learning_rate": 1.9966643406970633e-06, "loss": 0.9239801168441772, "step": 93 }, { "ce_loss": 0.0002095820236718282, "cls_loss": 0.05908203125, "epoch": 0.055855855855855854, "mask_bce_loss": 0.6273679733276367, "mask_dice_loss": 0.0935085266828537, "mask_loss": 0.7208765149116516, "step": 93 }, { "epoch": 0.05645645645645646, "grad_norm": 43.799129486083984, "learning_rate": 1.9965037025396823e-06, "loss": 1.0121986865997314, "step": 94 }, { "ce_loss": 6.316378858173266e-05, "cls_loss": 0.05517578125, "epoch": 0.05645645645645646, "mask_bce_loss": 0.7444838285446167, "mask_dice_loss": 0.11520715057849884, "mask_loss": 0.8596909642219543, "step": 94 }, { "epoch": 0.057057057057057055, "grad_norm": 24.01621437072754, "learning_rate": 1.996339293582533e-06, "loss": 0.8694986701011658, "step": 95 }, { "ce_loss": 4.265653114998713e-05, "cls_loss": 0.0458984375, "epoch": 0.057057057057057055, "mask_bce_loss": 0.737576961517334, "mask_dice_loss": 0.08345189690589905, "mask_loss": 0.8210288286209106, "step": 95 }, { "epoch": 0.05765765765765766, "grad_norm": 30.850618362426758, "learning_rate": 1.9961711144477425e-06, "loss": 0.9324817657470703, "step": 96 }, { "ce_loss": 9.077792492462322e-05, "cls_loss": 0.05615234375, "epoch": 0.05765765765765766, "mask_bce_loss": 0.6992975473403931, "mask_dice_loss": 0.05720030143857002, "mask_loss": 0.756497859954834, "step": 96 }, { "epoch": 0.058258258258258255, "grad_norm": 27.92517852783203, "learning_rate": 1.995999165771707e-06, "loss": 1.0251307487487793, "step": 97 }, { "ce_loss": 4.5004668209003285e-05, "cls_loss": 0.04541015625, "epoch": 0.058258258258258255, "mask_bce_loss": 0.874725341796875, "mask_dice_loss": 0.07459896057844162, "mask_loss": 0.9493243098258972, "step": 97 }, { "epoch": 0.05885885885885886, "grad_norm": 25.155128479003906, "learning_rate": 1.995823448205085e-06, "loss": 1.1288833618164062, "step": 98 }, { "ce_loss": 3.579667827580124e-05, "cls_loss": 0.04443359375, "epoch": 0.05885885885885886, "mask_bce_loss": 0.3074103891849518, "mask_dice_loss": 0.13456366956233978, "mask_loss": 0.44197404384613037, "step": 98 }, { "epoch": 0.05945945945945946, "grad_norm": 22.638973236083984, "learning_rate": 1.995643962412797e-06, "loss": 0.939802348613739, "step": 99 }, { "ce_loss": 0.0004025427042506635, "cls_loss": 0.055908203125, "epoch": 0.05945945945945946, "mask_bce_loss": 0.8608024716377258, "mask_dice_loss": 0.07682748883962631, "mask_loss": 0.9376299381256104, "step": 99 }, { "epoch": 0.06006006006006006, "grad_norm": 13.704671859741211, "learning_rate": 1.9954607090740217e-06, "loss": 0.7687138915061951, "step": 100 }, { "ce_loss": 3.399871275178157e-05, "cls_loss": 0.056640625, "epoch": 0.06006006006006006, "mask_bce_loss": 0.7698205709457397, "mask_dice_loss": 0.1409829556941986, "mask_loss": 0.9108035564422607, "step": 100 }, { "epoch": 0.06066066066066066, "grad_norm": 28.906551361083984, "learning_rate": 1.9952736888821967e-06, "loss": 1.0610966682434082, "step": 101 }, { "ce_loss": 2.8794174795621075e-05, "cls_loss": 0.061279296875, "epoch": 0.06066066066066066, "mask_bce_loss": 0.7676653861999512, "mask_dice_loss": 0.12659069895744324, "mask_loss": 0.8942561149597168, "step": 101 }, { "epoch": 0.06126126126126126, "grad_norm": 17.444965362548828, "learning_rate": 1.9950829025450113e-06, "loss": 0.9832133650779724, "step": 102 }, { "ce_loss": 0.00020542321726679802, "cls_loss": 0.05908203125, "epoch": 0.06126126126126126, "mask_bce_loss": 0.37241992354393005, "mask_dice_loss": 0.1141527071595192, "mask_loss": 0.48657262325286865, "step": 102 }, { "epoch": 0.061861861861861864, "grad_norm": 32.544403076171875, "learning_rate": 1.9948883507844068e-06, "loss": 0.9697765707969666, "step": 103 }, { "ce_loss": 0.029938139021396637, "cls_loss": 0.06787109375, "epoch": 0.061861861861861864, "mask_bce_loss": 1.7673851251602173, "mask_dice_loss": 0.13493631780147552, "mask_loss": 1.902321457862854, "step": 103 }, { "epoch": 0.06246246246246246, "grad_norm": 35.48094940185547, "learning_rate": 1.9946900343365726e-06, "loss": 1.013505220413208, "step": 104 }, { "ce_loss": 7.487950642826036e-05, "cls_loss": 0.05419921875, "epoch": 0.06246246246246246, "mask_bce_loss": 1.206525206565857, "mask_dice_loss": 0.1909182369709015, "mask_loss": 1.397443413734436, "step": 104 }, { "epoch": 0.06306306306306306, "grad_norm": 16.067180633544922, "learning_rate": 1.9944879539519443e-06, "loss": 1.1398217678070068, "step": 105 }, { "ce_loss": 0.00039791606832295656, "cls_loss": 0.05078125, "epoch": 0.06306306306306306, "mask_bce_loss": 1.450212836265564, "mask_dice_loss": 0.08233171701431274, "mask_loss": 1.5325446128845215, "step": 105 }, { "epoch": 0.06366366366366366, "grad_norm": 19.080183029174805, "learning_rate": 1.9942821103952e-06, "loss": 0.8802724480628967, "step": 106 }, { "ce_loss": 5.4060950787970796e-05, "cls_loss": 0.0498046875, "epoch": 0.06366366366366366, "mask_bce_loss": 0.783575713634491, "mask_dice_loss": 0.05099291726946831, "mask_loss": 0.8345686197280884, "step": 106 }, { "epoch": 0.06426426426426426, "grad_norm": 34.35637664794922, "learning_rate": 1.994072504445258e-06, "loss": 0.843559741973877, "step": 107 }, { "ce_loss": 0.00741796987131238, "cls_loss": 0.03466796875, "epoch": 0.06426426426426426, "mask_bce_loss": 0.7555691599845886, "mask_dice_loss": 0.05996738001704216, "mask_loss": 0.8155365586280823, "step": 107 }, { "epoch": 0.06486486486486487, "grad_norm": 21.558565139770508, "learning_rate": 1.9938591368952737e-06, "loss": 0.8418678641319275, "step": 108 }, { "ce_loss": 0.00026349734980612993, "cls_loss": 0.035888671875, "epoch": 0.06486486486486487, "mask_bce_loss": 0.3331778347492218, "mask_dice_loss": 0.03533888980746269, "mask_loss": 0.3685167133808136, "step": 108 }, { "epoch": 0.06546546546546547, "grad_norm": 20.290485382080078, "learning_rate": 1.993642008552636e-06, "loss": 0.885840117931366, "step": 109 }, { "ce_loss": 0.0003045905032195151, "cls_loss": 0.06396484375, "epoch": 0.06546546546546547, "mask_bce_loss": 1.0321632623672485, "mask_dice_loss": 0.1099090725183487, "mask_loss": 1.142072319984436, "step": 109 }, { "epoch": 0.06606606606606606, "grad_norm": 19.055091857910156, "learning_rate": 1.9934211202389643e-06, "loss": 0.841249942779541, "step": 110 }, { "ce_loss": 8.646293281344697e-05, "cls_loss": 0.0498046875, "epoch": 0.06606606606606606, "mask_bce_loss": 0.3436753451824188, "mask_dice_loss": 0.049419473856687546, "mask_loss": 0.3930948078632355, "step": 110 }, { "epoch": 0.06666666666666667, "grad_norm": 28.19367790222168, "learning_rate": 1.993196472790108e-06, "loss": 0.8836717009544373, "step": 111 }, { "ce_loss": 0.00015148684906307608, "cls_loss": 0.05029296875, "epoch": 0.06666666666666667, "mask_bce_loss": 1.1667112112045288, "mask_dice_loss": 0.1020989790558815, "mask_loss": 1.2688101530075073, "step": 111 }, { "epoch": 0.06726726726726727, "grad_norm": 18.131668090820312, "learning_rate": 1.9929680670561385e-06, "loss": 0.8503056764602661, "step": 112 }, { "ce_loss": 2.8371070584398694e-05, "cls_loss": 0.060546875, "epoch": 0.06726726726726727, "mask_bce_loss": 1.2708624601364136, "mask_dice_loss": 0.13017533719539642, "mask_loss": 1.4010378122329712, "step": 112 }, { "epoch": 0.06786786786786787, "grad_norm": 40.45888900756836, "learning_rate": 1.992735903901351e-06, "loss": 0.9231182932853699, "step": 113 }, { "ce_loss": 0.00013486930401995778, "cls_loss": 0.06201171875, "epoch": 0.06786786786786787, "mask_bce_loss": 0.1685463935136795, "mask_dice_loss": 0.18010617792606354, "mask_loss": 0.34865257143974304, "step": 113 }, { "epoch": 0.06846846846846846, "grad_norm": 26.80726432800293, "learning_rate": 1.9924999842042573e-06, "loss": 0.8404207825660706, "step": 114 }, { "ce_loss": 0.000104749713500496, "cls_loss": 0.055908203125, "epoch": 0.06846846846846846, "mask_bce_loss": 1.9492815732955933, "mask_dice_loss": 0.08912631869316101, "mask_loss": 2.038407802581787, "step": 114 }, { "epoch": 0.06906906906906907, "grad_norm": 23.163307189941406, "learning_rate": 1.9922603088575844e-06, "loss": 0.865681529045105, "step": 115 }, { "ce_loss": 2.971014691866003e-05, "cls_loss": 0.09814453125, "epoch": 0.06906906906906907, "mask_bce_loss": 0.6503251194953918, "mask_dice_loss": 0.07251034677028656, "mask_loss": 0.7228354811668396, "step": 115 }, { "epoch": 0.06966966966966967, "grad_norm": 13.773252487182617, "learning_rate": 1.992016878768271e-06, "loss": 0.8787857294082642, "step": 116 }, { "ce_loss": 5.3260086133377627e-05, "cls_loss": 0.05419921875, "epoch": 0.06966966966966967, "mask_bce_loss": 0.6207614541053772, "mask_dice_loss": 0.06087351590394974, "mask_loss": 0.6816349625587463, "step": 116 }, { "epoch": 0.07027027027027027, "grad_norm": 73.23187255859375, "learning_rate": 1.9917696948574634e-06, "loss": 0.8564697504043579, "step": 117 }, { "ce_loss": 0.00015477431588806212, "cls_loss": 0.0439453125, "epoch": 0.07027027027027027, "mask_bce_loss": 0.5638112425804138, "mask_dice_loss": 0.04892183095216751, "mask_loss": 0.6127330660820007, "step": 117 }, { "epoch": 0.07087087087087086, "grad_norm": 33.6517219543457, "learning_rate": 1.9915187580605146e-06, "loss": 0.9501999020576477, "step": 118 }, { "ce_loss": 5.971776772639714e-05, "cls_loss": 0.06689453125, "epoch": 0.07087087087087086, "mask_bce_loss": 0.4523064196109772, "mask_dice_loss": 0.10040696710348129, "mask_loss": 0.5527133941650391, "step": 118 }, { "epoch": 0.07147147147147147, "grad_norm": 37.94041442871094, "learning_rate": 1.991264069326975e-06, "loss": 1.0229988098144531, "step": 119 }, { "ce_loss": 0.0002024211862590164, "cls_loss": 0.0289306640625, "epoch": 0.07147147147147147, "mask_bce_loss": 0.30918145179748535, "mask_dice_loss": 0.02215016447007656, "mask_loss": 0.33133161067962646, "step": 119 }, { "epoch": 0.07207207207207207, "grad_norm": 15.308018684387207, "learning_rate": 1.9910056296205957e-06, "loss": 0.8801918029785156, "step": 120 }, { "ce_loss": 0.00011185010225744918, "cls_loss": 0.04150390625, "epoch": 0.07207207207207207, "mask_bce_loss": 0.647178053855896, "mask_dice_loss": 0.05699244886636734, "mask_loss": 0.7041705250740051, "step": 120 }, { "epoch": 0.07267267267267268, "grad_norm": 19.575429916381836, "learning_rate": 1.990743439919319e-06, "loss": 0.8242614269256592, "step": 121 }, { "ce_loss": 0.00010036920866696164, "cls_loss": 0.052734375, "epoch": 0.07267267267267268, "mask_bce_loss": 1.0383965969085693, "mask_dice_loss": 0.08272313326597214, "mask_loss": 1.121119737625122, "step": 121 }, { "epoch": 0.07327327327327328, "grad_norm": 30.14656639099121, "learning_rate": 1.99047750121528e-06, "loss": 0.775652289390564, "step": 122 }, { "ce_loss": 4.6768185711698607e-05, "cls_loss": 0.0546875, "epoch": 0.07327327327327328, "mask_bce_loss": 0.6259490847587585, "mask_dice_loss": 0.07987018674612045, "mask_loss": 0.7058192491531372, "step": 122 }, { "epoch": 0.07387387387387387, "grad_norm": 22.776403427124023, "learning_rate": 1.990207814514798e-06, "loss": 0.8516128063201904, "step": 123 }, { "ce_loss": 0.00024033556110225618, "cls_loss": 0.06201171875, "epoch": 0.07387387387387387, "mask_bce_loss": 1.1710931062698364, "mask_dice_loss": 0.12008779495954514, "mask_loss": 1.2911808490753174, "step": 123 }, { "epoch": 0.07447447447447447, "grad_norm": 19.22648811340332, "learning_rate": 1.9899343808383754e-06, "loss": 0.8205176591873169, "step": 124 }, { "ce_loss": 0.0015223994851112366, "cls_loss": 0.06591796875, "epoch": 0.07447447447447447, "mask_bce_loss": 1.018286108970642, "mask_dice_loss": 0.15277497470378876, "mask_loss": 1.1710610389709473, "step": 124 }, { "epoch": 0.07507507507507508, "grad_norm": 26.875865936279297, "learning_rate": 1.989657201220693e-06, "loss": 0.8434844017028809, "step": 125 }, { "ce_loss": 3.288321750005707e-05, "cls_loss": 0.04296875, "epoch": 0.07507507507507508, "mask_bce_loss": 0.8500728011131287, "mask_dice_loss": 0.0461137555539608, "mask_loss": 0.8961865305900574, "step": 125 }, { "epoch": 0.07567567567567568, "grad_norm": 20.15418243408203, "learning_rate": 1.9893762767106076e-06, "loss": 0.9366012811660767, "step": 126 }, { "ce_loss": 2.9156104574212804e-05, "cls_loss": 0.057861328125, "epoch": 0.07567567567567568, "mask_bce_loss": 0.5609766840934753, "mask_dice_loss": 0.14509639143943787, "mask_loss": 0.7060730457305908, "step": 126 }, { "epoch": 0.07627627627627627, "grad_norm": 17.474639892578125, "learning_rate": 1.989091608371146e-06, "loss": 0.9462077617645264, "step": 127 }, { "ce_loss": 2.8203769034007564e-05, "cls_loss": 0.09521484375, "epoch": 0.07627627627627627, "mask_bce_loss": 1.1023681163787842, "mask_dice_loss": 0.16188080608844757, "mask_loss": 1.2642489671707153, "step": 127 }, { "epoch": 0.07687687687687687, "grad_norm": 40.368980407714844, "learning_rate": 1.988803197279501e-06, "loss": 1.082580804824829, "step": 128 }, { "ce_loss": 4.2982399463653564e-05, "cls_loss": 0.05322265625, "epoch": 0.07687687687687687, "mask_bce_loss": 1.5507673025131226, "mask_dice_loss": 0.06983587890863419, "mask_loss": 1.6206032037734985, "step": 128 }, { "epoch": 0.07747747747747748, "grad_norm": 25.51866912841797, "learning_rate": 1.988511044527029e-06, "loss": 0.9220490455627441, "step": 129 }, { "ce_loss": 2.3917746148072183e-05, "cls_loss": 0.0693359375, "epoch": 0.07747747747747748, "mask_bce_loss": 0.8839642405509949, "mask_dice_loss": 0.10151471197605133, "mask_loss": 0.985478937625885, "step": 129 }, { "epoch": 0.07807807807807808, "grad_norm": 24.059642791748047, "learning_rate": 1.988215151219245e-06, "loss": 0.8117259740829468, "step": 130 }, { "ce_loss": 4.995604103896767e-05, "cls_loss": 0.029541015625, "epoch": 0.07807807807807808, "mask_bce_loss": 0.650440514087677, "mask_dice_loss": 0.029463350772857666, "mask_loss": 0.6799038648605347, "step": 130 }, { "epoch": 0.07867867867867868, "grad_norm": 16.48142433166504, "learning_rate": 1.9879155184758176e-06, "loss": 0.7253130078315735, "step": 131 }, { "ce_loss": 4.92359213239979e-05, "cls_loss": 0.06103515625, "epoch": 0.07867867867867868, "mask_bce_loss": 0.5303481817245483, "mask_dice_loss": 0.10546465218067169, "mask_loss": 0.6358128190040588, "step": 131 }, { "epoch": 0.07927927927927927, "grad_norm": 27.390382766723633, "learning_rate": 1.9876121474305663e-06, "loss": 0.8190069794654846, "step": 132 }, { "ce_loss": 2.6162513677263632e-05, "cls_loss": 0.053466796875, "epoch": 0.07927927927927927, "mask_bce_loss": 0.7320281863212585, "mask_dice_loss": 0.06778999418020248, "mask_loss": 0.7998181581497192, "step": 132 }, { "epoch": 0.07987987987987988, "grad_norm": 29.31528091430664, "learning_rate": 1.9873050392314566e-06, "loss": 1.0874347686767578, "step": 133 }, { "ce_loss": 3.370274862390943e-05, "cls_loss": 0.03857421875, "epoch": 0.07987987987987988, "mask_bce_loss": 1.1374777555465698, "mask_dice_loss": 0.036372698843479156, "mask_loss": 1.173850417137146, "step": 133 }, { "epoch": 0.08048048048048048, "grad_norm": 19.20989418029785, "learning_rate": 1.986994195040595e-06, "loss": 0.7924736738204956, "step": 134 }, { "ce_loss": 4.79560112580657e-05, "cls_loss": 0.060546875, "epoch": 0.08048048048048048, "mask_bce_loss": 0.3524957299232483, "mask_dice_loss": 0.07142358273267746, "mask_loss": 0.42391932010650635, "step": 134 }, { "epoch": 0.08108108108108109, "grad_norm": 35.39643478393555, "learning_rate": 1.9866796160342245e-06, "loss": 0.9140077829360962, "step": 135 }, { "ce_loss": 2.064753971353639e-05, "cls_loss": 0.04052734375, "epoch": 0.08108108108108109, "mask_bce_loss": 0.4489274024963379, "mask_dice_loss": 0.03792823106050491, "mask_loss": 0.4868556261062622, "step": 135 }, { "epoch": 0.08168168168168168, "grad_norm": 24.068910598754883, "learning_rate": 1.9863613034027223e-06, "loss": 0.8377690315246582, "step": 136 }, { "ce_loss": 2.7512291126186028e-05, "cls_loss": 0.0269775390625, "epoch": 0.08168168168168168, "mask_bce_loss": 0.18669584393501282, "mask_dice_loss": 0.020502375438809395, "mask_loss": 0.20719821751117706, "step": 136 }, { "epoch": 0.08228228228228228, "grad_norm": 25.38892364501953, "learning_rate": 1.9860392583505925e-06, "loss": 0.9775824546813965, "step": 137 }, { "ce_loss": 0.00043907889630645514, "cls_loss": 0.0361328125, "epoch": 0.08228228228228228, "mask_bce_loss": 0.31484419107437134, "mask_dice_loss": 0.03251729533076286, "mask_loss": 0.3473614752292633, "step": 137 }, { "epoch": 0.08288288288288288, "grad_norm": 28.209543228149414, "learning_rate": 1.985713482096463e-06, "loss": 1.0288143157958984, "step": 138 }, { "ce_loss": 0.00015250100113917142, "cls_loss": 0.03759765625, "epoch": 0.08288288288288288, "mask_bce_loss": 0.43907347321510315, "mask_dice_loss": 0.0584295354783535, "mask_loss": 0.49750301241874695, "step": 138 }, { "epoch": 0.08348348348348349, "grad_norm": 37.3709602355957, "learning_rate": 1.985383975873081e-06, "loss": 0.9430665969848633, "step": 139 }, { "ce_loss": 1.8933031242340803e-05, "cls_loss": 0.06005859375, "epoch": 0.08348348348348349, "mask_bce_loss": 1.6116828918457031, "mask_dice_loss": 0.0894361361861229, "mask_loss": 1.701119065284729, "step": 139 }, { "epoch": 0.08408408408408409, "grad_norm": 32.14317321777344, "learning_rate": 1.9850507409273087e-06, "loss": 1.0180997848510742, "step": 140 }, { "ce_loss": 0.00013013249554205686, "cls_loss": 0.0693359375, "epoch": 0.08408408408408409, "mask_bce_loss": 0.7450197339057922, "mask_dice_loss": 0.1314474642276764, "mask_loss": 0.876467227935791, "step": 140 }, { "epoch": 0.08468468468468468, "grad_norm": 14.96213150024414, "learning_rate": 1.9847137785201155e-06, "loss": 0.832472562789917, "step": 141 }, { "ce_loss": 9.642042277846485e-05, "cls_loss": 0.0830078125, "epoch": 0.08468468468468468, "mask_bce_loss": 0.8686081767082214, "mask_dice_loss": 0.09915093332529068, "mask_loss": 0.9677591323852539, "step": 141 }, { "epoch": 0.08528528528528528, "grad_norm": 31.235963821411133, "learning_rate": 1.9843730899265783e-06, "loss": 1.0808894634246826, "step": 142 }, { "ce_loss": 4.2765394027810544e-05, "cls_loss": 0.055908203125, "epoch": 0.08528528528528528, "mask_bce_loss": 0.9345149397850037, "mask_dice_loss": 0.08005532622337341, "mask_loss": 1.0145702362060547, "step": 142 }, { "epoch": 0.08588588588588589, "grad_norm": 26.361038208007812, "learning_rate": 1.9840286764358732e-06, "loss": 0.9584890007972717, "step": 143 }, { "ce_loss": 2.999777461809572e-05, "cls_loss": 0.04150390625, "epoch": 0.08588588588588589, "mask_bce_loss": 0.34405359625816345, "mask_dice_loss": 0.036001306027173996, "mask_loss": 0.38005489110946655, "step": 143 }, { "epoch": 0.08648648648648649, "grad_norm": 31.638948440551758, "learning_rate": 1.9836805393512705e-06, "loss": 1.0084209442138672, "step": 144 }, { "ce_loss": 5.6131593737518415e-05, "cls_loss": 0.04736328125, "epoch": 0.08648648648648649, "mask_bce_loss": 0.6732069849967957, "mask_dice_loss": 0.13145646452903748, "mask_loss": 0.8046634197235107, "step": 144 }, { "epoch": 0.08708708708708708, "grad_norm": 58.04975891113281, "learning_rate": 1.9833286799901308e-06, "loss": 0.8973804116249084, "step": 145 }, { "ce_loss": 8.126778266159818e-05, "cls_loss": 0.041259765625, "epoch": 0.08708708708708708, "mask_bce_loss": 0.366247296333313, "mask_dice_loss": 0.03982849046587944, "mask_loss": 0.40607577562332153, "step": 145 }, { "epoch": 0.08768768768768768, "grad_norm": 26.276430130004883, "learning_rate": 1.982973099683902e-06, "loss": 0.8902683258056641, "step": 146 }, { "ce_loss": 0.05384266376495361, "cls_loss": 0.06640625, "epoch": 0.08768768768768768, "mask_bce_loss": 1.6902745962142944, "mask_dice_loss": 0.11530786007642746, "mask_loss": 1.8055824041366577, "step": 146 }, { "epoch": 0.08828828828828829, "grad_norm": 31.379310607910156, "learning_rate": 1.982613799778109e-06, "loss": 0.9339362382888794, "step": 147 }, { "ce_loss": 2.8285454391152598e-05, "cls_loss": 0.05322265625, "epoch": 0.08828828828828829, "mask_bce_loss": 0.820722222328186, "mask_dice_loss": 0.09065067023038864, "mask_loss": 0.9113729000091553, "step": 147 }, { "epoch": 0.08888888888888889, "grad_norm": 18.721960067749023, "learning_rate": 1.9822507816323543e-06, "loss": 0.9546916484832764, "step": 148 }, { "ce_loss": 2.8173040846013464e-05, "cls_loss": 0.05078125, "epoch": 0.08888888888888889, "mask_bce_loss": 0.36409667134284973, "mask_dice_loss": 0.10530239343643188, "mask_loss": 0.4693990647792816, "step": 148 }, { "epoch": 0.0894894894894895, "grad_norm": 23.661008834838867, "learning_rate": 1.981884046620309e-06, "loss": 0.965652346611023, "step": 149 }, { "ce_loss": 0.00012227580009493977, "cls_loss": 0.037353515625, "epoch": 0.0894894894894895, "mask_bce_loss": 0.19620749354362488, "mask_dice_loss": 0.030255962163209915, "mask_loss": 0.2264634519815445, "step": 149 }, { "epoch": 0.09009009009009009, "grad_norm": 20.511754989624023, "learning_rate": 1.9815135961297097e-06, "loss": 0.9848958253860474, "step": 150 }, { "ce_loss": 6.656722689513117e-05, "cls_loss": 0.0546875, "epoch": 0.09009009009009009, "mask_bce_loss": 0.9323148727416992, "mask_dice_loss": 0.07454495131969452, "mask_loss": 1.0068597793579102, "step": 150 }, { "epoch": 0.09069069069069069, "grad_norm": 16.90283203125, "learning_rate": 1.981139431562352e-06, "loss": 0.9164067506790161, "step": 151 }, { "ce_loss": 0.0002159346331609413, "cls_loss": 0.05419921875, "epoch": 0.09069069069069069, "mask_bce_loss": 0.8215644955635071, "mask_dice_loss": 0.0813373401761055, "mask_loss": 0.902901828289032, "step": 151 }, { "epoch": 0.0912912912912913, "grad_norm": 31.650230407714844, "learning_rate": 1.9807615543340853e-06, "loss": 0.9034360647201538, "step": 152 }, { "ce_loss": 0.00010811853280756623, "cls_loss": 0.0634765625, "epoch": 0.0912912912912913, "mask_bce_loss": 0.7421200275421143, "mask_dice_loss": 0.101187564432621, "mask_loss": 0.843307614326477, "step": 152 }, { "epoch": 0.0918918918918919, "grad_norm": 16.6004638671875, "learning_rate": 1.9803799658748095e-06, "loss": 0.7942208647727966, "step": 153 }, { "ce_loss": 5.271506597637199e-05, "cls_loss": 0.05859375, "epoch": 0.0918918918918919, "mask_bce_loss": 1.1103099584579468, "mask_dice_loss": 0.10293933004140854, "mask_loss": 1.2132493257522583, "step": 153 }, { "epoch": 0.09249249249249249, "grad_norm": 21.55391502380371, "learning_rate": 1.9799946676284653e-06, "loss": 1.0032879114151, "step": 154 }, { "ce_loss": 4.0907529182732105e-05, "cls_loss": 0.05615234375, "epoch": 0.09249249249249249, "mask_bce_loss": 0.6717383861541748, "mask_dice_loss": 0.07857032865285873, "mask_loss": 0.7503086924552917, "step": 154 }, { "epoch": 0.09309309309309309, "grad_norm": 49.78294372558594, "learning_rate": 1.9796056610530338e-06, "loss": 0.9221913814544678, "step": 155 }, { "ce_loss": 7.689862104598433e-05, "cls_loss": 0.04052734375, "epoch": 0.09309309309309309, "mask_bce_loss": 0.7244579195976257, "mask_dice_loss": 0.049519337713718414, "mask_loss": 0.7739772796630859, "step": 155 }, { "epoch": 0.0936936936936937, "grad_norm": 14.814277648925781, "learning_rate": 1.9792129476205267e-06, "loss": 0.7364766001701355, "step": 156 }, { "ce_loss": 2.7568836230784655e-05, "cls_loss": 0.039794921875, "epoch": 0.0936936936936937, "mask_bce_loss": 0.186964213848114, "mask_dice_loss": 0.03362220525741577, "mask_loss": 0.22058641910552979, "step": 156 }, { "epoch": 0.0942942942942943, "grad_norm": 35.13548278808594, "learning_rate": 1.978816528816984e-06, "loss": 0.9599290490150452, "step": 157 }, { "ce_loss": 3.946137803723104e-05, "cls_loss": 0.033447265625, "epoch": 0.0942942942942943, "mask_bce_loss": 0.4063805043697357, "mask_dice_loss": 0.024740802124142647, "mask_loss": 0.4311213195323944, "step": 157 }, { "epoch": 0.0948948948948949, "grad_norm": 35.9882698059082, "learning_rate": 1.9784164061424663e-06, "loss": 0.9573612809181213, "step": 158 }, { "ce_loss": 2.7915139071410522e-05, "cls_loss": 0.04638671875, "epoch": 0.0948948948948949, "mask_bce_loss": 0.4023536741733551, "mask_dice_loss": 0.045331746339797974, "mask_loss": 0.4476854205131531, "step": 158 }, { "epoch": 0.09549549549549549, "grad_norm": 31.07865333557129, "learning_rate": 1.978012581111049e-06, "loss": 0.8425774574279785, "step": 159 }, { "ce_loss": 0.00014026147255208343, "cls_loss": 0.037353515625, "epoch": 0.09549549549549549, "mask_bce_loss": 0.40935036540031433, "mask_dice_loss": 0.054884787648916245, "mask_loss": 0.4642351567745209, "step": 159 }, { "epoch": 0.0960960960960961, "grad_norm": 22.568147659301758, "learning_rate": 1.977605055250819e-06, "loss": 0.9094167351722717, "step": 160 }, { "ce_loss": 3.173911682097241e-05, "cls_loss": 0.057373046875, "epoch": 0.0960960960960961, "mask_bce_loss": 0.7235059142112732, "mask_dice_loss": 0.09501215815544128, "mask_loss": 0.8185180425643921, "step": 160 }, { "epoch": 0.0966966966966967, "grad_norm": 24.23436164855957, "learning_rate": 1.977193830103866e-06, "loss": 0.7962015271186829, "step": 161 }, { "ce_loss": 0.0004900170024484396, "cls_loss": 0.053466796875, "epoch": 0.0966966966966967, "mask_bce_loss": 0.8109734654426575, "mask_dice_loss": 0.06537149101495743, "mask_loss": 0.8763449788093567, "step": 161 }, { "epoch": 0.0972972972972973, "grad_norm": 83.92549133300781, "learning_rate": 1.9767789072262778e-06, "loss": 1.1465191841125488, "step": 162 }, { "ce_loss": 0.0169149711728096, "cls_loss": 0.05810546875, "epoch": 0.0972972972972973, "mask_bce_loss": 1.3563741445541382, "mask_dice_loss": 0.12156333774328232, "mask_loss": 1.4779374599456787, "step": 162 }, { "epoch": 0.09789789789789789, "grad_norm": 19.973968505859375, "learning_rate": 1.9763602881881355e-06, "loss": 0.8219994902610779, "step": 163 }, { "ce_loss": 4.2556956032058224e-05, "cls_loss": 0.0888671875, "epoch": 0.09789789789789789, "mask_bce_loss": 1.9208297729492188, "mask_dice_loss": 0.0761343464255333, "mask_loss": 1.9969640970230103, "step": 163 }, { "epoch": 0.0984984984984985, "grad_norm": 31.228954315185547, "learning_rate": 1.9759379745735064e-06, "loss": 1.0329885482788086, "step": 164 }, { "ce_loss": 7.132199971238151e-05, "cls_loss": 0.0673828125, "epoch": 0.0984984984984985, "mask_bce_loss": 0.38353583216667175, "mask_dice_loss": 0.10583440214395523, "mask_loss": 0.4893702268600464, "step": 164 }, { "epoch": 0.0990990990990991, "grad_norm": 20.48762321472168, "learning_rate": 1.9755119679804367e-06, "loss": 0.7633225321769714, "step": 165 }, { "ce_loss": 3.305119025753811e-05, "cls_loss": 0.031982421875, "epoch": 0.0990990990990991, "mask_bce_loss": 0.3299034535884857, "mask_dice_loss": 0.02559463120996952, "mask_loss": 0.3554980754852295, "step": 165 }, { "epoch": 0.0996996996996997, "grad_norm": 32.072303771972656, "learning_rate": 1.9750822700209486e-06, "loss": 0.9295129776000977, "step": 166 }, { "ce_loss": 2.280065746163018e-05, "cls_loss": 0.0361328125, "epoch": 0.0996996996996997, "mask_bce_loss": 0.2501488924026489, "mask_dice_loss": 0.027947694063186646, "mask_loss": 0.27809658646583557, "step": 166 }, { "epoch": 0.1003003003003003, "grad_norm": 45.00312042236328, "learning_rate": 1.9746488823210325e-06, "loss": 0.7872883081436157, "step": 167 }, { "ce_loss": 0.00021264530369080603, "cls_loss": 0.0791015625, "epoch": 0.1003003003003003, "mask_bce_loss": 2.152376890182495, "mask_dice_loss": 0.14645560085773468, "mask_loss": 2.298832416534424, "step": 167 }, { "epoch": 0.1009009009009009, "grad_norm": 22.464391708374023, "learning_rate": 1.9742118065206393e-06, "loss": 0.8729379177093506, "step": 168 }, { "ce_loss": 0.0004399017780087888, "cls_loss": 0.0830078125, "epoch": 0.1009009009009009, "mask_bce_loss": 0.940934419631958, "mask_dice_loss": 0.07672716677188873, "mask_loss": 1.0176615715026855, "step": 168 }, { "epoch": 0.1015015015015015, "grad_norm": 21.4300537109375, "learning_rate": 1.973771044273678e-06, "loss": 0.8873287439346313, "step": 169 }, { "ce_loss": 0.00012473725655581802, "cls_loss": 0.041015625, "epoch": 0.1015015015015015, "mask_bce_loss": 0.8629245162010193, "mask_dice_loss": 0.04814659431576729, "mask_loss": 0.9110711216926575, "step": 169 }, { "epoch": 0.1021021021021021, "grad_norm": 48.085487365722656, "learning_rate": 1.9733265972480058e-06, "loss": 1.0013819932937622, "step": 170 }, { "ce_loss": 0.03218424320220947, "cls_loss": 0.0634765625, "epoch": 0.1021021021021021, "mask_bce_loss": 1.3897206783294678, "mask_dice_loss": 0.1576126664876938, "mask_loss": 1.5473333597183228, "step": 170 }, { "epoch": 0.10270270270270271, "grad_norm": 38.19533920288086, "learning_rate": 1.9728784671254233e-06, "loss": 0.86553955078125, "step": 171 }, { "ce_loss": 4.94942105433438e-05, "cls_loss": 0.0546875, "epoch": 0.10270270270270271, "mask_bce_loss": 0.5811690092086792, "mask_dice_loss": 0.13456933200359344, "mask_loss": 0.7157383561134338, "step": 171 }, { "epoch": 0.1033033033033033, "grad_norm": 19.64702606201172, "learning_rate": 1.972426655601668e-06, "loss": 0.8088476657867432, "step": 172 }, { "ce_loss": 4.533532774075866e-05, "cls_loss": 0.052001953125, "epoch": 0.1033033033033033, "mask_bce_loss": 0.7507021427154541, "mask_dice_loss": 0.07279609888792038, "mask_loss": 0.8234982490539551, "step": 172 }, { "epoch": 0.1039039039039039, "grad_norm": 68.59932708740234, "learning_rate": 1.9719711643864087e-06, "loss": 1.1291075944900513, "step": 173 }, { "ce_loss": 2.591317388578318e-05, "cls_loss": 0.052734375, "epoch": 0.1039039039039039, "mask_bce_loss": 0.8106512427330017, "mask_dice_loss": 0.06875898689031601, "mask_loss": 0.8794102072715759, "step": 173 }, { "epoch": 0.1045045045045045, "grad_norm": 28.556447982788086, "learning_rate": 1.971511995203238e-06, "loss": 0.9481885433197021, "step": 174 }, { "ce_loss": 0.004828756209462881, "cls_loss": 0.044921875, "epoch": 0.1045045045045045, "mask_bce_loss": 0.6180534958839417, "mask_dice_loss": 0.05888466164469719, "mask_loss": 0.6769381761550903, "step": 174 }, { "epoch": 0.10510510510510511, "grad_norm": 46.14301300048828, "learning_rate": 1.971049149789665e-06, "loss": 0.9197049140930176, "step": 175 }, { "ce_loss": 1.528377651993651e-05, "cls_loss": 0.026123046875, "epoch": 0.10510510510510511, "mask_bce_loss": 0.2427414208650589, "mask_dice_loss": 0.04080522805452347, "mask_loss": 0.28354665637016296, "step": 175 }, { "epoch": 0.1057057057057057, "grad_norm": 28.797109603881836, "learning_rate": 1.970582629897111e-06, "loss": 0.8465726375579834, "step": 176 }, { "ce_loss": 6.306828436208889e-05, "cls_loss": 0.05078125, "epoch": 0.1057057057057057, "mask_bce_loss": 0.5317959785461426, "mask_dice_loss": 0.06949187070131302, "mask_loss": 0.601287841796875, "step": 176 }, { "epoch": 0.1063063063063063, "grad_norm": 20.875246047973633, "learning_rate": 1.970112437290901e-06, "loss": 0.8030381798744202, "step": 177 }, { "ce_loss": 3.0213606805773452e-05, "cls_loss": 0.061279296875, "epoch": 0.1063063063063063, "mask_bce_loss": 1.1507494449615479, "mask_dice_loss": 0.1427956372499466, "mask_loss": 1.293545126914978, "step": 177 }, { "epoch": 0.1069069069069069, "grad_norm": 35.47929000854492, "learning_rate": 1.9696385737502584e-06, "loss": 0.861383318901062, "step": 178 }, { "ce_loss": 5.300628254190087e-05, "cls_loss": 0.039794921875, "epoch": 0.1069069069069069, "mask_bce_loss": 0.31387433409690857, "mask_dice_loss": 0.03750455006957054, "mask_loss": 0.3513788878917694, "step": 178 }, { "epoch": 0.10750750750750751, "grad_norm": 22.84060287475586, "learning_rate": 1.9691610410682964e-06, "loss": 0.8383473753929138, "step": 179 }, { "ce_loss": 2.2298969270195812e-05, "cls_loss": 0.05712890625, "epoch": 0.10750750750750751, "mask_bce_loss": 1.0615142583847046, "mask_dice_loss": 0.11658895015716553, "mask_loss": 1.1781032085418701, "step": 179 }, { "epoch": 0.10810810810810811, "grad_norm": 22.836769104003906, "learning_rate": 1.9686798410520133e-06, "loss": 0.7232077121734619, "step": 180 }, { "ce_loss": 3.685730916913599e-05, "cls_loss": 0.041748046875, "epoch": 0.10810810810810811, "mask_bce_loss": 0.6132872104644775, "mask_dice_loss": 0.06623297184705734, "mask_loss": 0.6795201897621155, "step": 180 }, { "epoch": 0.1087087087087087, "grad_norm": 16.986297607421875, "learning_rate": 1.968194975522284e-06, "loss": 0.7880321145057678, "step": 181 }, { "ce_loss": 0.01039371732622385, "cls_loss": 0.05078125, "epoch": 0.1087087087087087, "mask_bce_loss": 0.47914955019950867, "mask_dice_loss": 0.0850035771727562, "mask_loss": 0.5641531348228455, "step": 181 }, { "epoch": 0.10930930930930931, "grad_norm": 29.967737197875977, "learning_rate": 1.9677064463138546e-06, "loss": 0.8752521872520447, "step": 182 }, { "ce_loss": 2.2061116396798752e-05, "cls_loss": 0.048095703125, "epoch": 0.10930930930930931, "mask_bce_loss": 0.3885654807090759, "mask_dice_loss": 0.05138930678367615, "mask_loss": 0.4399547874927521, "step": 182 }, { "epoch": 0.10990990990990991, "grad_norm": 27.597322463989258, "learning_rate": 1.967214255275334e-06, "loss": 0.8572582602500916, "step": 183 }, { "ce_loss": 0.00697388406842947, "cls_loss": 0.0546875, "epoch": 0.10990990990990991, "mask_bce_loss": 0.32455742359161377, "mask_dice_loss": 0.07211451232433319, "mask_loss": 0.39667195081710815, "step": 183 }, { "epoch": 0.11051051051051052, "grad_norm": 30.982982635498047, "learning_rate": 1.9667184042691875e-06, "loss": 0.7828026413917542, "step": 184 }, { "ce_loss": 3.855912291328423e-05, "cls_loss": 0.06982421875, "epoch": 0.11051051051051052, "mask_bce_loss": 1.7740188837051392, "mask_dice_loss": 0.13409629464149475, "mask_loss": 1.9081151485443115, "step": 184 }, { "epoch": 0.1111111111111111, "grad_norm": 13.643583297729492, "learning_rate": 1.9662188951717304e-06, "loss": 0.760838508605957, "step": 185 }, { "ce_loss": 0.00013058654440101236, "cls_loss": 0.04150390625, "epoch": 0.1111111111111111, "mask_bce_loss": 1.9294099807739258, "mask_dice_loss": 0.06974196434020996, "mask_loss": 1.9991519451141357, "step": 185 }, { "epoch": 0.11171171171171171, "grad_norm": 15.644401550292969, "learning_rate": 1.96571572987312e-06, "loss": 1.0069599151611328, "step": 186 }, { "ce_loss": 3.6360917874844745e-05, "cls_loss": 0.04443359375, "epoch": 0.11171171171171171, "mask_bce_loss": 1.0131648778915405, "mask_dice_loss": 0.08159618824720383, "mask_loss": 1.0947610139846802, "step": 186 }, { "epoch": 0.11231231231231231, "grad_norm": 18.968067169189453, "learning_rate": 1.9652089102773484e-06, "loss": 0.9087978005409241, "step": 187 }, { "ce_loss": 3.065450073336251e-05, "cls_loss": 0.03955078125, "epoch": 0.11231231231231231, "mask_bce_loss": 0.5119089484214783, "mask_dice_loss": 0.04586147889494896, "mask_loss": 0.5577704310417175, "step": 187 }, { "epoch": 0.11291291291291292, "grad_norm": 22.550085067749023, "learning_rate": 1.964698438302237e-06, "loss": 0.8600544929504395, "step": 188 }, { "ce_loss": 0.00014770872076041996, "cls_loss": 0.0289306640625, "epoch": 0.11291291291291292, "mask_bce_loss": 0.39572229981422424, "mask_dice_loss": 0.03381771966814995, "mask_loss": 0.4295400083065033, "step": 188 }, { "epoch": 0.11351351351351352, "grad_norm": 37.752925872802734, "learning_rate": 1.9641843158794266e-06, "loss": 1.0281672477722168, "step": 189 }, { "ce_loss": 5.897740993532352e-05, "cls_loss": 0.03466796875, "epoch": 0.11351351351351352, "mask_bce_loss": 0.591586172580719, "mask_dice_loss": 0.06317750364542007, "mask_loss": 0.6547636985778809, "step": 189 }, { "epoch": 0.11411411411411411, "grad_norm": 31.1929931640625, "learning_rate": 1.963666544954371e-06, "loss": 0.9628913998603821, "step": 190 }, { "ce_loss": 9.215588943334296e-05, "cls_loss": 0.05322265625, "epoch": 0.11411411411411411, "mask_bce_loss": 0.6208127737045288, "mask_dice_loss": 0.09192865341901779, "mask_loss": 0.7127414345741272, "step": 190 }, { "epoch": 0.11471471471471471, "grad_norm": 21.984169006347656, "learning_rate": 1.9631451274863322e-06, "loss": 0.7893231511116028, "step": 191 }, { "ce_loss": 0.000629036221653223, "cls_loss": 0.06298828125, "epoch": 0.11471471471471471, "mask_bce_loss": 0.7059471011161804, "mask_dice_loss": 0.09597029536962509, "mask_loss": 0.8019173741340637, "step": 191 }, { "epoch": 0.11531531531531532, "grad_norm": 63.757354736328125, "learning_rate": 1.9626200654483687e-06, "loss": 0.9295555949211121, "step": 192 }, { "ce_loss": 0.0001597110094735399, "cls_loss": 0.053955078125, "epoch": 0.11531531531531532, "mask_bce_loss": 1.0383596420288086, "mask_dice_loss": 0.0709841400384903, "mask_loss": 1.1093437671661377, "step": 192 }, { "epoch": 0.11591591591591592, "grad_norm": 15.326539993286133, "learning_rate": 1.9620913608273308e-06, "loss": 0.7506568431854248, "step": 193 }, { "ce_loss": 0.0001735159457894042, "cls_loss": 0.07568359375, "epoch": 0.11591591591591592, "mask_bce_loss": 0.986118733882904, "mask_dice_loss": 0.15347474813461304, "mask_loss": 1.139593482017517, "step": 193 }, { "epoch": 0.11651651651651651, "grad_norm": 36.344444274902344, "learning_rate": 1.961559015623853e-06, "loss": 0.867622971534729, "step": 194 }, { "ce_loss": 6.069532173569314e-05, "cls_loss": 0.0361328125, "epoch": 0.11651651651651651, "mask_bce_loss": 0.31585821509361267, "mask_dice_loss": 0.033091798424720764, "mask_loss": 0.34895002841949463, "step": 194 }, { "epoch": 0.11711711711711711, "grad_norm": 27.540212631225586, "learning_rate": 1.9610230318523453e-06, "loss": 0.7941890954971313, "step": 195 }, { "ce_loss": 2.0887087885057554e-05, "cls_loss": 0.06884765625, "epoch": 0.11711711711711711, "mask_bce_loss": 0.98511803150177, "mask_dice_loss": 0.09392394870519638, "mask_loss": 1.0790419578552246, "step": 195 }, { "epoch": 0.11771771771771772, "grad_norm": 34.90920639038086, "learning_rate": 1.9604834115409857e-06, "loss": 1.0372711420059204, "step": 196 }, { "ce_loss": 0.0003565413935575634, "cls_loss": 0.06005859375, "epoch": 0.11771771771771772, "mask_bce_loss": 0.575587272644043, "mask_dice_loss": 0.09374924749135971, "mask_loss": 0.6693364977836609, "step": 196 }, { "epoch": 0.11831831831831832, "grad_norm": 15.079602241516113, "learning_rate": 1.9599401567317145e-06, "loss": 0.7465895414352417, "step": 197 }, { "ce_loss": 0.0001672690996201709, "cls_loss": 0.08203125, "epoch": 0.11831831831831832, "mask_bce_loss": 1.5179177522659302, "mask_dice_loss": 0.0822213888168335, "mask_loss": 1.6001391410827637, "step": 197 }, { "epoch": 0.11891891891891893, "grad_norm": 11.710132598876953, "learning_rate": 1.959393269480224e-06, "loss": 0.6801038384437561, "step": 198 }, { "ce_loss": 2.2922777134226635e-05, "cls_loss": 0.0478515625, "epoch": 0.11891891891891893, "mask_bce_loss": 0.8387079238891602, "mask_dice_loss": 0.05720863491296768, "mask_loss": 0.8959165811538696, "step": 198 }, { "epoch": 0.11951951951951952, "grad_norm": 62.096893310546875, "learning_rate": 1.9588427518559513e-06, "loss": 0.918253481388092, "step": 199 }, { "ce_loss": 4.73442196380347e-05, "cls_loss": 0.037109375, "epoch": 0.11951951951951952, "mask_bce_loss": 0.4613240659236908, "mask_dice_loss": 0.03127896785736084, "mask_loss": 0.49260303378105164, "step": 199 }, { "epoch": 0.12012012012012012, "grad_norm": 28.948488235473633, "learning_rate": 1.958288605942072e-06, "loss": 0.8748531341552734, "step": 200 }, { "ce_loss": 4.349397568148561e-05, "cls_loss": 0.057861328125, "epoch": 0.12012012012012012, "mask_bce_loss": 0.4998346269130707, "mask_dice_loss": 0.0901232361793518, "mask_loss": 0.5899578332901001, "step": 200 }, { "epoch": 0.12072072072072072, "grad_norm": 17.91237449645996, "learning_rate": 1.9577308338354907e-06, "loss": 0.8979141712188721, "step": 201 }, { "ce_loss": 3.732255572685972e-05, "cls_loss": 0.06103515625, "epoch": 0.12072072072072072, "mask_bce_loss": 0.5373799800872803, "mask_dice_loss": 0.10608190298080444, "mask_loss": 0.6434618830680847, "step": 201 }, { "epoch": 0.12132132132132133, "grad_norm": 24.002965927124023, "learning_rate": 1.957169437646834e-06, "loss": 0.7035423517227173, "step": 202 }, { "ce_loss": 0.0024443164002150297, "cls_loss": 0.21875, "epoch": 0.12132132132132133, "mask_bce_loss": 0.6442286372184753, "mask_dice_loss": 0.11617112904787064, "mask_loss": 0.7603997588157654, "step": 202 }, { "epoch": 0.12192192192192192, "grad_norm": 17.161834716796875, "learning_rate": 1.9566044195004407e-06, "loss": 0.8427568078041077, "step": 203 }, { "ce_loss": 3.751779513549991e-05, "cls_loss": 0.041015625, "epoch": 0.12192192192192192, "mask_bce_loss": 0.7424851059913635, "mask_dice_loss": 0.05747232586145401, "mask_loss": 0.7999574542045593, "step": 203 }, { "epoch": 0.12252252252252252, "grad_norm": 26.98950958251953, "learning_rate": 1.9560357815343576e-06, "loss": 0.8652293682098389, "step": 204 }, { "ce_loss": 0.00010234179353574291, "cls_loss": 0.0419921875, "epoch": 0.12252252252252252, "mask_bce_loss": 0.5991503000259399, "mask_dice_loss": 0.05552471801638603, "mask_loss": 0.6546750068664551, "step": 204 }, { "epoch": 0.12312312312312312, "grad_norm": 18.781618118286133, "learning_rate": 1.9554635259003273e-06, "loss": 0.9370501041412354, "step": 205 }, { "ce_loss": 4.9530841351952404e-05, "cls_loss": 0.057861328125, "epoch": 0.12312312312312312, "mask_bce_loss": 0.5390082597732544, "mask_dice_loss": 0.08636614680290222, "mask_loss": 0.625374436378479, "step": 205 }, { "epoch": 0.12372372372372373, "grad_norm": 29.53583335876465, "learning_rate": 1.954887654763782e-06, "loss": 0.8825756907463074, "step": 206 }, { "ce_loss": 5.555451934924349e-05, "cls_loss": 0.05859375, "epoch": 0.12372372372372373, "mask_bce_loss": 0.6065228581428528, "mask_dice_loss": 0.0725567489862442, "mask_loss": 0.6790795922279358, "step": 206 }, { "epoch": 0.12432432432432433, "grad_norm": 18.15776824951172, "learning_rate": 1.9543081703038357e-06, "loss": 0.7043869495391846, "step": 207 }, { "ce_loss": 3.910524173988961e-05, "cls_loss": 0.055908203125, "epoch": 0.12432432432432433, "mask_bce_loss": 0.9227821230888367, "mask_dice_loss": 0.08570069074630737, "mask_loss": 1.008482813835144, "step": 207 }, { "epoch": 0.12492492492492492, "grad_norm": 28.508651733398438, "learning_rate": 1.9537250747132747e-06, "loss": 0.926513671875, "step": 208 }, { "ce_loss": 6.926255446160212e-05, "cls_loss": 0.05029296875, "epoch": 0.12492492492492492, "mask_bce_loss": 0.5864166021347046, "mask_dice_loss": 0.10338306427001953, "mask_loss": 0.6897996664047241, "step": 208 }, { "epoch": 0.12552552552552554, "grad_norm": 20.62690544128418, "learning_rate": 1.9531383701985497e-06, "loss": 0.8640689849853516, "step": 209 }, { "ce_loss": 3.3239033655263484e-05, "cls_loss": 0.03955078125, "epoch": 0.12552552552552554, "mask_bce_loss": 0.37545037269592285, "mask_dice_loss": 0.061974406242370605, "mask_loss": 0.43742477893829346, "step": 209 }, { "epoch": 0.12612612612612611, "grad_norm": 19.1878662109375, "learning_rate": 1.952548058979769e-06, "loss": 0.8493676781654358, "step": 210 }, { "ce_loss": 3.109568206127733e-05, "cls_loss": 0.04443359375, "epoch": 0.12612612612612611, "mask_bce_loss": 0.5511075854301453, "mask_dice_loss": 0.04577948898077011, "mask_loss": 0.5968870520591736, "step": 210 }, { "epoch": 0.12672672672672672, "grad_norm": 39.66432189941406, "learning_rate": 1.9519541432906868e-06, "loss": 0.9235525131225586, "step": 211 }, { "ce_loss": 5.231994509813376e-05, "cls_loss": 0.04248046875, "epoch": 0.12672672672672672, "mask_bce_loss": 0.5986132025718689, "mask_dice_loss": 0.047025054693222046, "mask_loss": 0.6456382274627686, "step": 211 }, { "epoch": 0.12732732732732732, "grad_norm": 18.680267333984375, "learning_rate": 1.951356625378699e-06, "loss": 0.8163272738456726, "step": 212 }, { "ce_loss": 2.447086626489181e-05, "cls_loss": 0.0654296875, "epoch": 0.12732732732732732, "mask_bce_loss": 0.45880326628685, "mask_dice_loss": 0.05754583328962326, "mask_loss": 0.5163490772247314, "step": 212 }, { "epoch": 0.12792792792792793, "grad_norm": 31.946556091308594, "learning_rate": 1.9507555075048305e-06, "loss": 1.0470069646835327, "step": 213 }, { "ce_loss": 4.275753963156603e-05, "cls_loss": 0.09326171875, "epoch": 0.12792792792792793, "mask_bce_loss": 0.5375936627388, "mask_dice_loss": 0.038787126541137695, "mask_loss": 0.5763807892799377, "step": 213 }, { "epoch": 0.12852852852852853, "grad_norm": 26.326919555664062, "learning_rate": 1.9501507919437293e-06, "loss": 0.8956798315048218, "step": 214 }, { "ce_loss": 5.8588200772646815e-05, "cls_loss": 0.03369140625, "epoch": 0.12852852852852853, "mask_bce_loss": 0.665438711643219, "mask_dice_loss": 0.05073515325784683, "mask_loss": 0.7161738872528076, "step": 214 }, { "epoch": 0.12912912912912913, "grad_norm": 25.804264068603516, "learning_rate": 1.9495424809836573e-06, "loss": 0.9088098406791687, "step": 215 }, { "ce_loss": 7.599771925015375e-05, "cls_loss": 0.05712890625, "epoch": 0.12912912912912913, "mask_bce_loss": 0.911206066608429, "mask_dice_loss": 0.08364228904247284, "mask_loss": 0.994848370552063, "step": 215 }, { "epoch": 0.12972972972972974, "grad_norm": 27.66560173034668, "learning_rate": 1.948930576926481e-06, "loss": 0.8578746318817139, "step": 216 }, { "ce_loss": 2.7249214326729998e-05, "cls_loss": 0.0625, "epoch": 0.12972972972972974, "mask_bce_loss": 1.342075228691101, "mask_dice_loss": 0.1430782675743103, "mask_loss": 1.4851534366607666, "step": 216 }, { "epoch": 0.13033033033033034, "grad_norm": 22.986801147460938, "learning_rate": 1.948315082087664e-06, "loss": 0.9379733800888062, "step": 217 }, { "ce_loss": 0.000300725398119539, "cls_loss": 0.050048828125, "epoch": 0.13033033033033034, "mask_bce_loss": 0.3219219148159027, "mask_dice_loss": 0.06190449744462967, "mask_loss": 0.3838264048099518, "step": 217 }, { "epoch": 0.13093093093093094, "grad_norm": 17.082033157348633, "learning_rate": 1.9476959987962572e-06, "loss": 0.8603769540786743, "step": 218 }, { "ce_loss": 7.32607368263416e-05, "cls_loss": 0.036865234375, "epoch": 0.13093093093093094, "mask_bce_loss": 0.47028347849845886, "mask_dice_loss": 0.035090357065200806, "mask_loss": 0.5053738355636597, "step": 218 }, { "epoch": 0.13153153153153152, "grad_norm": 31.05573844909668, "learning_rate": 1.9470733293948902e-06, "loss": 0.8510897159576416, "step": 219 }, { "ce_loss": 2.5348015697090887e-05, "cls_loss": 0.0390625, "epoch": 0.13153153153153152, "mask_bce_loss": 0.47728973627090454, "mask_dice_loss": 0.042267508804798126, "mask_loss": 0.5195572376251221, "step": 219 }, { "epoch": 0.13213213213213212, "grad_norm": 19.774009704589844, "learning_rate": 1.9464470762397626e-06, "loss": 0.7960196137428284, "step": 220 }, { "ce_loss": 0.0005569122731685638, "cls_loss": 0.036865234375, "epoch": 0.13213213213213212, "mask_bce_loss": 0.6020805239677429, "mask_dice_loss": 0.04602283984422684, "mask_loss": 0.6481033563613892, "step": 220 }, { "epoch": 0.13273273273273273, "grad_norm": 30.163738250732422, "learning_rate": 1.9458172417006346e-06, "loss": 0.9146794080734253, "step": 221 }, { "ce_loss": 0.00011975997767876834, "cls_loss": 0.031494140625, "epoch": 0.13273273273273273, "mask_bce_loss": 0.3289167881011963, "mask_dice_loss": 0.04603268951177597, "mask_loss": 0.37494948506355286, "step": 221 }, { "epoch": 0.13333333333333333, "grad_norm": 35.52772521972656, "learning_rate": 1.9451838281608196e-06, "loss": 0.7934065461158752, "step": 222 }, { "ce_loss": 4.126750354771502e-05, "cls_loss": 0.0546875, "epoch": 0.13333333333333333, "mask_bce_loss": 0.8925997018814087, "mask_dice_loss": 0.11733143776655197, "mask_loss": 1.0099310874938965, "step": 222 }, { "epoch": 0.13393393393393394, "grad_norm": 30.455989837646484, "learning_rate": 1.944546838017173e-06, "loss": 0.8957576751708984, "step": 223 }, { "ce_loss": 3.6687513784272596e-05, "cls_loss": 0.044921875, "epoch": 0.13393393393393394, "mask_bce_loss": 0.1620636284351349, "mask_dice_loss": 0.026670334860682487, "mask_loss": 0.18873396515846252, "step": 223 }, { "epoch": 0.13453453453453454, "grad_norm": 20.064407348632812, "learning_rate": 1.9439062736800844e-06, "loss": 0.8580139875411987, "step": 224 }, { "ce_loss": 7.188644667621702e-05, "cls_loss": 0.03662109375, "epoch": 0.13453453453453454, "mask_bce_loss": 0.5552011132240295, "mask_dice_loss": 0.06320410221815109, "mask_loss": 0.6184052228927612, "step": 224 }, { "epoch": 0.13513513513513514, "grad_norm": 34.267398834228516, "learning_rate": 1.9432621375734683e-06, "loss": 0.974747359752655, "step": 225 }, { "ce_loss": 2.6873216484091245e-05, "cls_loss": 0.05419921875, "epoch": 0.13513513513513514, "mask_bce_loss": 0.8653661012649536, "mask_dice_loss": 0.08427553623914719, "mask_loss": 0.9496416449546814, "step": 225 }, { "epoch": 0.13573573573573575, "grad_norm": 23.032363891601562, "learning_rate": 1.9426144321347554e-06, "loss": 0.8152979612350464, "step": 226 }, { "ce_loss": 0.03419021517038345, "cls_loss": 0.03369140625, "epoch": 0.13573573573573575, "mask_bce_loss": 0.254309743642807, "mask_dice_loss": 0.02356691285967827, "mask_loss": 0.2778766453266144, "step": 226 }, { "epoch": 0.13633633633633635, "grad_norm": 21.60953140258789, "learning_rate": 1.9419631598148826e-06, "loss": 0.8950212001800537, "step": 227 }, { "ce_loss": 6.233329622773454e-05, "cls_loss": 0.057861328125, "epoch": 0.13633633633633635, "mask_bce_loss": 0.7484661340713501, "mask_dice_loss": 0.09682205319404602, "mask_loss": 0.8452881574630737, "step": 227 }, { "epoch": 0.13693693693693693, "grad_norm": 21.017375946044922, "learning_rate": 1.941308323078283e-06, "loss": 1.0351446866989136, "step": 228 }, { "ce_loss": 0.00024846766609698534, "cls_loss": 0.055908203125, "epoch": 0.13693693693693693, "mask_bce_loss": 2.0128281116485596, "mask_dice_loss": 0.20649385452270508, "mask_loss": 2.2193219661712646, "step": 228 }, { "epoch": 0.13753753753753753, "grad_norm": 32.161277770996094, "learning_rate": 1.940649924402879e-06, "loss": 0.8755577802658081, "step": 229 }, { "ce_loss": 0.0003055866400245577, "cls_loss": 0.0400390625, "epoch": 0.13753753753753753, "mask_bce_loss": 0.2639738619327545, "mask_dice_loss": 0.03750677406787872, "mask_loss": 0.30148065090179443, "step": 229 }, { "epoch": 0.13813813813813813, "grad_norm": 17.70762825012207, "learning_rate": 1.93998796628007e-06, "loss": 0.8269757032394409, "step": 230 }, { "ce_loss": 6.107806257205084e-05, "cls_loss": 0.050048828125, "epoch": 0.13813813813813813, "mask_bce_loss": 1.0210449695587158, "mask_dice_loss": 0.14977166056632996, "mask_loss": 1.1708166599273682, "step": 230 }, { "epoch": 0.13873873873873874, "grad_norm": 24.088542938232422, "learning_rate": 1.9393224512147267e-06, "loss": 0.7005638480186462, "step": 231 }, { "ce_loss": 9.730947203934193e-05, "cls_loss": 0.05908203125, "epoch": 0.13873873873873874, "mask_bce_loss": 0.8308659791946411, "mask_dice_loss": 0.09232459962368011, "mask_loss": 0.9231905937194824, "step": 231 }, { "epoch": 0.13933933933933934, "grad_norm": 27.45137596130371, "learning_rate": 1.938653381725177e-06, "loss": 0.8291493058204651, "step": 232 }, { "ce_loss": 3.195125100319274e-05, "cls_loss": 0.03173828125, "epoch": 0.13933933933933934, "mask_bce_loss": 0.3097558915615082, "mask_dice_loss": 0.025664672255516052, "mask_loss": 0.33542054891586304, "step": 232 }, { "epoch": 0.13993993993993994, "grad_norm": 29.489028930664062, "learning_rate": 1.9379807603432e-06, "loss": 0.8972260355949402, "step": 233 }, { "ce_loss": 3.34874821419362e-05, "cls_loss": 0.053955078125, "epoch": 0.13993993993993994, "mask_bce_loss": 0.7459731101989746, "mask_dice_loss": 0.16206274926662445, "mask_loss": 0.9080358743667603, "step": 233 }, { "epoch": 0.14054054054054055, "grad_norm": 40.03779983520508, "learning_rate": 1.937304589614015e-06, "loss": 0.9146872758865356, "step": 234 }, { "ce_loss": 3.124958311673254e-05, "cls_loss": 0.0634765625, "epoch": 0.14054054054054055, "mask_bce_loss": 0.9501029253005981, "mask_dice_loss": 0.1083638072013855, "mask_loss": 1.0584666728973389, "step": 234 }, { "epoch": 0.14114114114114115, "grad_norm": 19.248842239379883, "learning_rate": 1.9366248720962727e-06, "loss": 0.8182445764541626, "step": 235 }, { "ce_loss": 7.267509499797598e-05, "cls_loss": 0.057861328125, "epoch": 0.14114114114114115, "mask_bce_loss": 1.337107539176941, "mask_dice_loss": 0.10082607716321945, "mask_loss": 1.4379335641860962, "step": 235 }, { "epoch": 0.14174174174174173, "grad_norm": 21.447141647338867, "learning_rate": 1.9359416103620442e-06, "loss": 0.7994502782821655, "step": 236 }, { "ce_loss": 2.957335891551338e-05, "cls_loss": 0.052001953125, "epoch": 0.14174174174174173, "mask_bce_loss": 0.8302595019340515, "mask_dice_loss": 0.08345561474561691, "mask_loss": 0.913715124130249, "step": 236 }, { "epoch": 0.14234234234234233, "grad_norm": 12.912201881408691, "learning_rate": 1.935254806996812e-06, "loss": 0.680029034614563, "step": 237 }, { "ce_loss": 2.780376780719962e-05, "cls_loss": 0.055908203125, "epoch": 0.14234234234234233, "mask_bce_loss": 1.3360451459884644, "mask_dice_loss": 0.08958058804273605, "mask_loss": 1.4256256818771362, "step": 237 }, { "epoch": 0.14294294294294294, "grad_norm": 22.342559814453125, "learning_rate": 1.934564464599461e-06, "loss": 0.8215990662574768, "step": 238 }, { "ce_loss": 0.0002702339261304587, "cls_loss": 0.049072265625, "epoch": 0.14294294294294294, "mask_bce_loss": 0.6481857299804688, "mask_dice_loss": 0.11833363026380539, "mask_loss": 0.7665193676948547, "step": 238 }, { "epoch": 0.14354354354354354, "grad_norm": 25.060258865356445, "learning_rate": 1.933870585782267e-06, "loss": 0.9159284234046936, "step": 239 }, { "ce_loss": 3.7369907659012824e-05, "cls_loss": 0.0771484375, "epoch": 0.14354354354354354, "mask_bce_loss": 0.2103029042482376, "mask_dice_loss": 0.02577810548245907, "mask_loss": 0.23608100414276123, "step": 239 }, { "epoch": 0.14414414414414414, "grad_norm": 36.98003005981445, "learning_rate": 1.933173173170889e-06, "loss": 1.0006487369537354, "step": 240 }, { "ce_loss": 0.0002012750192079693, "cls_loss": 0.08056640625, "epoch": 0.14414414414414414, "mask_bce_loss": 0.6725789308547974, "mask_dice_loss": 0.09667545557022095, "mask_loss": 0.7692543864250183, "step": 240 }, { "epoch": 0.14474474474474475, "grad_norm": 25.7158260345459, "learning_rate": 1.932472229404356e-06, "loss": 0.8864552974700928, "step": 241 }, { "ce_loss": 3.469392322585918e-05, "cls_loss": 0.0230712890625, "epoch": 0.14474474474474475, "mask_bce_loss": 0.18123403191566467, "mask_dice_loss": 0.01471267081797123, "mask_loss": 0.19594670832157135, "step": 241 }, { "epoch": 0.14534534534534535, "grad_norm": 23.612043380737305, "learning_rate": 1.9317677571350604e-06, "loss": 0.8931270837783813, "step": 242 }, { "ce_loss": 3.79109442292247e-05, "cls_loss": 0.061767578125, "epoch": 0.14534534534534535, "mask_bce_loss": 0.5304180979728699, "mask_dice_loss": 0.1626545786857605, "mask_loss": 0.6930726766586304, "step": 242 }, { "epoch": 0.14594594594594595, "grad_norm": 67.47167205810547, "learning_rate": 1.9310597590287473e-06, "loss": 0.9238914847373962, "step": 243 }, { "ce_loss": 0.0003489031223580241, "cls_loss": 0.06494140625, "epoch": 0.14594594594594595, "mask_bce_loss": 0.5016066431999207, "mask_dice_loss": 0.10651472955942154, "mask_loss": 0.608121395111084, "step": 243 }, { "epoch": 0.14654654654654656, "grad_norm": 28.51106071472168, "learning_rate": 1.930348237764502e-06, "loss": 0.834000289440155, "step": 244 }, { "ce_loss": 2.2970796635490842e-05, "cls_loss": 0.05712890625, "epoch": 0.14654654654654656, "mask_bce_loss": 1.3702259063720703, "mask_dice_loss": 0.18668578565120697, "mask_loss": 1.5569117069244385, "step": 244 }, { "epoch": 0.14714714714714713, "grad_norm": 29.898067474365234, "learning_rate": 1.929633196034742e-06, "loss": 0.8673421144485474, "step": 245 }, { "ce_loss": 0.00023982718994375318, "cls_loss": 0.09326171875, "epoch": 0.14714714714714713, "mask_bce_loss": 0.7706180810928345, "mask_dice_loss": 0.0817875862121582, "mask_loss": 0.8524056673049927, "step": 245 }, { "epoch": 0.14774774774774774, "grad_norm": 25.38163185119629, "learning_rate": 1.9289146365452067e-06, "loss": 0.9079372882843018, "step": 246 }, { "ce_loss": 3.752532575163059e-05, "cls_loss": 0.043212890625, "epoch": 0.14774774774774774, "mask_bce_loss": 0.3776664137840271, "mask_dice_loss": 0.03545583412051201, "mask_loss": 0.4131222367286682, "step": 246 }, { "epoch": 0.14834834834834834, "grad_norm": 31.87659454345703, "learning_rate": 1.9281925620149467e-06, "loss": 0.956485390663147, "step": 247 }, { "ce_loss": 4.6820488933008164e-05, "cls_loss": 0.05908203125, "epoch": 0.14834834834834834, "mask_bce_loss": 1.009015679359436, "mask_dice_loss": 0.07725010067224503, "mask_loss": 1.0862658023834229, "step": 247 }, { "epoch": 0.14894894894894894, "grad_norm": 19.2994384765625, "learning_rate": 1.927466975176314e-06, "loss": 0.7731269001960754, "step": 248 }, { "ce_loss": 7.497208571294323e-05, "cls_loss": 0.06103515625, "epoch": 0.14894894894894894, "mask_bce_loss": 0.8538575172424316, "mask_dice_loss": 0.13585524260997772, "mask_loss": 0.9897127747535706, "step": 248 }, { "epoch": 0.14954954954954955, "grad_norm": 23.666311264038086, "learning_rate": 1.92673787877495e-06, "loss": 0.7632424831390381, "step": 249 }, { "ce_loss": 2.8798838684451766e-05, "cls_loss": 0.042236328125, "epoch": 0.14954954954954955, "mask_bce_loss": 0.2841022312641144, "mask_dice_loss": 0.03679477050900459, "mask_loss": 0.32089701294898987, "step": 249 }, { "epoch": 0.15015015015015015, "grad_norm": 18.034526824951172, "learning_rate": 1.9260052755697783e-06, "loss": 0.7658596038818359, "step": 250 }, { "ce_loss": 0.005658265203237534, "cls_loss": 0.06396484375, "epoch": 0.15015015015015015, "mask_bce_loss": 0.4604777991771698, "mask_dice_loss": 0.12151086330413818, "mask_loss": 0.5819886922836304, "step": 250 }, { "epoch": 0.15075075075075076, "grad_norm": 20.383708953857422, "learning_rate": 1.9252691683329902e-06, "loss": 0.7879098057746887, "step": 251 }, { "ce_loss": 7.926462421892211e-05, "cls_loss": 0.05322265625, "epoch": 0.15075075075075076, "mask_bce_loss": 0.9265584945678711, "mask_dice_loss": 0.08352448046207428, "mask_loss": 1.0100829601287842, "step": 251 }, { "epoch": 0.15135135135135136, "grad_norm": 30.465890884399414, "learning_rate": 1.9245295598500384e-06, "loss": 0.9258317351341248, "step": 252 }, { "ce_loss": 6.306808063527569e-05, "cls_loss": 0.057373046875, "epoch": 0.15135135135135136, "mask_bce_loss": 0.6243155598640442, "mask_dice_loss": 0.07737918943166733, "mask_loss": 0.7016947269439697, "step": 252 }, { "epoch": 0.15195195195195196, "grad_norm": 27.152448654174805, "learning_rate": 1.923786452919623e-06, "loss": 1.020339012145996, "step": 253 }, { "ce_loss": 2.6582751161186025e-05, "cls_loss": 0.05322265625, "epoch": 0.15195195195195196, "mask_bce_loss": 1.4171096086502075, "mask_dice_loss": 0.1805235743522644, "mask_loss": 1.5976331233978271, "step": 253 }, { "epoch": 0.15255255255255254, "grad_norm": 17.709903717041016, "learning_rate": 1.9230398503536832e-06, "loss": 0.8702489137649536, "step": 254 }, { "ce_loss": 2.8034744900651276e-05, "cls_loss": 0.059326171875, "epoch": 0.15255255255255254, "mask_bce_loss": 0.6503230929374695, "mask_dice_loss": 0.11137622594833374, "mask_loss": 0.7616993188858032, "step": 254 }, { "epoch": 0.15315315315315314, "grad_norm": 27.547117233276367, "learning_rate": 1.9222897549773846e-06, "loss": 0.957504153251648, "step": 255 }, { "ce_loss": 5.985750976833515e-05, "cls_loss": 0.05712890625, "epoch": 0.15315315315315314, "mask_bce_loss": 0.5778894424438477, "mask_dice_loss": 0.06680836528539658, "mask_loss": 0.6446977853775024, "step": 255 }, { "epoch": 0.15375375375375375, "grad_norm": 29.900821685791016, "learning_rate": 1.9215361696291112e-06, "loss": 1.0643537044525146, "step": 256 }, { "ce_loss": 0.00013396533904597163, "cls_loss": 0.0478515625, "epoch": 0.15375375375375375, "mask_bce_loss": 0.5882830023765564, "mask_dice_loss": 0.08770862221717834, "mask_loss": 0.6759916543960571, "step": 256 }, { "epoch": 0.15435435435435435, "grad_norm": 17.948665618896484, "learning_rate": 1.9207790971604524e-06, "loss": 0.9184110760688782, "step": 257 }, { "ce_loss": 3.108628516201861e-05, "cls_loss": 0.05615234375, "epoch": 0.15435435435435435, "mask_bce_loss": 0.4904567301273346, "mask_dice_loss": 0.09170817583799362, "mask_loss": 0.5821648836135864, "step": 257 }, { "epoch": 0.15495495495495495, "grad_norm": 31.320228576660156, "learning_rate": 1.920018540436193e-06, "loss": 0.935221791267395, "step": 258 }, { "ce_loss": 0.00010723471496021375, "cls_loss": 0.0673828125, "epoch": 0.15495495495495495, "mask_bce_loss": 1.3124960660934448, "mask_dice_loss": 0.11925309151411057, "mask_loss": 1.4317491054534912, "step": 258 }, { "epoch": 0.15555555555555556, "grad_norm": 26.15648078918457, "learning_rate": 1.919254502334302e-06, "loss": 0.7456363439559937, "step": 259 }, { "ce_loss": 0.00010916146857198328, "cls_loss": 0.0341796875, "epoch": 0.15555555555555556, "mask_bce_loss": 0.22927060723304749, "mask_dice_loss": 0.02610352635383606, "mask_loss": 0.25537413358688354, "step": 259 }, { "epoch": 0.15615615615615616, "grad_norm": 26.28302764892578, "learning_rate": 1.918486985745923e-06, "loss": 0.7725459337234497, "step": 260 }, { "ce_loss": 3.506094799377024e-05, "cls_loss": 0.0810546875, "epoch": 0.15615615615615616, "mask_bce_loss": 1.4803494215011597, "mask_dice_loss": 0.1327158659696579, "mask_loss": 1.613065242767334, "step": 260 }, { "epoch": 0.15675675675675677, "grad_norm": 30.15642547607422, "learning_rate": 1.917715993575361e-06, "loss": 0.9059398770332336, "step": 261 }, { "ce_loss": 2.3359018086921424e-05, "cls_loss": 0.04833984375, "epoch": 0.15675675675675677, "mask_bce_loss": 0.40869012475013733, "mask_dice_loss": 0.06462729722261429, "mask_loss": 0.473317414522171, "step": 261 }, { "epoch": 0.15735735735735737, "grad_norm": 33.47853088378906, "learning_rate": 1.916941528740074e-06, "loss": 0.9316349029541016, "step": 262 }, { "ce_loss": 6.304634734988213e-05, "cls_loss": 0.050048828125, "epoch": 0.15735735735735737, "mask_bce_loss": 1.1523182392120361, "mask_dice_loss": 0.10237343609333038, "mask_loss": 1.25469172000885, "step": 262 }, { "epoch": 0.15795795795795795, "grad_norm": 16.250707626342773, "learning_rate": 1.9161635941706596e-06, "loss": 0.7312716841697693, "step": 263 }, { "ce_loss": 0.00011893564806086943, "cls_loss": 0.04833984375, "epoch": 0.15795795795795795, "mask_bce_loss": 0.743201732635498, "mask_dice_loss": 0.08326777070760727, "mask_loss": 0.8264694809913635, "step": 263 }, { "epoch": 0.15855855855855855, "grad_norm": 40.97865676879883, "learning_rate": 1.915382192810846e-06, "loss": 0.9177247285842896, "step": 264 }, { "ce_loss": 3.800244303420186e-05, "cls_loss": 0.06640625, "epoch": 0.15855855855855855, "mask_bce_loss": 1.5795804262161255, "mask_dice_loss": 0.0902739092707634, "mask_loss": 1.6698542833328247, "step": 264 }, { "epoch": 0.15915915915915915, "grad_norm": 18.072656631469727, "learning_rate": 1.914597327617479e-06, "loss": 0.9450907707214355, "step": 265 }, { "ce_loss": 3.0724884709343314e-05, "cls_loss": 0.030517578125, "epoch": 0.15915915915915915, "mask_bce_loss": 0.35523656010627747, "mask_dice_loss": 0.022728972136974335, "mask_loss": 0.3779655396938324, "step": 265 }, { "epoch": 0.15975975975975976, "grad_norm": 19.71143913269043, "learning_rate": 1.9138090015605115e-06, "loss": 0.8762261271476746, "step": 266 }, { "ce_loss": 0.0006458857678808272, "cls_loss": 0.06396484375, "epoch": 0.15975975975975976, "mask_bce_loss": 0.691317617893219, "mask_dice_loss": 0.12452571839094162, "mask_loss": 0.8158433437347412, "step": 266 }, { "epoch": 0.16036036036036036, "grad_norm": 16.00835609436035, "learning_rate": 1.9130172176229934e-06, "loss": 0.8546750545501709, "step": 267 }, { "ce_loss": 2.7358310035197064e-05, "cls_loss": 0.052734375, "epoch": 0.16036036036036036, "mask_bce_loss": 0.9639039039611816, "mask_dice_loss": 0.10644562542438507, "mask_loss": 1.0703495740890503, "step": 267 }, { "epoch": 0.16096096096096096, "grad_norm": 86.60547637939453, "learning_rate": 1.9122219788010585e-06, "loss": 0.8705011606216431, "step": 268 }, { "ce_loss": 0.00151939841452986, "cls_loss": 0.06201171875, "epoch": 0.16096096096096096, "mask_bce_loss": 0.34538713097572327, "mask_dice_loss": 0.12741811573505402, "mask_loss": 0.4728052616119385, "step": 268 }, { "epoch": 0.16156156156156157, "grad_norm": 29.808320999145508, "learning_rate": 1.9114232881039147e-06, "loss": 0.923174262046814, "step": 269 }, { "ce_loss": 2.3517195586464368e-05, "cls_loss": 0.025634765625, "epoch": 0.16156156156156157, "mask_bce_loss": 0.22264504432678223, "mask_dice_loss": 0.01750277169048786, "mask_loss": 0.24014781415462494, "step": 269 }, { "epoch": 0.16216216216216217, "grad_norm": 19.289960861206055, "learning_rate": 1.9106211485538304e-06, "loss": 0.8934556245803833, "step": 270 }, { "ce_loss": 6.047653005225584e-05, "cls_loss": 0.0673828125, "epoch": 0.16216216216216217, "mask_bce_loss": 1.9294782876968384, "mask_dice_loss": 0.11900156736373901, "mask_loss": 2.0484797954559326, "step": 270 }, { "epoch": 0.16276276276276277, "grad_norm": 47.45904541015625, "learning_rate": 1.909815563186127e-06, "loss": 1.013728380203247, "step": 271 }, { "ce_loss": 3.9025762816891074e-05, "cls_loss": 0.068359375, "epoch": 0.16276276276276277, "mask_bce_loss": 0.4542181193828583, "mask_dice_loss": 0.1475251168012619, "mask_loss": 0.601743221282959, "step": 271 }, { "epoch": 0.16336336336336335, "grad_norm": 25.340938568115234, "learning_rate": 1.9090065350491626e-06, "loss": 0.8776445388793945, "step": 272 }, { "ce_loss": 4.813397754332982e-05, "cls_loss": 0.06201171875, "epoch": 0.16336336336336335, "mask_bce_loss": 1.7644903659820557, "mask_dice_loss": 0.11406315863132477, "mask_loss": 1.8785535097122192, "step": 272 }, { "epoch": 0.16396396396396395, "grad_norm": 26.668752670288086, "learning_rate": 1.908194067204324e-06, "loss": 0.8448218703269958, "step": 273 }, { "ce_loss": 6.722627585986629e-05, "cls_loss": 0.06298828125, "epoch": 0.16396396396396395, "mask_bce_loss": 1.4073072671890259, "mask_dice_loss": 0.14371663331985474, "mask_loss": 1.5510239601135254, "step": 273 }, { "epoch": 0.16456456456456456, "grad_norm": 19.3915958404541, "learning_rate": 1.907378162726014e-06, "loss": 0.7643231749534607, "step": 274 }, { "ce_loss": 2.8037387892254628e-05, "cls_loss": 0.055419921875, "epoch": 0.16456456456456456, "mask_bce_loss": 0.548106849193573, "mask_dice_loss": 0.06273605674505234, "mask_loss": 0.6108428835868835, "step": 274 }, { "epoch": 0.16516516516516516, "grad_norm": 32.35401916503906, "learning_rate": 1.9065588247016393e-06, "loss": 0.9217748045921326, "step": 275 }, { "ce_loss": 0.005641249008476734, "cls_loss": 0.0771484375, "epoch": 0.16516516516516516, "mask_bce_loss": 1.628453254699707, "mask_dice_loss": 0.11996733397245407, "mask_loss": 1.7484205961227417, "step": 275 }, { "epoch": 0.16576576576576577, "grad_norm": 26.378599166870117, "learning_rate": 1.9057360562315995e-06, "loss": 0.7566087245941162, "step": 276 }, { "ce_loss": 0.0019601250533014536, "cls_loss": 0.049072265625, "epoch": 0.16576576576576577, "mask_bce_loss": 1.2266749143600464, "mask_dice_loss": 0.05932440981268883, "mask_loss": 1.2859992980957031, "step": 276 }, { "epoch": 0.16636636636636637, "grad_norm": 31.49818992614746, "learning_rate": 1.9049098604292752e-06, "loss": 1.0486005544662476, "step": 277 }, { "ce_loss": 2.5921543056028895e-05, "cls_loss": 0.06103515625, "epoch": 0.16636636636636637, "mask_bce_loss": 0.8684541583061218, "mask_dice_loss": 0.0881054624915123, "mask_loss": 0.9565595984458923, "step": 277 }, { "epoch": 0.16696696696696697, "grad_norm": 19.853862762451172, "learning_rate": 1.904080240421016e-06, "loss": 0.8144301176071167, "step": 278 }, { "ce_loss": 3.970962279709056e-05, "cls_loss": 0.03857421875, "epoch": 0.16696696696696697, "mask_bce_loss": 0.7017722725868225, "mask_dice_loss": 0.07186613231897354, "mask_loss": 0.7736384272575378, "step": 278 }, { "epoch": 0.16756756756756758, "grad_norm": 26.080116271972656, "learning_rate": 1.9032471993461288e-06, "loss": 0.9275593757629395, "step": 279 }, { "ce_loss": 0.00024085520999506116, "cls_loss": 0.03955078125, "epoch": 0.16756756756756758, "mask_bce_loss": 0.5645617246627808, "mask_dice_loss": 0.054741647094488144, "mask_loss": 0.6193033456802368, "step": 279 }, { "epoch": 0.16816816816816818, "grad_norm": 24.61683464050293, "learning_rate": 1.9024107403568658e-06, "loss": 0.82042396068573, "step": 280 }, { "ce_loss": 0.0004479968047235161, "cls_loss": 0.057373046875, "epoch": 0.16816816816816818, "mask_bce_loss": 0.19391633570194244, "mask_dice_loss": 0.07608436793088913, "mask_loss": 0.270000696182251, "step": 280 }, { "epoch": 0.16876876876876876, "grad_norm": 21.634511947631836, "learning_rate": 1.9015708666184125e-06, "loss": 0.9345443844795227, "step": 281 }, { "ce_loss": 4.042625369038433e-05, "cls_loss": 0.017333984375, "epoch": 0.16876876876876876, "mask_bce_loss": 0.09005454927682877, "mask_dice_loss": 0.010323193855583668, "mask_loss": 0.10037774592638016, "step": 281 }, { "epoch": 0.16936936936936936, "grad_norm": 37.22487258911133, "learning_rate": 1.9007275813088768e-06, "loss": 0.8848097324371338, "step": 282 }, { "ce_loss": 2.6525040084379725e-05, "cls_loss": 0.04736328125, "epoch": 0.16936936936936936, "mask_bce_loss": 0.4630919098854065, "mask_dice_loss": 0.09056874364614487, "mask_loss": 0.5536606311798096, "step": 282 }, { "epoch": 0.16996996996996996, "grad_norm": 21.294363021850586, "learning_rate": 1.899880887619275e-06, "loss": 0.8760217428207397, "step": 283 }, { "ce_loss": 0.004578791093081236, "cls_loss": 0.0419921875, "epoch": 0.16996996996996996, "mask_bce_loss": 0.6642252802848816, "mask_dice_loss": 0.05953902006149292, "mask_loss": 0.7237643003463745, "step": 283 }, { "epoch": 0.17057057057057057, "grad_norm": 22.240741729736328, "learning_rate": 1.8990307887535216e-06, "loss": 0.7608683109283447, "step": 284 }, { "ce_loss": 9.779600804904476e-05, "cls_loss": 0.059814453125, "epoch": 0.17057057057057057, "mask_bce_loss": 0.6420416831970215, "mask_dice_loss": 0.0961371585726738, "mask_loss": 0.7381788492202759, "step": 284 }, { "epoch": 0.17117117117117117, "grad_norm": 33.08860778808594, "learning_rate": 1.8981772879284158e-06, "loss": 0.9293307662010193, "step": 285 }, { "ce_loss": 3.320080213597976e-05, "cls_loss": 0.0201416015625, "epoch": 0.17117117117117117, "mask_bce_loss": 0.19312207400798798, "mask_dice_loss": 0.01292011421173811, "mask_loss": 0.20604218542575836, "step": 285 }, { "epoch": 0.17177177177177178, "grad_norm": 29.037317276000977, "learning_rate": 1.8973203883736306e-06, "loss": 0.9151370525360107, "step": 286 }, { "ce_loss": 7.272457150975242e-05, "cls_loss": 0.04833984375, "epoch": 0.17177177177177178, "mask_bce_loss": 0.5862894654273987, "mask_dice_loss": 0.05081179365515709, "mask_loss": 0.6371012330055237, "step": 286 }, { "epoch": 0.17237237237237238, "grad_norm": 19.67023277282715, "learning_rate": 1.8964600933316995e-06, "loss": 0.8060640096664429, "step": 287 }, { "ce_loss": 0.09275515377521515, "cls_loss": 0.04345703125, "epoch": 0.17237237237237238, "mask_bce_loss": 0.5367159843444824, "mask_dice_loss": 0.04017702490091324, "mask_loss": 0.5768930315971375, "step": 287 }, { "epoch": 0.17297297297297298, "grad_norm": 14.024560928344727, "learning_rate": 1.8955964060580044e-06, "loss": 0.7908521294593811, "step": 288 }, { "ce_loss": 0.00025057035963982344, "cls_loss": 0.04443359375, "epoch": 0.17297297297297298, "mask_bce_loss": 1.4083945751190186, "mask_dice_loss": 0.14561574161052704, "mask_loss": 1.554010272026062, "step": 288 }, { "epoch": 0.1735735735735736, "grad_norm": 24.73820686340332, "learning_rate": 1.8947293298207633e-06, "loss": 0.7270078659057617, "step": 289 }, { "ce_loss": 6.353144999593496e-05, "cls_loss": 0.041748046875, "epoch": 0.1735735735735736, "mask_bce_loss": 0.708664059638977, "mask_dice_loss": 0.04461383447051048, "mask_loss": 0.7532778978347778, "step": 289 }, { "epoch": 0.17417417417417416, "grad_norm": 29.977596282958984, "learning_rate": 1.8938588679010194e-06, "loss": 0.8611330986022949, "step": 290 }, { "ce_loss": 6.0137965192552656e-05, "cls_loss": 0.056640625, "epoch": 0.17417417417417416, "mask_bce_loss": 1.3353337049484253, "mask_dice_loss": 0.12322293967008591, "mask_loss": 1.4585566520690918, "step": 290 }, { "epoch": 0.17477477477477477, "grad_norm": 44.30724334716797, "learning_rate": 1.8929850235926263e-06, "loss": 0.8356292843818665, "step": 291 }, { "ce_loss": 3.220617873012088e-05, "cls_loss": 0.034423828125, "epoch": 0.17477477477477477, "mask_bce_loss": 0.9302794337272644, "mask_dice_loss": 0.039075229316949844, "mask_loss": 0.9693546891212463, "step": 291 }, { "epoch": 0.17537537537537537, "grad_norm": 48.01213073730469, "learning_rate": 1.8921078002022367e-06, "loss": 0.9421887397766113, "step": 292 }, { "ce_loss": 3.6449830076890066e-05, "cls_loss": 0.06640625, "epoch": 0.17537537537537537, "mask_bce_loss": 1.0106266736984253, "mask_dice_loss": 0.13662932813167572, "mask_loss": 1.1472560167312622, "step": 292 }, { "epoch": 0.17597597597597597, "grad_norm": 15.161629676818848, "learning_rate": 1.8912272010492906e-06, "loss": 0.7981972694396973, "step": 293 }, { "ce_loss": 0.007075680419802666, "cls_loss": 0.05419921875, "epoch": 0.17597597597597597, "mask_bce_loss": 0.6998148560523987, "mask_dice_loss": 0.077815942466259, "mask_loss": 0.7776308059692383, "step": 293 }, { "epoch": 0.17657657657657658, "grad_norm": 27.265993118286133, "learning_rate": 1.8903432294660008e-06, "loss": 0.9764852523803711, "step": 294 }, { "ce_loss": 7.821038161637262e-05, "cls_loss": 0.03369140625, "epoch": 0.17657657657657658, "mask_bce_loss": 0.37340056896209717, "mask_dice_loss": 0.042827606201171875, "mask_loss": 0.41622817516326904, "step": 294 }, { "epoch": 0.17717717717717718, "grad_norm": 14.02545166015625, "learning_rate": 1.8894558887973432e-06, "loss": 0.8163813948631287, "step": 295 }, { "ce_loss": 5.517762474482879e-05, "cls_loss": 0.03857421875, "epoch": 0.17717717717717718, "mask_bce_loss": 0.27441248297691345, "mask_dice_loss": 0.04007657989859581, "mask_loss": 0.31448906660079956, "step": 295 }, { "epoch": 0.17777777777777778, "grad_norm": 24.54128074645996, "learning_rate": 1.8885651824010402e-06, "loss": 0.9123126864433289, "step": 296 }, { "ce_loss": 3.573377398424782e-05, "cls_loss": 0.0283203125, "epoch": 0.17777777777777778, "mask_bce_loss": 0.18939794600009918, "mask_dice_loss": 0.04260075464844704, "mask_loss": 0.23199869692325592, "step": 296 }, { "epoch": 0.1783783783783784, "grad_norm": 21.422494888305664, "learning_rate": 1.8876711136475523e-06, "loss": 0.8481888771057129, "step": 297 }, { "ce_loss": 6.056494021322578e-05, "cls_loss": 0.038818359375, "epoch": 0.1783783783783784, "mask_bce_loss": 0.36323362588882446, "mask_dice_loss": 0.06332721561193466, "mask_loss": 0.4265608489513397, "step": 297 }, { "epoch": 0.178978978978979, "grad_norm": 23.217731475830078, "learning_rate": 1.8867736859200618e-06, "loss": 0.9321725368499756, "step": 298 }, { "ce_loss": 4.423605423653498e-05, "cls_loss": 0.05615234375, "epoch": 0.178978978978979, "mask_bce_loss": 0.6979722380638123, "mask_dice_loss": 0.11779556423425674, "mask_loss": 0.8157678246498108, "step": 298 }, { "epoch": 0.17957957957957957, "grad_norm": 18.45623779296875, "learning_rate": 1.8858729026144625e-06, "loss": 0.9097012281417847, "step": 299 }, { "ce_loss": 0.00010278630361426622, "cls_loss": 0.048095703125, "epoch": 0.17957957957957957, "mask_bce_loss": 0.7791029810905457, "mask_dice_loss": 0.07808426767587662, "mask_loss": 0.8571872711181641, "step": 299 }, { "epoch": 0.18018018018018017, "grad_norm": 18.817949295043945, "learning_rate": 1.8849687671393448e-06, "loss": 0.7348836660385132, "step": 300 }, { "ce_loss": 0.0005463321576826274, "cls_loss": 0.05517578125, "epoch": 0.18018018018018017, "mask_bce_loss": 0.6496272087097168, "mask_dice_loss": 0.07939384132623672, "mask_loss": 0.7290210723876953, "step": 300 }, { "epoch": 0.18078078078078078, "grad_norm": 54.84593963623047, "learning_rate": 1.8840612829159849e-06, "loss": 0.959036111831665, "step": 301 }, { "ce_loss": 4.691944195656106e-05, "cls_loss": 0.036865234375, "epoch": 0.18078078078078078, "mask_bce_loss": 0.38006240129470825, "mask_dice_loss": 0.03233053907752037, "mask_loss": 0.4123929440975189, "step": 301 }, { "epoch": 0.18138138138138138, "grad_norm": 42.99876022338867, "learning_rate": 1.8831504533783296e-06, "loss": 1.0280773639678955, "step": 302 }, { "ce_loss": 0.0003800969570875168, "cls_loss": 0.055908203125, "epoch": 0.18138138138138138, "mask_bce_loss": 0.8473971486091614, "mask_dice_loss": 0.11051386594772339, "mask_loss": 0.9579110145568848, "step": 302 }, { "epoch": 0.18198198198198198, "grad_norm": 18.20298957824707, "learning_rate": 1.8822362819729856e-06, "loss": 0.7222795486450195, "step": 303 }, { "ce_loss": 0.0001150208554463461, "cls_loss": 0.04443359375, "epoch": 0.18198198198198198, "mask_bce_loss": 0.8432151675224304, "mask_dice_loss": 0.0538436658680439, "mask_loss": 0.8970588445663452, "step": 303 }, { "epoch": 0.1825825825825826, "grad_norm": 14.948430061340332, "learning_rate": 1.8813187721592045e-06, "loss": 0.8759507536888123, "step": 304 }, { "ce_loss": 4.914875898975879e-05, "cls_loss": 0.0654296875, "epoch": 0.1825825825825826, "mask_bce_loss": 1.1438268423080444, "mask_dice_loss": 0.10970952361822128, "mask_loss": 1.253536343574524, "step": 304 }, { "epoch": 0.1831831831831832, "grad_norm": 37.28888702392578, "learning_rate": 1.8803979274088709e-06, "loss": 0.8592754602432251, "step": 305 }, { "ce_loss": 6.026294067851268e-05, "cls_loss": 0.04248046875, "epoch": 0.1831831831831832, "mask_bce_loss": 0.2987963557243347, "mask_dice_loss": 0.04396688938140869, "mask_loss": 0.3427632451057434, "step": 305 }, { "epoch": 0.1837837837837838, "grad_norm": 25.518085479736328, "learning_rate": 1.8794737512064888e-06, "loss": 0.8309555053710938, "step": 306 }, { "ce_loss": 0.0002526906318962574, "cls_loss": 0.05517578125, "epoch": 0.1837837837837838, "mask_bce_loss": 1.0497968196868896, "mask_dice_loss": 0.05424086004495621, "mask_loss": 1.1040376424789429, "step": 306 }, { "epoch": 0.1843843843843844, "grad_norm": 20.556734085083008, "learning_rate": 1.8785462470491688e-06, "loss": 0.9727604985237122, "step": 307 }, { "ce_loss": 4.3081181502202526e-05, "cls_loss": 0.04443359375, "epoch": 0.1843843843843844, "mask_bce_loss": 0.43879780173301697, "mask_dice_loss": 0.05772171542048454, "mask_loss": 0.4965195059776306, "step": 307 }, { "epoch": 0.18498498498498497, "grad_norm": 18.213600158691406, "learning_rate": 1.8776154184466143e-06, "loss": 0.8383730053901672, "step": 308 }, { "ce_loss": 0.00022162016830407083, "cls_loss": 0.06396484375, "epoch": 0.18498498498498497, "mask_bce_loss": 0.6638882756233215, "mask_dice_loss": 0.10904481261968613, "mask_loss": 0.7729330658912659, "step": 308 }, { "epoch": 0.18558558558558558, "grad_norm": 15.92707347869873, "learning_rate": 1.876681268921108e-06, "loss": 0.7732937335968018, "step": 309 }, { "ce_loss": 4.417903255671263e-05, "cls_loss": 0.03759765625, "epoch": 0.18558558558558558, "mask_bce_loss": 0.5414816737174988, "mask_dice_loss": 0.04435382038354874, "mask_loss": 0.5858355164527893, "step": 309 }, { "epoch": 0.18618618618618618, "grad_norm": 33.2056999206543, "learning_rate": 1.8757438020075e-06, "loss": 1.1601550579071045, "step": 310 }, { "ce_loss": 4.336065467214212e-05, "cls_loss": 0.0546875, "epoch": 0.18618618618618618, "mask_bce_loss": 1.2459951639175415, "mask_dice_loss": 0.07255951315164566, "mask_loss": 1.3185546398162842, "step": 310 }, { "epoch": 0.18678678678678678, "grad_norm": 27.656553268432617, "learning_rate": 1.8748030212531933e-06, "loss": 0.7696192860603333, "step": 311 }, { "ce_loss": 4.150990935158916e-05, "cls_loss": 0.06494140625, "epoch": 0.18678678678678678, "mask_bce_loss": 0.7261914610862732, "mask_dice_loss": 0.11253774166107178, "mask_loss": 0.838729202747345, "step": 311 }, { "epoch": 0.1873873873873874, "grad_norm": 20.597444534301758, "learning_rate": 1.8738589302181299e-06, "loss": 0.9077986478805542, "step": 312 }, { "ce_loss": 3.622499207267538e-05, "cls_loss": 0.046630859375, "epoch": 0.1873873873873874, "mask_bce_loss": 1.0609567165374756, "mask_dice_loss": 0.06548478454351425, "mask_loss": 1.126441478729248, "step": 312 }, { "epoch": 0.187987987987988, "grad_norm": 18.115386962890625, "learning_rate": 1.8729115324747787e-06, "loss": 0.8477374315261841, "step": 313 }, { "ce_loss": 0.00013631403271574527, "cls_loss": 0.040283203125, "epoch": 0.187987987987988, "mask_bce_loss": 0.5795779228210449, "mask_dice_loss": 0.05274156481027603, "mask_loss": 0.6323195099830627, "step": 313 }, { "epoch": 0.1885885885885886, "grad_norm": 42.64849853515625, "learning_rate": 1.871960831608121e-06, "loss": 0.9290210008621216, "step": 314 }, { "ce_loss": 0.0008361681248061359, "cls_loss": 0.036865234375, "epoch": 0.1885885885885886, "mask_bce_loss": 0.42854976654052734, "mask_dice_loss": 0.06385860592126846, "mask_loss": 0.4924083650112152, "step": 314 }, { "epoch": 0.1891891891891892, "grad_norm": 67.1750717163086, "learning_rate": 1.8710068312156375e-06, "loss": 0.9550722241401672, "step": 315 }, { "ce_loss": 2.004177622438874e-05, "cls_loss": 0.03759765625, "epoch": 0.1891891891891892, "mask_bce_loss": 0.3677433431148529, "mask_dice_loss": 0.033144641667604446, "mask_loss": 0.40088799595832825, "step": 315 }, { "epoch": 0.1897897897897898, "grad_norm": 28.86111831665039, "learning_rate": 1.870049534907294e-06, "loss": 0.881196141242981, "step": 316 }, { "ce_loss": 3.336356894578785e-05, "cls_loss": 0.05810546875, "epoch": 0.1897897897897898, "mask_bce_loss": 1.5280998945236206, "mask_dice_loss": 0.11693301051855087, "mask_loss": 1.6450328826904297, "step": 316 }, { "epoch": 0.19039039039039038, "grad_norm": 34.78274154663086, "learning_rate": 1.8690889463055281e-06, "loss": 0.8328103423118591, "step": 317 }, { "ce_loss": 2.6891391826211475e-05, "cls_loss": 0.0322265625, "epoch": 0.19039039039039038, "mask_bce_loss": 0.23843708634376526, "mask_dice_loss": 0.02413339912891388, "mask_loss": 0.26257050037384033, "step": 317 }, { "epoch": 0.19099099099099098, "grad_norm": 19.084943771362305, "learning_rate": 1.8681250690452362e-06, "loss": 0.7293296456336975, "step": 318 }, { "ce_loss": 0.00012753679766319692, "cls_loss": 0.0400390625, "epoch": 0.19099099099099098, "mask_bce_loss": 0.3485320210456848, "mask_dice_loss": 0.046052541583776474, "mask_loss": 0.3945845663547516, "step": 318 }, { "epoch": 0.1915915915915916, "grad_norm": 23.769758224487305, "learning_rate": 1.8671579067737583e-06, "loss": 0.7442349195480347, "step": 319 }, { "ce_loss": 0.00011777569307014346, "cls_loss": 0.06591796875, "epoch": 0.1915915915915916, "mask_bce_loss": 0.5630499124526978, "mask_dice_loss": 0.0759003683924675, "mask_loss": 0.6389502882957458, "step": 319 }, { "epoch": 0.1921921921921922, "grad_norm": 40.690162658691406, "learning_rate": 1.8661874631508656e-06, "loss": 1.2781070470809937, "step": 320 }, { "ce_loss": 2.7550942832021974e-05, "cls_loss": 0.05810546875, "epoch": 0.1921921921921922, "mask_bce_loss": 0.652239978313446, "mask_dice_loss": 0.10599683970212936, "mask_loss": 0.758236825466156, "step": 320 }, { "epoch": 0.1927927927927928, "grad_norm": 25.143367767333984, "learning_rate": 1.8652137418487451e-06, "loss": 0.8925929069519043, "step": 321 }, { "ce_loss": 7.67186502343975e-05, "cls_loss": 0.0693359375, "epoch": 0.1927927927927928, "mask_bce_loss": 1.3825604915618896, "mask_dice_loss": 0.12970435619354248, "mask_loss": 1.5122648477554321, "step": 321 }, { "epoch": 0.1933933933933934, "grad_norm": 34.61191177368164, "learning_rate": 1.864236746551988e-06, "loss": 1.0428472757339478, "step": 322 }, { "ce_loss": 3.658635978354141e-05, "cls_loss": 0.038818359375, "epoch": 0.1933933933933934, "mask_bce_loss": 0.25333401560783386, "mask_dice_loss": 0.03461773321032524, "mask_loss": 0.2879517376422882, "step": 322 }, { "epoch": 0.193993993993994, "grad_norm": 54.08476257324219, "learning_rate": 1.8632564809575738e-06, "loss": 0.884190559387207, "step": 323 }, { "ce_loss": 8.652218093629926e-05, "cls_loss": 0.05517578125, "epoch": 0.193993993993994, "mask_bce_loss": 0.8283694386482239, "mask_dice_loss": 0.12111016362905502, "mask_loss": 0.9494795799255371, "step": 323 }, { "epoch": 0.1945945945945946, "grad_norm": 60.93608856201172, "learning_rate": 1.862272948774856e-06, "loss": 0.7724469900131226, "step": 324 }, { "ce_loss": 6.504459452116862e-05, "cls_loss": 0.0498046875, "epoch": 0.1945945945945946, "mask_bce_loss": 0.5344745516777039, "mask_dice_loss": 0.05489708110690117, "mask_loss": 0.5893716216087341, "step": 324 }, { "epoch": 0.19519519519519518, "grad_norm": 29.08200454711914, "learning_rate": 1.8612861537255504e-06, "loss": 0.8754069209098816, "step": 325 }, { "ce_loss": 4.9911315727513283e-05, "cls_loss": 0.046142578125, "epoch": 0.19519519519519518, "mask_bce_loss": 0.5110889673233032, "mask_dice_loss": 0.0500049851834774, "mask_loss": 0.5610939264297485, "step": 325 }, { "epoch": 0.19579579579579579, "grad_norm": 28.365564346313477, "learning_rate": 1.8602960995437184e-06, "loss": 0.8384197950363159, "step": 326 }, { "ce_loss": 3.495362034300342e-05, "cls_loss": 0.06494140625, "epoch": 0.19579579579579579, "mask_bce_loss": 0.8514156341552734, "mask_dice_loss": 0.11684443056583405, "mask_loss": 0.9682600498199463, "step": 326 }, { "epoch": 0.1963963963963964, "grad_norm": 19.269142150878906, "learning_rate": 1.8593027899757555e-06, "loss": 0.7247942090034485, "step": 327 }, { "ce_loss": 2.971506728499662e-05, "cls_loss": 0.059326171875, "epoch": 0.1963963963963964, "mask_bce_loss": 1.0459983348846436, "mask_dice_loss": 0.11222483217716217, "mask_loss": 1.1582231521606445, "step": 327 }, { "epoch": 0.196996996996997, "grad_norm": 21.245851516723633, "learning_rate": 1.8583062287803736e-06, "loss": 0.874213695526123, "step": 328 }, { "ce_loss": 1.799330857465975e-05, "cls_loss": 0.030029296875, "epoch": 0.196996996996997, "mask_bce_loss": 0.2934335768222809, "mask_dice_loss": 0.022125571966171265, "mask_loss": 0.31555914878845215, "step": 328 }, { "epoch": 0.1975975975975976, "grad_norm": 24.422000885009766, "learning_rate": 1.8573064197285905e-06, "loss": 0.7467973828315735, "step": 329 }, { "ce_loss": 0.00011598399578360841, "cls_loss": 0.03515625, "epoch": 0.1975975975975976, "mask_bce_loss": 0.35478994250297546, "mask_dice_loss": 0.03459532558917999, "mask_loss": 0.38938528299331665, "step": 329 }, { "epoch": 0.1981981981981982, "grad_norm": 17.80179214477539, "learning_rate": 1.856303366603714e-06, "loss": 0.7508030533790588, "step": 330 }, { "ce_loss": 0.00017865344125311822, "cls_loss": 0.059326171875, "epoch": 0.1981981981981982, "mask_bce_loss": 1.3334940671920776, "mask_dice_loss": 0.10796060413122177, "mask_loss": 1.4414546489715576, "step": 330 }, { "epoch": 0.1987987987987988, "grad_norm": 127.92483520507812, "learning_rate": 1.8552970732013265e-06, "loss": 0.8026978969573975, "step": 331 }, { "ce_loss": 6.977170414756984e-05, "cls_loss": 0.05029296875, "epoch": 0.1987987987987988, "mask_bce_loss": 0.7431357502937317, "mask_dice_loss": 0.09103663265705109, "mask_loss": 0.8341723680496216, "step": 331 }, { "epoch": 0.1993993993993994, "grad_norm": 27.56947135925293, "learning_rate": 1.8542875433292724e-06, "loss": 0.8238087892532349, "step": 332 }, { "ce_loss": 1.8038832422462292e-05, "cls_loss": 0.059326171875, "epoch": 0.1993993993993994, "mask_bce_loss": 0.7993395328521729, "mask_dice_loss": 0.10745211690664291, "mask_loss": 0.906791627407074, "step": 332 }, { "epoch": 0.2, "grad_norm": 27.916406631469727, "learning_rate": 1.853274780807643e-06, "loss": 0.8945143222808838, "step": 333 }, { "ce_loss": 7.923603698145598e-05, "cls_loss": 0.06591796875, "epoch": 0.2, "mask_bce_loss": 0.5048395395278931, "mask_dice_loss": 0.10759972780942917, "mask_loss": 0.6124392747879028, "step": 333 }, { "epoch": 0.2006006006006006, "grad_norm": 22.832008361816406, "learning_rate": 1.8522587894687617e-06, "loss": 0.9354227781295776, "step": 334 }, { "ce_loss": 3.206499241059646e-05, "cls_loss": 0.0849609375, "epoch": 0.2006006006006006, "mask_bce_loss": 0.6602001786231995, "mask_dice_loss": 0.08285895735025406, "mask_loss": 0.7430591583251953, "step": 334 }, { "epoch": 0.2012012012012012, "grad_norm": 13.668551445007324, "learning_rate": 1.8512395731571706e-06, "loss": 0.7660577893257141, "step": 335 }, { "ce_loss": 3.364485746715218e-05, "cls_loss": 0.0703125, "epoch": 0.2012012012012012, "mask_bce_loss": 1.8069229125976562, "mask_dice_loss": 0.1142597571015358, "mask_loss": 1.921182632446289, "step": 335 }, { "epoch": 0.2018018018018018, "grad_norm": 75.6722183227539, "learning_rate": 1.8502171357296142e-06, "loss": 0.8287543058395386, "step": 336 }, { "ce_loss": 7.99888075562194e-05, "cls_loss": 0.05517578125, "epoch": 0.2018018018018018, "mask_bce_loss": 0.7365630269050598, "mask_dice_loss": 0.09906300902366638, "mask_loss": 0.8356260061264038, "step": 336 }, { "epoch": 0.2024024024024024, "grad_norm": 24.969730377197266, "learning_rate": 1.8491914810550261e-06, "loss": 0.8475737571716309, "step": 337 }, { "ce_loss": 0.0004771831736434251, "cls_loss": 0.048828125, "epoch": 0.2024024024024024, "mask_bce_loss": 0.851601243019104, "mask_dice_loss": 0.1002691388130188, "mask_loss": 0.9518703818321228, "step": 337 }, { "epoch": 0.203003003003003, "grad_norm": 19.748706817626953, "learning_rate": 1.8481626130145145e-06, "loss": 0.9362854361534119, "step": 338 }, { "ce_loss": 3.685442788992077e-05, "cls_loss": 0.0693359375, "epoch": 0.203003003003003, "mask_bce_loss": 0.9629837274551392, "mask_dice_loss": 0.13664723932743073, "mask_loss": 1.0996309518814087, "step": 338 }, { "epoch": 0.2036036036036036, "grad_norm": 23.284156799316406, "learning_rate": 1.8471305355013471e-06, "loss": 0.7799349427223206, "step": 339 }, { "ce_loss": 0.0005375450709834695, "cls_loss": 0.060546875, "epoch": 0.2036036036036036, "mask_bce_loss": 1.0042105913162231, "mask_dice_loss": 0.11133121699094772, "mask_loss": 1.1155418157577515, "step": 339 }, { "epoch": 0.2042042042042042, "grad_norm": 22.811481475830078, "learning_rate": 1.8460952524209353e-06, "loss": 0.8401907682418823, "step": 340 }, { "ce_loss": 2.002583642024547e-05, "cls_loss": 0.0322265625, "epoch": 0.2042042042042042, "mask_bce_loss": 0.29723721742630005, "mask_dice_loss": 0.02266654558479786, "mask_loss": 0.31990376114845276, "step": 340 }, { "epoch": 0.2048048048048048, "grad_norm": 35.534305572509766, "learning_rate": 1.845056767690822e-06, "loss": 0.9013881683349609, "step": 341 }, { "ce_loss": 0.013604472391307354, "cls_loss": 0.04638671875, "epoch": 0.2048048048048048, "mask_bce_loss": 0.3189810812473297, "mask_dice_loss": 0.04989127442240715, "mask_loss": 0.36887234449386597, "step": 341 }, { "epoch": 0.20540540540540542, "grad_norm": 25.97873878479004, "learning_rate": 1.8440150852406635e-06, "loss": 1.0286461114883423, "step": 342 }, { "ce_loss": 3.642536103143357e-05, "cls_loss": 0.048095703125, "epoch": 0.20540540540540542, "mask_bce_loss": 1.002820611000061, "mask_dice_loss": 0.05585373565554619, "mask_loss": 1.0586743354797363, "step": 342 }, { "epoch": 0.206006006006006, "grad_norm": 29.41646957397461, "learning_rate": 1.8429702090122182e-06, "loss": 0.9537880420684814, "step": 343 }, { "ce_loss": 3.276179268141277e-05, "cls_loss": 0.1474609375, "epoch": 0.206006006006006, "mask_bce_loss": 0.8458601236343384, "mask_dice_loss": 0.06282173097133636, "mask_loss": 0.9086818695068359, "step": 343 }, { "epoch": 0.2066066066066066, "grad_norm": 32.850589752197266, "learning_rate": 1.8419221429593283e-06, "loss": 1.1642249822616577, "step": 344 }, { "ce_loss": 7.691881182836369e-05, "cls_loss": 0.11474609375, "epoch": 0.2066066066066066, "mask_bce_loss": 0.579067051410675, "mask_dice_loss": 0.058791451156139374, "mask_loss": 0.637858510017395, "step": 344 }, { "epoch": 0.2072072072072072, "grad_norm": 51.05511474609375, "learning_rate": 1.8408708910479075e-06, "loss": 0.9601329565048218, "step": 345 }, { "ce_loss": 0.00011025741696357727, "cls_loss": 0.03564453125, "epoch": 0.2072072072072072, "mask_bce_loss": 0.305345356464386, "mask_dice_loss": 0.030132001265883446, "mask_loss": 0.335477352142334, "step": 345 }, { "epoch": 0.2078078078078078, "grad_norm": 37.15113067626953, "learning_rate": 1.839816457255924e-06, "loss": 0.7629492282867432, "step": 346 }, { "ce_loss": 3.11590192723088e-05, "cls_loss": 0.0712890625, "epoch": 0.2078078078078078, "mask_bce_loss": 0.6195343136787415, "mask_dice_loss": 0.1082386001944542, "mask_loss": 0.7277728915214539, "step": 346 }, { "epoch": 0.2084084084084084, "grad_norm": 31.095054626464844, "learning_rate": 1.8387588455733873e-06, "loss": 0.8400722146034241, "step": 347 }, { "ce_loss": 2.9849041311535984e-05, "cls_loss": 0.06787109375, "epoch": 0.2084084084084084, "mask_bce_loss": 1.0828725099563599, "mask_dice_loss": 0.12140537798404694, "mask_loss": 1.2042778730392456, "step": 347 }, { "epoch": 0.209009009009009, "grad_norm": 17.05228042602539, "learning_rate": 1.837698060002331e-06, "loss": 0.8222233653068542, "step": 348 }, { "ce_loss": 0.00012876020628027618, "cls_loss": 0.0458984375, "epoch": 0.209009009009009, "mask_bce_loss": 0.28245848417282104, "mask_dice_loss": 0.06907566636800766, "mask_loss": 0.3515341579914093, "step": 348 }, { "epoch": 0.20960960960960962, "grad_norm": 16.02701187133789, "learning_rate": 1.8366341045568e-06, "loss": 0.8645038604736328, "step": 349 }, { "ce_loss": 5.205212073633447e-05, "cls_loss": 0.05615234375, "epoch": 0.20960960960960962, "mask_bce_loss": 0.8514220118522644, "mask_dice_loss": 0.09369261562824249, "mask_loss": 0.9451146125793457, "step": 349 }, { "epoch": 0.21021021021021022, "grad_norm": 18.804824829101562, "learning_rate": 1.8355669832628329e-06, "loss": 0.9188249111175537, "step": 350 }, { "ce_loss": 0.004676527343690395, "cls_loss": 0.051513671875, "epoch": 0.21021021021021022, "mask_bce_loss": 0.4656953811645508, "mask_dice_loss": 0.07893943041563034, "mask_loss": 0.5446348190307617, "step": 350 }, { "epoch": 0.21081081081081082, "grad_norm": 20.431392669677734, "learning_rate": 1.8344967001584488e-06, "loss": 0.8923936486244202, "step": 351 }, { "ce_loss": 0.0003283513360656798, "cls_loss": 0.05322265625, "epoch": 0.21081081081081082, "mask_bce_loss": 0.5199295878410339, "mask_dice_loss": 0.08092938363552094, "mask_loss": 0.6008589863777161, "step": 351 }, { "epoch": 0.2114114114114114, "grad_norm": 31.545705795288086, "learning_rate": 1.8334232592936313e-06, "loss": 0.8531376123428345, "step": 352 }, { "ce_loss": 0.0001261788565898314, "cls_loss": 0.03173828125, "epoch": 0.2114114114114114, "mask_bce_loss": 0.3438553512096405, "mask_dice_loss": 0.06199140474200249, "mask_loss": 0.4058467447757721, "step": 352 }, { "epoch": 0.212012012012012, "grad_norm": 20.438302993774414, "learning_rate": 1.8323466647303117e-06, "loss": 0.8672056198120117, "step": 353 }, { "ce_loss": 0.00010999249207088724, "cls_loss": 0.047119140625, "epoch": 0.212012012012012, "mask_bce_loss": 0.5996617674827576, "mask_dice_loss": 0.05988258868455887, "mask_loss": 0.6595443487167358, "step": 353 }, { "epoch": 0.2126126126126126, "grad_norm": 16.878576278686523, "learning_rate": 1.8312669205423567e-06, "loss": 0.9760626554489136, "step": 354 }, { "ce_loss": 4.377979348646477e-05, "cls_loss": 0.0478515625, "epoch": 0.2126126126126126, "mask_bce_loss": 0.9474528431892395, "mask_dice_loss": 0.08161884546279907, "mask_loss": 1.0290716886520386, "step": 354 }, { "epoch": 0.2132132132132132, "grad_norm": 17.856098175048828, "learning_rate": 1.8301840308155506e-06, "loss": 0.8688492178916931, "step": 355 }, { "ce_loss": 0.00031722179846838117, "cls_loss": 0.037841796875, "epoch": 0.2132132132132132, "mask_bce_loss": 0.5203946232795715, "mask_dice_loss": 0.0384010411798954, "mask_loss": 0.558795690536499, "step": 355 }, { "epoch": 0.2138138138138138, "grad_norm": 29.848665237426758, "learning_rate": 1.8290979996475802e-06, "loss": 0.8386516571044922, "step": 356 }, { "ce_loss": 0.0008350819698534906, "cls_loss": 0.040283203125, "epoch": 0.2138138138138138, "mask_bce_loss": 0.4109286963939667, "mask_dice_loss": 0.03233465179800987, "mask_loss": 0.44326335191726685, "step": 356 }, { "epoch": 0.21441441441441442, "grad_norm": 24.65736961364746, "learning_rate": 1.82800883114802e-06, "loss": 0.740216851234436, "step": 357 }, { "ce_loss": 5.78937579120975e-05, "cls_loss": 0.0306396484375, "epoch": 0.21441441441441442, "mask_bce_loss": 0.23452065885066986, "mask_dice_loss": 0.03836342692375183, "mask_loss": 0.2728840708732605, "step": 357 }, { "epoch": 0.21501501501501502, "grad_norm": 20.695955276489258, "learning_rate": 1.826916529438316e-06, "loss": 0.7593165636062622, "step": 358 }, { "ce_loss": 9.861075523076579e-05, "cls_loss": 0.059326171875, "epoch": 0.21501501501501502, "mask_bce_loss": 0.9974374175071716, "mask_dice_loss": 0.10834123194217682, "mask_loss": 1.105778694152832, "step": 358 }, { "epoch": 0.21561561561561562, "grad_norm": 21.32268714904785, "learning_rate": 1.8258210986517705e-06, "loss": 0.8469101190567017, "step": 359 }, { "ce_loss": 4.204966899123974e-05, "cls_loss": 0.05419921875, "epoch": 0.21561561561561562, "mask_bce_loss": 0.8730104565620422, "mask_dice_loss": 0.12647949159145355, "mask_loss": 0.999489963054657, "step": 359 }, { "epoch": 0.21621621621621623, "grad_norm": 28.72339630126953, "learning_rate": 1.8247225429335268e-06, "loss": 0.839093804359436, "step": 360 }, { "ce_loss": 0.00023592147044837475, "cls_loss": 0.040771484375, "epoch": 0.21621621621621623, "mask_bce_loss": 0.5155500769615173, "mask_dice_loss": 0.036694008857011795, "mask_loss": 0.5522440671920776, "step": 360 }, { "epoch": 0.2168168168168168, "grad_norm": 47.966209411621094, "learning_rate": 1.8236208664405517e-06, "loss": 0.9870973825454712, "step": 361 }, { "ce_loss": 0.00015204479859676212, "cls_loss": 0.0400390625, "epoch": 0.2168168168168168, "mask_bce_loss": 0.2738364636898041, "mask_dice_loss": 0.03987567499279976, "mask_loss": 0.31371214985847473, "step": 361 }, { "epoch": 0.2174174174174174, "grad_norm": 15.270723342895508, "learning_rate": 1.8225160733416227e-06, "loss": 0.8615419864654541, "step": 362 }, { "ce_loss": 8.37283514556475e-05, "cls_loss": 0.059814453125, "epoch": 0.2174174174174174, "mask_bce_loss": 0.6689139008522034, "mask_dice_loss": 0.17775705456733704, "mask_loss": 0.8466709852218628, "step": 362 }, { "epoch": 0.218018018018018, "grad_norm": 19.474092483520508, "learning_rate": 1.8214081678173094e-06, "loss": 0.7989219427108765, "step": 363 }, { "ce_loss": 7.878137694206089e-05, "cls_loss": 0.06396484375, "epoch": 0.218018018018018, "mask_bce_loss": 0.5175501704216003, "mask_dice_loss": 0.07632488757371902, "mask_loss": 0.5938750505447388, "step": 363 }, { "epoch": 0.21861861861861862, "grad_norm": 20.678300857543945, "learning_rate": 1.8202971540599594e-06, "loss": 0.9371017217636108, "step": 364 }, { "ce_loss": 2.461592885083519e-05, "cls_loss": 0.059326171875, "epoch": 0.21861861861861862, "mask_bce_loss": 0.5166398882865906, "mask_dice_loss": 0.0972733423113823, "mask_loss": 0.6139132380485535, "step": 364 }, { "epoch": 0.21921921921921922, "grad_norm": 19.374710083007812, "learning_rate": 1.8191830362736824e-06, "loss": 0.8513679504394531, "step": 365 }, { "ce_loss": 2.6132755010621622e-05, "cls_loss": 0.0625, "epoch": 0.21921921921921922, "mask_bce_loss": 0.548323929309845, "mask_dice_loss": 0.12085805088281631, "mask_loss": 0.6691820025444031, "step": 365 }, { "epoch": 0.21981981981981982, "grad_norm": 34.599571228027344, "learning_rate": 1.818065818674333e-06, "loss": 0.94530189037323, "step": 366 }, { "ce_loss": 0.00011284211359452456, "cls_loss": 0.0390625, "epoch": 0.21981981981981982, "mask_bce_loss": 0.5612978339195251, "mask_dice_loss": 0.05391070246696472, "mask_loss": 0.6152085065841675, "step": 366 }, { "epoch": 0.22042042042042043, "grad_norm": 25.887435913085938, "learning_rate": 1.8169455054894958e-06, "loss": 0.839824914932251, "step": 367 }, { "ce_loss": 4.829425233765505e-05, "cls_loss": 0.03515625, "epoch": 0.22042042042042043, "mask_bce_loss": 0.6034703850746155, "mask_dice_loss": 0.04744745418429375, "mask_loss": 0.6509178280830383, "step": 367 }, { "epoch": 0.22102102102102103, "grad_norm": 22.36952781677246, "learning_rate": 1.8158221009584698e-06, "loss": 0.7655526399612427, "step": 368 }, { "ce_loss": 2.143060919479467e-05, "cls_loss": 0.03125, "epoch": 0.22102102102102103, "mask_bce_loss": 0.2676992118358612, "mask_dice_loss": 0.04942075535655022, "mask_loss": 0.3171199560165405, "step": 368 }, { "epoch": 0.22162162162162163, "grad_norm": 20.176496505737305, "learning_rate": 1.8146956093322508e-06, "loss": 0.6814196705818176, "step": 369 }, { "ce_loss": 2.3047357899486087e-05, "cls_loss": 0.036376953125, "epoch": 0.22162162162162163, "mask_bce_loss": 0.4363795816898346, "mask_dice_loss": 0.030868351459503174, "mask_loss": 0.46724793314933777, "step": 369 }, { "epoch": 0.2222222222222222, "grad_norm": 53.63217544555664, "learning_rate": 1.813566034873517e-06, "loss": 0.8701708316802979, "step": 370 }, { "ce_loss": 0.001741676707752049, "cls_loss": 0.04833984375, "epoch": 0.2222222222222222, "mask_bce_loss": 0.6182557940483093, "mask_dice_loss": 0.06265222281217575, "mask_loss": 0.6809080243110657, "step": 370 }, { "epoch": 0.2228228228228228, "grad_norm": 25.003700256347656, "learning_rate": 1.8124333818566121e-06, "loss": 0.972842812538147, "step": 371 }, { "ce_loss": 6.077011857996695e-05, "cls_loss": 0.04150390625, "epoch": 0.2228228228228228, "mask_bce_loss": 0.3836922347545624, "mask_dice_loss": 0.056484173983335495, "mask_loss": 0.440176397562027, "step": 371 }, { "epoch": 0.22342342342342342, "grad_norm": 22.72211456298828, "learning_rate": 1.8112976545675287e-06, "loss": 0.8125523328781128, "step": 372 }, { "ce_loss": 3.9899918192531914e-05, "cls_loss": 0.048583984375, "epoch": 0.22342342342342342, "mask_bce_loss": 0.28549104928970337, "mask_dice_loss": 0.035674333572387695, "mask_loss": 0.32116538286209106, "step": 372 }, { "epoch": 0.22402402402402402, "grad_norm": 19.691463470458984, "learning_rate": 1.810158857303893e-06, "loss": 0.8393401503562927, "step": 373 }, { "ce_loss": 4.492112202569842e-05, "cls_loss": 0.06494140625, "epoch": 0.22402402402402402, "mask_bce_loss": 1.4969185590744019, "mask_dice_loss": 0.13304471969604492, "mask_loss": 1.6299632787704468, "step": 373 }, { "epoch": 0.22462462462462462, "grad_norm": 38.7220458984375, "learning_rate": 1.8090169943749474e-06, "loss": 0.9329726696014404, "step": 374 }, { "ce_loss": 2.3604032321600243e-05, "cls_loss": 0.03955078125, "epoch": 0.22462462462462462, "mask_bce_loss": 0.6208832859992981, "mask_dice_loss": 0.031768798828125, "mask_loss": 0.6526520848274231, "step": 374 }, { "epoch": 0.22522522522522523, "grad_norm": 34.08349609375, "learning_rate": 1.8078720701015357e-06, "loss": 1.0883140563964844, "step": 375 }, { "ce_loss": 0.005393756087869406, "cls_loss": 0.04833984375, "epoch": 0.22522522522522523, "mask_bce_loss": 0.4340980052947998, "mask_dice_loss": 0.08612998574972153, "mask_loss": 0.5202279686927795, "step": 375 }, { "epoch": 0.22582582582582583, "grad_norm": 14.193582534790039, "learning_rate": 1.8067240888160852e-06, "loss": 0.7412997484207153, "step": 376 }, { "ce_loss": 5.487970702233724e-05, "cls_loss": 0.05712890625, "epoch": 0.22582582582582583, "mask_bce_loss": 0.35071370005607605, "mask_dice_loss": 0.07419341802597046, "mask_loss": 0.4249071180820465, "step": 376 }, { "epoch": 0.22642642642642644, "grad_norm": 25.848642349243164, "learning_rate": 1.8055730548625919e-06, "loss": 0.7788676619529724, "step": 377 }, { "ce_loss": 0.00015731310122646391, "cls_loss": 0.031005859375, "epoch": 0.22642642642642644, "mask_bce_loss": 0.5298776030540466, "mask_dice_loss": 0.05686146765947342, "mask_loss": 0.5867390632629395, "step": 377 }, { "epoch": 0.22702702702702704, "grad_norm": 22.230024337768555, "learning_rate": 1.8044189725966022e-06, "loss": 0.7612347602844238, "step": 378 }, { "ce_loss": 3.776826270041056e-05, "cls_loss": 0.0654296875, "epoch": 0.22702702702702704, "mask_bce_loss": 1.391499638557434, "mask_dice_loss": 0.1420947015285492, "mask_loss": 1.5335943698883057, "step": 378 }, { "epoch": 0.22762762762762762, "grad_norm": 38.28535079956055, "learning_rate": 1.803261846385198e-06, "loss": 0.907224178314209, "step": 379 }, { "ce_loss": 6.487544305855408e-05, "cls_loss": 0.043212890625, "epoch": 0.22762762762762762, "mask_bce_loss": 0.9222144484519958, "mask_dice_loss": 0.16304226219654083, "mask_loss": 1.0852566957473755, "step": 379 }, { "epoch": 0.22822822822822822, "grad_norm": 27.984703063964844, "learning_rate": 1.8021016806069796e-06, "loss": 0.8675044178962708, "step": 380 }, { "ce_loss": 3.115953950327821e-05, "cls_loss": 0.0439453125, "epoch": 0.22822822822822822, "mask_bce_loss": 0.7046317458152771, "mask_dice_loss": 0.05339245870709419, "mask_loss": 0.7580242156982422, "step": 380 }, { "epoch": 0.22882882882882882, "grad_norm": 24.794729232788086, "learning_rate": 1.8009384796520486e-06, "loss": 0.7865103483200073, "step": 381 }, { "ce_loss": 0.0002153529494535178, "cls_loss": 0.0703125, "epoch": 0.22882882882882882, "mask_bce_loss": 1.7335585355758667, "mask_dice_loss": 0.20694367587566376, "mask_loss": 1.9405021667480469, "step": 381 }, { "epoch": 0.22942942942942943, "grad_norm": 37.81327819824219, "learning_rate": 1.7997722479219925e-06, "loss": 0.8808525800704956, "step": 382 }, { "ce_loss": 2.5440034733037464e-05, "cls_loss": 0.025146484375, "epoch": 0.22942942942942943, "mask_bce_loss": 0.16593873500823975, "mask_dice_loss": 0.016212811693549156, "mask_loss": 0.18215154111385345, "step": 382 }, { "epoch": 0.23003003003003003, "grad_norm": 20.48589515686035, "learning_rate": 1.7986029898298667e-06, "loss": 0.8405716419219971, "step": 383 }, { "ce_loss": 5.1222457841504365e-05, "cls_loss": 0.05419921875, "epoch": 0.23003003003003003, "mask_bce_loss": 0.41240188479423523, "mask_dice_loss": 0.07904287427663803, "mask_loss": 0.49144476652145386, "step": 383 }, { "epoch": 0.23063063063063063, "grad_norm": 18.186813354492188, "learning_rate": 1.797430709800179e-06, "loss": 0.7690770626068115, "step": 384 }, { "ce_loss": 0.0002218482259195298, "cls_loss": 0.142578125, "epoch": 0.23063063063063063, "mask_bce_loss": 1.2886689901351929, "mask_dice_loss": 0.1024039015173912, "mask_loss": 1.3910728693008423, "step": 384 }, { "epoch": 0.23123123123123124, "grad_norm": 33.661216735839844, "learning_rate": 1.7962554122688719e-06, "loss": 0.937052309513092, "step": 385 }, { "ce_loss": 7.1788250352256e-05, "cls_loss": 0.05517578125, "epoch": 0.23123123123123124, "mask_bce_loss": 1.7916887998580933, "mask_dice_loss": 0.10682173818349838, "mask_loss": 1.8985105752944946, "step": 385 }, { "epoch": 0.23183183183183184, "grad_norm": 17.455059051513672, "learning_rate": 1.7950771016833062e-06, "loss": 0.8452134132385254, "step": 386 }, { "ce_loss": 5.72551361983642e-05, "cls_loss": 0.0361328125, "epoch": 0.23183183183183184, "mask_bce_loss": 0.2534703314304352, "mask_dice_loss": 0.02873922698199749, "mask_loss": 0.2822095453739166, "step": 386 }, { "epoch": 0.23243243243243245, "grad_norm": 25.993654251098633, "learning_rate": 1.7938957825022446e-06, "loss": 0.8379316329956055, "step": 387 }, { "ce_loss": 0.00017203186871483922, "cls_loss": 0.05712890625, "epoch": 0.23243243243243245, "mask_bce_loss": 0.7193681597709656, "mask_dice_loss": 0.13405953347682953, "mask_loss": 0.8534277081489563, "step": 387 }, { "epoch": 0.23303303303303302, "grad_norm": 24.658964157104492, "learning_rate": 1.7927114591958344e-06, "loss": 0.904226541519165, "step": 388 }, { "ce_loss": 0.002367011271417141, "cls_loss": 0.06103515625, "epoch": 0.23303303303303302, "mask_bce_loss": 1.3717989921569824, "mask_dice_loss": 0.06917861104011536, "mask_loss": 1.4409775733947754, "step": 388 }, { "epoch": 0.23363363363363363, "grad_norm": 23.348976135253906, "learning_rate": 1.7915241362455896e-06, "loss": 0.7410115599632263, "step": 389 }, { "ce_loss": 5.918391252635047e-05, "cls_loss": 0.041015625, "epoch": 0.23363363363363363, "mask_bce_loss": 0.5856898427009583, "mask_dice_loss": 0.032335441559553146, "mask_loss": 0.6180253028869629, "step": 389 }, { "epoch": 0.23423423423423423, "grad_norm": 28.30428123474121, "learning_rate": 1.790333818144376e-06, "loss": 0.9224505424499512, "step": 390 }, { "ce_loss": 2.2466343580163084e-05, "cls_loss": 0.04248046875, "epoch": 0.23423423423423423, "mask_bce_loss": 0.704942524433136, "mask_dice_loss": 0.07009551674127579, "mask_loss": 0.7750380635261536, "step": 390 }, { "epoch": 0.23483483483483483, "grad_norm": 30.781620025634766, "learning_rate": 1.7891405093963937e-06, "loss": 0.8803932666778564, "step": 391 }, { "ce_loss": 3.892879612976685e-05, "cls_loss": 0.03125, "epoch": 0.23483483483483483, "mask_bce_loss": 0.3780241310596466, "mask_dice_loss": 0.043585002422332764, "mask_loss": 0.42160913348197937, "step": 391 }, { "epoch": 0.23543543543543544, "grad_norm": 26.976144790649414, "learning_rate": 1.7879442145171572e-06, "loss": 0.898232638835907, "step": 392 }, { "ce_loss": 0.0017212954116985202, "cls_loss": 0.03515625, "epoch": 0.23543543543543544, "mask_bce_loss": 0.3115611970424652, "mask_dice_loss": 0.030471378937363625, "mask_loss": 0.3420325815677643, "step": 392 }, { "epoch": 0.23603603603603604, "grad_norm": 30.510265350341797, "learning_rate": 1.7867449380334831e-06, "loss": 0.8038909435272217, "step": 393 }, { "ce_loss": 0.0012209357228130102, "cls_loss": 0.04833984375, "epoch": 0.23603603603603604, "mask_bce_loss": 0.24926818907260895, "mask_dice_loss": 0.06757200509309769, "mask_loss": 0.31684020161628723, "step": 393 }, { "epoch": 0.23663663663663664, "grad_norm": 15.489339828491211, "learning_rate": 1.7855426844834694e-06, "loss": 0.8019137978553772, "step": 394 }, { "ce_loss": 2.9394890589173883e-05, "cls_loss": 0.04833984375, "epoch": 0.23663663663663664, "mask_bce_loss": 0.4607534408569336, "mask_dice_loss": 0.06838985532522202, "mask_loss": 0.5291432738304138, "step": 394 }, { "epoch": 0.23723723723723725, "grad_norm": 35.13844680786133, "learning_rate": 1.7843374584164786e-06, "loss": 0.9625821709632874, "step": 395 }, { "ce_loss": 4.636387893697247e-05, "cls_loss": 0.04931640625, "epoch": 0.23723723723723725, "mask_bce_loss": 0.7047056555747986, "mask_dice_loss": 0.08020691573619843, "mask_loss": 0.7849125862121582, "step": 395 }, { "epoch": 0.23783783783783785, "grad_norm": 25.686569213867188, "learning_rate": 1.7831292643931226e-06, "loss": 0.847233235836029, "step": 396 }, { "ce_loss": 2.123374906659592e-05, "cls_loss": 0.0849609375, "epoch": 0.23783783783783785, "mask_bce_loss": 1.8883074522018433, "mask_dice_loss": 0.11394383758306503, "mask_loss": 2.002251386642456, "step": 396 }, { "epoch": 0.23843843843843843, "grad_norm": 28.626615524291992, "learning_rate": 1.7819181069852442e-06, "loss": 0.8770096302032471, "step": 397 }, { "ce_loss": 0.0003288336156401783, "cls_loss": 0.05859375, "epoch": 0.23843843843843843, "mask_bce_loss": 1.5124586820602417, "mask_dice_loss": 0.14263096451759338, "mask_loss": 1.6550896167755127, "step": 397 }, { "epoch": 0.23903903903903903, "grad_norm": 50.3541374206543, "learning_rate": 1.7807039907758983e-06, "loss": 0.8225818872451782, "step": 398 }, { "ce_loss": 6.250644946703687e-05, "cls_loss": 0.05419921875, "epoch": 0.23903903903903903, "mask_bce_loss": 0.7047428488731384, "mask_dice_loss": 0.06497085094451904, "mask_loss": 0.7697136998176575, "step": 398 }, { "epoch": 0.23963963963963963, "grad_norm": 19.694944381713867, "learning_rate": 1.7794869203593375e-06, "loss": 0.6990551352500916, "step": 399 }, { "ce_loss": 0.0021553724072873592, "cls_loss": 0.06591796875, "epoch": 0.23963963963963963, "mask_bce_loss": 0.8425316214561462, "mask_dice_loss": 0.12904374301433563, "mask_loss": 0.9715753793716431, "step": 399 }, { "epoch": 0.24024024024024024, "grad_norm": 33.36418914794922, "learning_rate": 1.7782669003409922e-06, "loss": 0.8718867301940918, "step": 400 }, { "ce_loss": 0.00016314699314534664, "cls_loss": 0.05615234375, "epoch": 0.24024024024024024, "mask_bce_loss": 0.4629083275794983, "mask_dice_loss": 0.14485982060432434, "mask_loss": 0.607768177986145, "step": 400 }, { "epoch": 0.24084084084084084, "grad_norm": 32.32306671142578, "learning_rate": 1.7770439353374553e-06, "loss": 0.7551383972167969, "step": 401 }, { "ce_loss": 8.090517076198012e-05, "cls_loss": 0.018798828125, "epoch": 0.24084084084084084, "mask_bce_loss": 0.15089157223701477, "mask_dice_loss": 0.011588843539357185, "mask_loss": 0.1624804139137268, "step": 401 }, { "epoch": 0.24144144144144145, "grad_norm": 32.6138916015625, "learning_rate": 1.775818029976462e-06, "loss": 0.905386209487915, "step": 402 }, { "ce_loss": 5.310952474246733e-05, "cls_loss": 0.04833984375, "epoch": 0.24144144144144145, "mask_bce_loss": 0.7146334052085876, "mask_dice_loss": 0.08069848269224167, "mask_loss": 0.7953318953514099, "step": 402 }, { "epoch": 0.24204204204204205, "grad_norm": 27.67317771911621, "learning_rate": 1.7745891888968752e-06, "loss": 0.7704877853393555, "step": 403 }, { "ce_loss": 2.851195677067153e-05, "cls_loss": 0.033203125, "epoch": 0.24204204204204205, "mask_bce_loss": 0.2591123878955841, "mask_dice_loss": 0.025578314438462257, "mask_loss": 0.2846907079219818, "step": 403 }, { "epoch": 0.24264264264264265, "grad_norm": 34.16339874267578, "learning_rate": 1.7733574167486665e-06, "loss": 0.963316798210144, "step": 404 }, { "ce_loss": 0.00018317739886697382, "cls_loss": 0.05615234375, "epoch": 0.24264264264264265, "mask_bce_loss": 0.6825423240661621, "mask_dice_loss": 0.08892562240362167, "mask_loss": 0.771467924118042, "step": 404 }, { "epoch": 0.24324324324324326, "grad_norm": 38.745174407958984, "learning_rate": 1.772122718192898e-06, "loss": 0.9447935819625854, "step": 405 }, { "ce_loss": 0.006939912214875221, "cls_loss": 0.052734375, "epoch": 0.24324324324324326, "mask_bce_loss": 0.34618133306503296, "mask_dice_loss": 0.04258756712079048, "mask_loss": 0.38876891136169434, "step": 405 }, { "epoch": 0.24384384384384383, "grad_norm": 51.344970703125, "learning_rate": 1.770885097901706e-06, "loss": 0.8506380915641785, "step": 406 }, { "ce_loss": 4.6583041694248095e-05, "cls_loss": 0.03955078125, "epoch": 0.24384384384384383, "mask_bce_loss": 0.4619557857513428, "mask_dice_loss": 0.03457896038889885, "mask_loss": 0.4965347349643707, "step": 406 }, { "epoch": 0.24444444444444444, "grad_norm": 48.67683029174805, "learning_rate": 1.7696445605582828e-06, "loss": 0.7551443576812744, "step": 407 }, { "ce_loss": 9.591285197529942e-05, "cls_loss": 0.059326171875, "epoch": 0.24444444444444444, "mask_bce_loss": 0.5554074645042419, "mask_dice_loss": 0.16900716722011566, "mask_loss": 0.7244146466255188, "step": 407 }, { "epoch": 0.24504504504504504, "grad_norm": 22.516599655151367, "learning_rate": 1.768401110856859e-06, "loss": 0.9584536552429199, "step": 408 }, { "ce_loss": 7.254476076923311e-05, "cls_loss": 0.06396484375, "epoch": 0.24504504504504504, "mask_bce_loss": 0.9433483481407166, "mask_dice_loss": 0.11332289129495621, "mask_loss": 1.0566712617874146, "step": 408 }, { "epoch": 0.24564564564564564, "grad_norm": 27.377880096435547, "learning_rate": 1.7671547535026847e-06, "loss": 0.9006087779998779, "step": 409 }, { "ce_loss": 2.922645944636315e-05, "cls_loss": 0.036376953125, "epoch": 0.24564564564564564, "mask_bce_loss": 0.43228575587272644, "mask_dice_loss": 0.04407372698187828, "mask_loss": 0.476359486579895, "step": 409 }, { "epoch": 0.24624624624624625, "grad_norm": 18.89347267150879, "learning_rate": 1.7659054932120143e-06, "loss": 0.6347947120666504, "step": 410 }, { "ce_loss": 3.7303332646843046e-05, "cls_loss": 0.0634765625, "epoch": 0.24624624624624625, "mask_bce_loss": 0.6588711738586426, "mask_dice_loss": 0.12260746955871582, "mask_loss": 0.7814786434173584, "step": 410 }, { "epoch": 0.24684684684684685, "grad_norm": 14.754931449890137, "learning_rate": 1.764653334712085e-06, "loss": 0.8137030005455017, "step": 411 }, { "ce_loss": 0.000727978884242475, "cls_loss": 0.04150390625, "epoch": 0.24684684684684685, "mask_bce_loss": 0.3987937867641449, "mask_dice_loss": 0.07957419008016586, "mask_loss": 0.47836798429489136, "step": 411 }, { "epoch": 0.24744744744744746, "grad_norm": 29.948760986328125, "learning_rate": 1.7633982827411028e-06, "loss": 0.8744723796844482, "step": 412 }, { "ce_loss": 0.0031467669177800417, "cls_loss": 0.0289306640625, "epoch": 0.24744744744744746, "mask_bce_loss": 0.3128868341445923, "mask_dice_loss": 0.02842055633664131, "mask_loss": 0.3413074016571045, "step": 412 }, { "epoch": 0.24804804804804806, "grad_norm": 18.56319808959961, "learning_rate": 1.7621403420482214e-06, "loss": 0.775260865688324, "step": 413 }, { "ce_loss": 7.003999780863523e-05, "cls_loss": 0.059814453125, "epoch": 0.24804804804804806, "mask_bce_loss": 0.5058442950248718, "mask_dice_loss": 0.06066526845097542, "mask_loss": 0.5665095448493958, "step": 413 }, { "epoch": 0.24864864864864866, "grad_norm": 39.24064636230469, "learning_rate": 1.7608795173935263e-06, "loss": 0.9646337032318115, "step": 414 }, { "ce_loss": 0.0009990108665078878, "cls_loss": 0.0458984375, "epoch": 0.24864864864864866, "mask_bce_loss": 0.8297109007835388, "mask_dice_loss": 0.055372532457113266, "mask_loss": 0.8850834369659424, "step": 414 }, { "epoch": 0.24924924924924924, "grad_norm": 23.772777557373047, "learning_rate": 1.7596158135480152e-06, "loss": 0.8645263314247131, "step": 415 }, { "ce_loss": 0.002874740632250905, "cls_loss": 0.056640625, "epoch": 0.24924924924924924, "mask_bce_loss": 0.7515501379966736, "mask_dice_loss": 0.10703437775373459, "mask_loss": 0.8585845232009888, "step": 415 }, { "epoch": 0.24984984984984984, "grad_norm": 21.15030288696289, "learning_rate": 1.7583492352935814e-06, "loss": 0.6850278377532959, "step": 416 }, { "ce_loss": 3.733261110028252e-05, "cls_loss": 0.0498046875, "epoch": 0.24984984984984984, "mask_bce_loss": 0.6105998158454895, "mask_dice_loss": 0.05915714055299759, "mask_loss": 0.6697569489479065, "step": 416 }, { "epoch": 0.25045045045045045, "grad_norm": 23.581750869750977, "learning_rate": 1.7570797874229948e-06, "loss": 0.8239184617996216, "step": 417 }, { "ce_loss": 0.0007149463635869324, "cls_loss": 0.05078125, "epoch": 0.25045045045045045, "mask_bce_loss": 0.3016577661037445, "mask_dice_loss": 0.10487014055252075, "mask_loss": 0.40652790665626526, "step": 417 }, { "epoch": 0.2510510510510511, "grad_norm": 30.99138832092285, "learning_rate": 1.7558074747398841e-06, "loss": 0.9901119470596313, "step": 418 }, { "ce_loss": 3.980739711550996e-05, "cls_loss": 0.034423828125, "epoch": 0.2510510510510511, "mask_bce_loss": 0.16625016927719116, "mask_dice_loss": 0.02969507686793804, "mask_loss": 0.19594524800777435, "step": 418 }, { "epoch": 0.25165165165165165, "grad_norm": 21.655603408813477, "learning_rate": 1.7545323020587186e-06, "loss": 0.7829588055610657, "step": 419 }, { "ce_loss": 1.7519030734547414e-05, "cls_loss": 0.05322265625, "epoch": 0.25165165165165165, "mask_bce_loss": 1.0526859760284424, "mask_dice_loss": 0.07369057834148407, "mask_loss": 1.1263765096664429, "step": 419 }, { "epoch": 0.25225225225225223, "grad_norm": 16.85330581665039, "learning_rate": 1.7532542742047896e-06, "loss": 0.80890953540802, "step": 420 }, { "ce_loss": 1.721754961181432e-05, "cls_loss": 0.06787109375, "epoch": 0.25225225225225223, "mask_bce_loss": 0.46241217851638794, "mask_dice_loss": 0.1167086735367775, "mask_loss": 0.5791208744049072, "step": 420 }, { "epoch": 0.25285285285285286, "grad_norm": 25.218446731567383, "learning_rate": 1.7519733960141924e-06, "loss": 0.8363372087478638, "step": 421 }, { "ce_loss": 3.993058271589689e-05, "cls_loss": 0.032958984375, "epoch": 0.25285285285285286, "mask_bce_loss": 0.2218640297651291, "mask_dice_loss": 0.02732761576771736, "mask_loss": 0.24919164180755615, "step": 421 }, { "epoch": 0.25345345345345344, "grad_norm": 50.43796157836914, "learning_rate": 1.750689672333809e-06, "loss": 0.9458003640174866, "step": 422 }, { "ce_loss": 2.5231745894416235e-05, "cls_loss": 0.046875, "epoch": 0.25345345345345344, "mask_bce_loss": 0.6582987308502197, "mask_dice_loss": 0.08618100732564926, "mask_loss": 0.7444797158241272, "step": 422 }, { "epoch": 0.25405405405405407, "grad_norm": 34.785850524902344, "learning_rate": 1.7494031080212878e-06, "loss": 0.9995620846748352, "step": 423 }, { "ce_loss": 7.764353358652443e-05, "cls_loss": 0.041748046875, "epoch": 0.25405405405405407, "mask_bce_loss": 0.9987382292747498, "mask_dice_loss": 0.08348153531551361, "mask_loss": 1.0822197198867798, "step": 423 }, { "epoch": 0.25465465465465464, "grad_norm": 27.885330200195312, "learning_rate": 1.7481137079450265e-06, "loss": 0.9043608903884888, "step": 424 }, { "ce_loss": 5.007625441066921e-05, "cls_loss": 0.05615234375, "epoch": 0.25465465465465464, "mask_bce_loss": 1.5018930435180664, "mask_dice_loss": 0.16301222145557404, "mask_loss": 1.664905309677124, "step": 424 }, { "epoch": 0.2552552552552553, "grad_norm": 25.942502975463867, "learning_rate": 1.7468214769841537e-06, "loss": 0.7317534685134888, "step": 425 }, { "ce_loss": 4.2232008127029985e-05, "cls_loss": 0.04150390625, "epoch": 0.2552552552552553, "mask_bce_loss": 0.5741642713546753, "mask_dice_loss": 0.06494005024433136, "mask_loss": 0.6391043066978455, "step": 425 }, { "epoch": 0.25585585585585585, "grad_norm": 26.090105056762695, "learning_rate": 1.74552642002851e-06, "loss": 0.90298992395401, "step": 426 }, { "ce_loss": 5.897844675928354e-05, "cls_loss": 0.03662109375, "epoch": 0.25585585585585585, "mask_bce_loss": 0.3746494948863983, "mask_dice_loss": 0.031369712203741074, "mask_loss": 0.4060192108154297, "step": 426 }, { "epoch": 0.2564564564564565, "grad_norm": 34.64594268798828, "learning_rate": 1.744228541978629e-06, "loss": 1.0686815977096558, "step": 427 }, { "ce_loss": 3.523194754961878e-05, "cls_loss": 0.05029296875, "epoch": 0.2564564564564565, "mask_bce_loss": 0.7382963299751282, "mask_dice_loss": 0.1005154401063919, "mask_loss": 0.8388117551803589, "step": 427 }, { "epoch": 0.25705705705705706, "grad_norm": 33.075889587402344, "learning_rate": 1.7429278477457206e-06, "loss": 0.9046100378036499, "step": 428 }, { "ce_loss": 4.3078154703835025e-05, "cls_loss": 0.05712890625, "epoch": 0.25705705705705706, "mask_bce_loss": 0.3139394223690033, "mask_dice_loss": 0.07273398339748383, "mask_loss": 0.3866733908653259, "step": 428 }, { "epoch": 0.25765765765765763, "grad_norm": 34.8673210144043, "learning_rate": 1.7416243422516505e-06, "loss": 0.9131046533584595, "step": 429 }, { "ce_loss": 0.0006987382075749338, "cls_loss": 0.045654296875, "epoch": 0.25765765765765763, "mask_bce_loss": 0.8523152470588684, "mask_dice_loss": 0.05230847746133804, "mask_loss": 0.9046237468719482, "step": 429 }, { "epoch": 0.25825825825825827, "grad_norm": 23.23445701599121, "learning_rate": 1.7403180304289223e-06, "loss": 0.9161455631256104, "step": 430 }, { "ce_loss": 3.7594756577163935e-05, "cls_loss": 0.04833984375, "epoch": 0.25825825825825827, "mask_bce_loss": 1.0938100814819336, "mask_dice_loss": 0.05419331416487694, "mask_loss": 1.148003339767456, "step": 430 }, { "epoch": 0.25885885885885884, "grad_norm": 25.93810272216797, "learning_rate": 1.739008917220659e-06, "loss": 0.8607907295227051, "step": 431 }, { "ce_loss": 2.2695247025694698e-05, "cls_loss": 0.050048828125, "epoch": 0.25885885885885884, "mask_bce_loss": 0.6145442128181458, "mask_dice_loss": 0.04929990693926811, "mask_loss": 0.663844108581543, "step": 431 }, { "epoch": 0.2594594594594595, "grad_norm": 60.02272033691406, "learning_rate": 1.7376970075805843e-06, "loss": 0.984490156173706, "step": 432 }, { "ce_loss": 2.7532048989087343e-05, "cls_loss": 0.05908203125, "epoch": 0.2594594594594595, "mask_bce_loss": 0.15621604025363922, "mask_dice_loss": 0.1764727383852005, "mask_loss": 0.3326887786388397, "step": 432 }, { "epoch": 0.26006006006006005, "grad_norm": 69.90774536132812, "learning_rate": 1.7363823064730027e-06, "loss": 0.8323874473571777, "step": 433 }, { "ce_loss": 2.1536159692914225e-05, "cls_loss": 0.05126953125, "epoch": 0.26006006006006005, "mask_bce_loss": 0.6433700323104858, "mask_dice_loss": 0.16711607575416565, "mask_loss": 0.8104860782623291, "step": 433 }, { "epoch": 0.2606606606606607, "grad_norm": 19.03218650817871, "learning_rate": 1.7350648188727836e-06, "loss": 0.8458631634712219, "step": 434 }, { "ce_loss": 2.8964683224330656e-05, "cls_loss": 0.05419921875, "epoch": 0.2606606606606607, "mask_bce_loss": 0.39922380447387695, "mask_dice_loss": 0.10631636530160904, "mask_loss": 0.5055401921272278, "step": 434 }, { "epoch": 0.26126126126126126, "grad_norm": 20.721132278442383, "learning_rate": 1.7337445497653385e-06, "loss": 0.7523762583732605, "step": 435 }, { "ce_loss": 3.2591029594186693e-05, "cls_loss": 0.04931640625, "epoch": 0.26126126126126126, "mask_bce_loss": 0.36569440364837646, "mask_dice_loss": 0.07386735081672668, "mask_loss": 0.43956175446510315, "step": 435 }, { "epoch": 0.2618618618618619, "grad_norm": 19.41510009765625, "learning_rate": 1.7324215041466053e-06, "loss": 0.8566108345985413, "step": 436 }, { "ce_loss": 9.677310299593955e-05, "cls_loss": 0.052734375, "epoch": 0.2618618618618619, "mask_bce_loss": 1.0289899110794067, "mask_dice_loss": 0.0968971773982048, "mask_loss": 1.1258870363235474, "step": 436 }, { "epoch": 0.26246246246246246, "grad_norm": 20.801061630249023, "learning_rate": 1.7310956870230285e-06, "loss": 0.8165611028671265, "step": 437 }, { "ce_loss": 7.700608693994582e-05, "cls_loss": 0.028564453125, "epoch": 0.26246246246246246, "mask_bce_loss": 0.37725135684013367, "mask_dice_loss": 0.03793811425566673, "mask_loss": 0.4151894748210907, "step": 437 }, { "epoch": 0.26306306306306304, "grad_norm": 13.95311164855957, "learning_rate": 1.7297671034115394e-06, "loss": 0.7697868347167969, "step": 438 }, { "ce_loss": 5.796417099190876e-05, "cls_loss": 0.036376953125, "epoch": 0.26306306306306304, "mask_bce_loss": 0.5579361319541931, "mask_dice_loss": 0.039939459413290024, "mask_loss": 0.5978755950927734, "step": 438 }, { "epoch": 0.26366366366366367, "grad_norm": 40.78001403808594, "learning_rate": 1.728435758339538e-06, "loss": 0.949318528175354, "step": 439 }, { "ce_loss": 3.542276317602955e-05, "cls_loss": 0.059814453125, "epoch": 0.26366366366366367, "mask_bce_loss": 0.3866804540157318, "mask_dice_loss": 0.09312720596790314, "mask_loss": 0.47980767488479614, "step": 439 }, { "epoch": 0.26426426426426425, "grad_norm": 20.826805114746094, "learning_rate": 1.7271016568448739e-06, "loss": 0.8223960399627686, "step": 440 }, { "ce_loss": 0.00027552220853976905, "cls_loss": 0.0458984375, "epoch": 0.26426426426426425, "mask_bce_loss": 0.20938746631145477, "mask_dice_loss": 0.04723597690463066, "mask_loss": 0.25662344694137573, "step": 440 }, { "epoch": 0.2648648648648649, "grad_norm": 29.18491554260254, "learning_rate": 1.7257648039758272e-06, "loss": 0.9437807202339172, "step": 441 }, { "ce_loss": 0.0010651197517290711, "cls_loss": 0.05224609375, "epoch": 0.2648648648648649, "mask_bce_loss": 0.39694806933403015, "mask_dice_loss": 0.05344853922724724, "mask_loss": 0.4503965973854065, "step": 441 }, { "epoch": 0.26546546546546546, "grad_norm": 21.203372955322266, "learning_rate": 1.724425204791089e-06, "loss": 0.7913420796394348, "step": 442 }, { "ce_loss": 5.8761212130775675e-05, "cls_loss": 0.064453125, "epoch": 0.26546546546546546, "mask_bce_loss": 0.9954151511192322, "mask_dice_loss": 0.12747883796691895, "mask_loss": 1.122894048690796, "step": 442 }, { "epoch": 0.2660660660660661, "grad_norm": 18.376136779785156, "learning_rate": 1.723082864359742e-06, "loss": 0.8564457893371582, "step": 443 }, { "ce_loss": 0.00044413661817088723, "cls_loss": 0.04248046875, "epoch": 0.2660660660660661, "mask_bce_loss": 0.46425700187683105, "mask_dice_loss": 0.08445985615253448, "mask_loss": 0.5487168431282043, "step": 443 }, { "epoch": 0.26666666666666666, "grad_norm": 10.28231430053711, "learning_rate": 1.7217377877612435e-06, "loss": 0.7272936701774597, "step": 444 }, { "ce_loss": 3.577738971216604e-05, "cls_loss": 0.0390625, "epoch": 0.26666666666666666, "mask_bce_loss": 0.9414971470832825, "mask_dice_loss": 0.07696977257728577, "mask_loss": 1.0184669494628906, "step": 444 }, { "epoch": 0.2672672672672673, "grad_norm": 37.81431198120117, "learning_rate": 1.7203899800854029e-06, "loss": 0.9617224931716919, "step": 445 }, { "ce_loss": 0.00026625191094353795, "cls_loss": 0.0439453125, "epoch": 0.2672672672672673, "mask_bce_loss": 0.7212640643119812, "mask_dice_loss": 0.07293087989091873, "mask_loss": 0.7941949367523193, "step": 445 }, { "epoch": 0.26786786786786787, "grad_norm": 44.57920837402344, "learning_rate": 1.7190394464323647e-06, "loss": 0.8501483201980591, "step": 446 }, { "ce_loss": 9.367702296003699e-05, "cls_loss": 0.060546875, "epoch": 0.26786786786786787, "mask_bce_loss": 1.6345160007476807, "mask_dice_loss": 0.11493652313947678, "mask_loss": 1.7494524717330933, "step": 446 }, { "epoch": 0.26846846846846845, "grad_norm": 22.82339096069336, "learning_rate": 1.717686191912589e-06, "loss": 0.8387662172317505, "step": 447 }, { "ce_loss": 4.3964329961454496e-05, "cls_loss": 0.061279296875, "epoch": 0.26846846846846845, "mask_bce_loss": 1.9222139120101929, "mask_dice_loss": 0.11959053575992584, "mask_loss": 2.041804552078247, "step": 447 }, { "epoch": 0.2690690690690691, "grad_norm": 30.27185821533203, "learning_rate": 1.7163302216468308e-06, "loss": 0.9542697668075562, "step": 448 }, { "ce_loss": 0.0001251399953616783, "cls_loss": 0.057373046875, "epoch": 0.2690690690690691, "mask_bce_loss": 0.9916836619377136, "mask_dice_loss": 0.10847177356481552, "mask_loss": 1.1001554727554321, "step": 448 }, { "epoch": 0.26966966966966965, "grad_norm": 28.083057403564453, "learning_rate": 1.714971540766122e-06, "loss": 0.7955169677734375, "step": 449 }, { "ce_loss": 8.995180542115122e-05, "cls_loss": 0.04931640625, "epoch": 0.26966966966966965, "mask_bce_loss": 0.31703251600265503, "mask_dice_loss": 0.1961982548236847, "mask_loss": 0.5132308006286621, "step": 449 }, { "epoch": 0.2702702702702703, "grad_norm": 14.961515426635742, "learning_rate": 1.7136101544117522e-06, "loss": 0.7965542078018188, "step": 450 }, { "ce_loss": 2.4326773200300522e-05, "cls_loss": 0.028564453125, "epoch": 0.2702702702702703, "mask_bce_loss": 0.4347716271877289, "mask_dice_loss": 0.022189367562532425, "mask_loss": 0.4569610059261322, "step": 450 }, { "epoch": 0.27087087087087086, "grad_norm": 40.87716293334961, "learning_rate": 1.7122460677352477e-06, "loss": 0.8922286033630371, "step": 451 }, { "ce_loss": 0.00026416112086735666, "cls_loss": 0.0625, "epoch": 0.27087087087087086, "mask_bce_loss": 1.1270149946212769, "mask_dice_loss": 0.1473761945962906, "mask_loss": 1.2743911743164062, "step": 451 }, { "epoch": 0.2714714714714715, "grad_norm": 17.69478416442871, "learning_rate": 1.7108792858983531e-06, "loss": 0.8175550103187561, "step": 452 }, { "ce_loss": 3.867112172883935e-05, "cls_loss": 0.06982421875, "epoch": 0.2714714714714715, "mask_bce_loss": 1.3414878845214844, "mask_dice_loss": 0.10002505779266357, "mask_loss": 1.441512942314148, "step": 452 }, { "epoch": 0.27207207207207207, "grad_norm": 16.062026977539062, "learning_rate": 1.709509814073012e-06, "loss": 0.8471211194992065, "step": 453 }, { "ce_loss": 6.730935274390504e-05, "cls_loss": 0.0888671875, "epoch": 0.27207207207207207, "mask_bce_loss": 1.06965172290802, "mask_dice_loss": 0.19035091996192932, "mask_loss": 1.260002613067627, "step": 453 }, { "epoch": 0.2726726726726727, "grad_norm": 29.4941463470459, "learning_rate": 1.7081376574413464e-06, "loss": 1.0208795070648193, "step": 454 }, { "ce_loss": 0.004019766580313444, "cls_loss": 0.06787109375, "epoch": 0.2726726726726727, "mask_bce_loss": 1.3142882585525513, "mask_dice_loss": 0.15967564284801483, "mask_loss": 1.4739638566970825, "step": 454 }, { "epoch": 0.2732732732732733, "grad_norm": 15.300200462341309, "learning_rate": 1.7067628211956382e-06, "loss": 0.8208324313163757, "step": 455 }, { "ce_loss": 6.597298488486558e-05, "cls_loss": 0.05419921875, "epoch": 0.2732732732732733, "mask_bce_loss": 0.9135918021202087, "mask_dice_loss": 0.08478570729494095, "mask_loss": 0.9983775019645691, "step": 455 }, { "epoch": 0.27387387387387385, "grad_norm": 16.947511672973633, "learning_rate": 1.7053853105383088e-06, "loss": 0.8052096366882324, "step": 456 }, { "ce_loss": 9.407369361724705e-05, "cls_loss": 0.051513671875, "epoch": 0.27387387387387385, "mask_bce_loss": 0.45886141061782837, "mask_dice_loss": 0.07736793905496597, "mask_loss": 0.5362293720245361, "step": 456 }, { "epoch": 0.2744744744744745, "grad_norm": 28.5454158782959, "learning_rate": 1.704005130681899e-06, "loss": 0.9127674698829651, "step": 457 }, { "ce_loss": 3.9854257920524105e-05, "cls_loss": 0.03515625, "epoch": 0.2744744744744745, "mask_bce_loss": 0.5105174779891968, "mask_dice_loss": 0.036579232662916183, "mask_loss": 0.5470967292785645, "step": 457 }, { "epoch": 0.27507507507507506, "grad_norm": 15.208949089050293, "learning_rate": 1.7026222868490518e-06, "loss": 0.7777485251426697, "step": 458 }, { "ce_loss": 0.001076703891158104, "cls_loss": 0.06494140625, "epoch": 0.27507507507507506, "mask_bce_loss": 1.4168938398361206, "mask_dice_loss": 0.10571908950805664, "mask_loss": 1.5226129293441772, "step": 458 }, { "epoch": 0.2756756756756757, "grad_norm": 59.257232666015625, "learning_rate": 1.7012367842724883e-06, "loss": 0.8796564936637878, "step": 459 }, { "ce_loss": 6.497511640191078e-05, "cls_loss": 0.025634765625, "epoch": 0.2756756756756757, "mask_bce_loss": 0.2845490872859955, "mask_dice_loss": 0.017573662102222443, "mask_loss": 0.30212274193763733, "step": 459 }, { "epoch": 0.27627627627627627, "grad_norm": 18.937240600585938, "learning_rate": 1.699848628194993e-06, "loss": 0.7821755409240723, "step": 460 }, { "ce_loss": 0.00021993089467287064, "cls_loss": 0.051513671875, "epoch": 0.27627627627627627, "mask_bce_loss": 1.0219477415084839, "mask_dice_loss": 0.06281565874814987, "mask_loss": 1.0847634077072144, "step": 460 }, { "epoch": 0.2768768768768769, "grad_norm": 27.138620376586914, "learning_rate": 1.6984578238693886e-06, "loss": 0.8629555702209473, "step": 461 }, { "ce_loss": 7.249377813423052e-05, "cls_loss": 0.034912109375, "epoch": 0.2768768768768769, "mask_bce_loss": 0.3161073327064514, "mask_dice_loss": 0.030368639156222343, "mask_loss": 0.3464759588241577, "step": 461 }, { "epoch": 0.2774774774774775, "grad_norm": 19.172887802124023, "learning_rate": 1.6970643765585208e-06, "loss": 0.6909236907958984, "step": 462 }, { "ce_loss": 5.048050661571324e-05, "cls_loss": 0.04833984375, "epoch": 0.2774774774774775, "mask_bce_loss": 0.7295002341270447, "mask_dice_loss": 0.04768509790301323, "mask_loss": 0.777185320854187, "step": 462 }, { "epoch": 0.27807807807807805, "grad_norm": 27.71856117248535, "learning_rate": 1.695668291535236e-06, "loss": 0.9113079905509949, "step": 463 }, { "ce_loss": 4.547228672890924e-05, "cls_loss": 0.05859375, "epoch": 0.27807807807807805, "mask_bce_loss": 0.6437970399856567, "mask_dice_loss": 0.06133991479873657, "mask_loss": 0.7051369547843933, "step": 463 }, { "epoch": 0.2786786786786787, "grad_norm": 49.335289001464844, "learning_rate": 1.6942695740823614e-06, "loss": 0.8650554418563843, "step": 464 }, { "ce_loss": 3.314045534352772e-05, "cls_loss": 0.053955078125, "epoch": 0.2786786786786787, "mask_bce_loss": 0.5648161768913269, "mask_dice_loss": 0.13963882625102997, "mask_loss": 0.7044550180435181, "step": 464 }, { "epoch": 0.27927927927927926, "grad_norm": 26.217247009277344, "learning_rate": 1.692868229492686e-06, "loss": 0.8477643728256226, "step": 465 }, { "ce_loss": 7.026939420029521e-05, "cls_loss": 0.064453125, "epoch": 0.27927927927927926, "mask_bce_loss": 0.7214921712875366, "mask_dice_loss": 0.1349392533302307, "mask_loss": 0.8564314246177673, "step": 465 }, { "epoch": 0.2798798798798799, "grad_norm": 27.53032112121582, "learning_rate": 1.6914642630689386e-06, "loss": 0.7758421301841736, "step": 466 }, { "ce_loss": 3.239160650991835e-05, "cls_loss": 0.03271484375, "epoch": 0.2798798798798799, "mask_bce_loss": 0.3583633601665497, "mask_dice_loss": 0.02372489497065544, "mask_loss": 0.38208824396133423, "step": 466 }, { "epoch": 0.28048048048048047, "grad_norm": 70.48978424072266, "learning_rate": 1.690057680123771e-06, "loss": 0.8909127712249756, "step": 467 }, { "ce_loss": 2.3116001102607697e-05, "cls_loss": 0.03369140625, "epoch": 0.28048048048048047, "mask_bce_loss": 0.2836533486843109, "mask_dice_loss": 0.028406186029314995, "mask_loss": 0.31205952167510986, "step": 467 }, { "epoch": 0.2810810810810811, "grad_norm": 24.295085906982422, "learning_rate": 1.6886484859797347e-06, "loss": 0.8402165770530701, "step": 468 }, { "ce_loss": 0.00011471889592939988, "cls_loss": 0.031494140625, "epoch": 0.2810810810810811, "mask_bce_loss": 0.44002047181129456, "mask_dice_loss": 0.02294677309691906, "mask_loss": 0.46296724677085876, "step": 468 }, { "epoch": 0.2816816816816817, "grad_norm": 33.73031997680664, "learning_rate": 1.6872366859692627e-06, "loss": 0.831180989742279, "step": 469 }, { "ce_loss": 4.2503979784669355e-05, "cls_loss": 0.04443359375, "epoch": 0.2816816816816817, "mask_bce_loss": 0.6631217002868652, "mask_dice_loss": 0.07448118180036545, "mask_loss": 0.7376028895378113, "step": 469 }, { "epoch": 0.2822822822822823, "grad_norm": 18.472881317138672, "learning_rate": 1.6858222854346483e-06, "loss": 0.788265585899353, "step": 470 }, { "ce_loss": 0.0010403729975223541, "cls_loss": 0.05419921875, "epoch": 0.2822822822822823, "mask_bce_loss": 0.9831787943840027, "mask_dice_loss": 0.06533878296613693, "mask_loss": 1.0485175848007202, "step": 470 }, { "epoch": 0.2828828828828829, "grad_norm": 21.582744598388672, "learning_rate": 1.6844052897280245e-06, "loss": 0.8701803684234619, "step": 471 }, { "ce_loss": 3.0382303521037102e-05, "cls_loss": 0.04345703125, "epoch": 0.2828828828828829, "mask_bce_loss": 0.5476503372192383, "mask_dice_loss": 0.04411201551556587, "mask_loss": 0.591762363910675, "step": 471 }, { "epoch": 0.28348348348348346, "grad_norm": 26.958566665649414, "learning_rate": 1.6829857042113468e-06, "loss": 0.8327588438987732, "step": 472 }, { "ce_loss": 6.641379150096327e-05, "cls_loss": 0.051513671875, "epoch": 0.28348348348348346, "mask_bce_loss": 0.7434741258621216, "mask_dice_loss": 0.05392003059387207, "mask_loss": 0.7973941564559937, "step": 472 }, { "epoch": 0.2840840840840841, "grad_norm": 55.73108673095703, "learning_rate": 1.6815635342563685e-06, "loss": 1.0053037405014038, "step": 473 }, { "ce_loss": 0.00016572177992202342, "cls_loss": 0.05810546875, "epoch": 0.2840840840840841, "mask_bce_loss": 1.4594635963439941, "mask_dice_loss": 0.09458845108747482, "mask_loss": 1.5540519952774048, "step": 473 }, { "epoch": 0.28468468468468466, "grad_norm": 29.676664352416992, "learning_rate": 1.680138785244623e-06, "loss": 0.8019285798072815, "step": 474 }, { "ce_loss": 0.0001176888108602725, "cls_loss": 0.0654296875, "epoch": 0.28468468468468466, "mask_bce_loss": 0.10733915865421295, "mask_dice_loss": 0.15640436112880707, "mask_loss": 0.26374351978302, "step": 474 }, { "epoch": 0.2852852852852853, "grad_norm": 36.514461517333984, "learning_rate": 1.6787114625674032e-06, "loss": 0.8803620338439941, "step": 475 }, { "ce_loss": 6.131133704911917e-05, "cls_loss": 0.05859375, "epoch": 0.2852852852852853, "mask_bce_loss": 1.0480413436889648, "mask_dice_loss": 0.15317034721374512, "mask_loss": 1.20121169090271, "step": 475 }, { "epoch": 0.28588588588588587, "grad_norm": 30.91511344909668, "learning_rate": 1.6772815716257411e-06, "loss": 0.908418595790863, "step": 476 }, { "ce_loss": 0.00013844399654772133, "cls_loss": 0.048583984375, "epoch": 0.28588588588588587, "mask_bce_loss": 0.4109920561313629, "mask_dice_loss": 0.05339783430099487, "mask_loss": 0.4643898904323578, "step": 476 }, { "epoch": 0.2864864864864865, "grad_norm": 39.86845397949219, "learning_rate": 1.675849117830386e-06, "loss": 0.8501884937286377, "step": 477 }, { "ce_loss": 3.310839019832201e-05, "cls_loss": 0.06396484375, "epoch": 0.2864864864864865, "mask_bce_loss": 0.6241127848625183, "mask_dice_loss": 0.1275259256362915, "mask_loss": 0.7516387104988098, "step": 477 }, { "epoch": 0.2870870870870871, "grad_norm": 21.55253791809082, "learning_rate": 1.674414106601787e-06, "loss": 0.8022564649581909, "step": 478 }, { "ce_loss": 0.0001592804619576782, "cls_loss": 0.036865234375, "epoch": 0.2870870870870871, "mask_bce_loss": 0.40608975291252136, "mask_dice_loss": 0.029810979962348938, "mask_loss": 0.4359007477760315, "step": 478 }, { "epoch": 0.2876876876876877, "grad_norm": 15.48462963104248, "learning_rate": 1.6729765433700683e-06, "loss": 0.7799710631370544, "step": 479 }, { "ce_loss": 2.6089417588082142e-05, "cls_loss": 0.06591796875, "epoch": 0.2876876876876877, "mask_bce_loss": 1.6623550653457642, "mask_dice_loss": 0.1073891744017601, "mask_loss": 1.7697442770004272, "step": 479 }, { "epoch": 0.2882882882882883, "grad_norm": 30.73182487487793, "learning_rate": 1.6715364335750125e-06, "loss": 0.8922914266586304, "step": 480 }, { "ce_loss": 1.8666434698388912e-05, "cls_loss": 0.039306640625, "epoch": 0.2882882882882883, "mask_bce_loss": 0.16449670493602753, "mask_dice_loss": 0.0337107852101326, "mask_loss": 0.19820749759674072, "step": 480 }, { "epoch": 0.28888888888888886, "grad_norm": 22.97128677368164, "learning_rate": 1.670093782666038e-06, "loss": 0.8615198135375977, "step": 481 }, { "ce_loss": 0.023326942697167397, "cls_loss": 0.0693359375, "epoch": 0.28888888888888886, "mask_bce_loss": 0.8905672430992126, "mask_dice_loss": 0.07950624078512192, "mask_loss": 0.9700734615325928, "step": 481 }, { "epoch": 0.2894894894894895, "grad_norm": 39.142608642578125, "learning_rate": 1.6686485961021794e-06, "loss": 0.7987855672836304, "step": 482 }, { "ce_loss": 0.016140708699822426, "cls_loss": 0.03857421875, "epoch": 0.2894894894894895, "mask_bce_loss": 0.40341687202453613, "mask_dice_loss": 0.04342649132013321, "mask_loss": 0.44684335589408875, "step": 482 }, { "epoch": 0.29009009009009007, "grad_norm": 42.72518539428711, "learning_rate": 1.6672008793520652e-06, "loss": 0.9087438583374023, "step": 483 }, { "ce_loss": 9.858921112027019e-05, "cls_loss": 0.0625, "epoch": 0.29009009009009007, "mask_bce_loss": 0.7191857695579529, "mask_dice_loss": 0.12433546036481857, "mask_loss": 0.843521237373352, "step": 483 }, { "epoch": 0.2906906906906907, "grad_norm": 26.96681785583496, "learning_rate": 1.6657506378938993e-06, "loss": 0.8868430852890015, "step": 484 }, { "ce_loss": 8.223476470448077e-05, "cls_loss": 0.07763671875, "epoch": 0.2906906906906907, "mask_bce_loss": 0.9667006731033325, "mask_dice_loss": 0.07696101814508438, "mask_loss": 1.0436617136001587, "step": 484 }, { "epoch": 0.2912912912912913, "grad_norm": 32.99191665649414, "learning_rate": 1.6642978772154383e-06, "loss": 0.9307250380516052, "step": 485 }, { "ce_loss": 9.349464380647987e-05, "cls_loss": 0.0296630859375, "epoch": 0.2912912912912913, "mask_bce_loss": 0.38628193736076355, "mask_dice_loss": 0.025242919102311134, "mask_loss": 0.41152486205101013, "step": 485 }, { "epoch": 0.2918918918918919, "grad_norm": 24.081117630004883, "learning_rate": 1.662842602813972e-06, "loss": 0.9865436553955078, "step": 486 }, { "ce_loss": 3.8036912883399054e-05, "cls_loss": 0.04833984375, "epoch": 0.2918918918918919, "mask_bce_loss": 0.7659689784049988, "mask_dice_loss": 0.09554006904363632, "mask_loss": 0.8615090250968933, "step": 486 }, { "epoch": 0.2924924924924925, "grad_norm": 30.898588180541992, "learning_rate": 1.6613848201963025e-06, "loss": 0.9321368932723999, "step": 487 }, { "ce_loss": 4.408407039591111e-05, "cls_loss": 0.05908203125, "epoch": 0.2924924924924925, "mask_bce_loss": 1.4767318964004517, "mask_dice_loss": 0.09721138328313828, "mask_loss": 1.5739432573318481, "step": 487 }, { "epoch": 0.2930930930930931, "grad_norm": 27.100860595703125, "learning_rate": 1.6599245348787228e-06, "loss": 0.810324490070343, "step": 488 }, { "ce_loss": 0.00035581650445237756, "cls_loss": 0.042236328125, "epoch": 0.2930930930930931, "mask_bce_loss": 0.7204774022102356, "mask_dice_loss": 0.06270336359739304, "mask_loss": 0.7831807732582092, "step": 488 }, { "epoch": 0.2936936936936937, "grad_norm": 23.064048767089844, "learning_rate": 1.6584617523869952e-06, "loss": 0.8839197158813477, "step": 489 }, { "ce_loss": 7.651899795746431e-05, "cls_loss": 0.04833984375, "epoch": 0.2936936936936937, "mask_bce_loss": 0.4368910491466522, "mask_dice_loss": 0.02526441588997841, "mask_loss": 0.46215546131134033, "step": 489 }, { "epoch": 0.29429429429429427, "grad_norm": 25.89473533630371, "learning_rate": 1.6569964782563337e-06, "loss": 0.8336597681045532, "step": 490 }, { "ce_loss": 3.273358379374258e-05, "cls_loss": 0.04052734375, "epoch": 0.29429429429429427, "mask_bce_loss": 0.35733747482299805, "mask_dice_loss": 0.03650181367993355, "mask_loss": 0.3938392996788025, "step": 490 }, { "epoch": 0.2948948948948949, "grad_norm": 15.206025123596191, "learning_rate": 1.6555287180313786e-06, "loss": 0.9283562898635864, "step": 491 }, { "ce_loss": 8.099801925709471e-05, "cls_loss": 0.05712890625, "epoch": 0.2948948948948949, "mask_bce_loss": 0.45966121554374695, "mask_dice_loss": 0.059203166514635086, "mask_loss": 0.5188643932342529, "step": 491 }, { "epoch": 0.2954954954954955, "grad_norm": 38.466064453125, "learning_rate": 1.6540584772661786e-06, "loss": 0.9015997648239136, "step": 492 }, { "ce_loss": 0.00016405220958404243, "cls_loss": 0.052734375, "epoch": 0.2954954954954955, "mask_bce_loss": 0.7980033755302429, "mask_dice_loss": 0.08265519887208939, "mask_loss": 0.8806585669517517, "step": 492 }, { "epoch": 0.2960960960960961, "grad_norm": 22.551156997680664, "learning_rate": 1.6525857615241685e-06, "loss": 0.9583300352096558, "step": 493 }, { "ce_loss": 2.5515053494018503e-05, "cls_loss": 0.056640625, "epoch": 0.2960960960960961, "mask_bce_loss": 0.740120530128479, "mask_dice_loss": 0.07967057079076767, "mask_loss": 0.8197910785675049, "step": 493 }, { "epoch": 0.2966966966966967, "grad_norm": 41.917903900146484, "learning_rate": 1.651110576378149e-06, "loss": 0.8409905433654785, "step": 494 }, { "ce_loss": 0.00015851345960982144, "cls_loss": 0.03857421875, "epoch": 0.2966966966966967, "mask_bce_loss": 0.36684349179267883, "mask_dice_loss": 0.03756866231560707, "mask_loss": 0.4044121503829956, "step": 494 }, { "epoch": 0.2972972972972973, "grad_norm": 40.02617645263672, "learning_rate": 1.649632927410264e-06, "loss": 0.8355262279510498, "step": 495 }, { "ce_loss": 0.00037585696554742754, "cls_loss": 0.0498046875, "epoch": 0.2972972972972973, "mask_bce_loss": 0.5832442045211792, "mask_dice_loss": 0.09935019165277481, "mask_loss": 0.6825944185256958, "step": 495 }, { "epoch": 0.2978978978978979, "grad_norm": 45.7310905456543, "learning_rate": 1.6481528202119824e-06, "loss": 1.102698802947998, "step": 496 }, { "ce_loss": 4.09053718612995e-05, "cls_loss": 0.031494140625, "epoch": 0.2978978978978979, "mask_bce_loss": 0.3013244569301605, "mask_dice_loss": 0.02261173725128174, "mask_loss": 0.32393619418144226, "step": 496 }, { "epoch": 0.2984984984984985, "grad_norm": 26.843610763549805, "learning_rate": 1.6466702603840733e-06, "loss": 0.7932410836219788, "step": 497 }, { "ce_loss": 5.4601117881247774e-05, "cls_loss": 0.0250244140625, "epoch": 0.2984984984984985, "mask_bce_loss": 0.2234831601381302, "mask_dice_loss": 0.01768561638891697, "mask_loss": 0.2411687821149826, "step": 497 }, { "epoch": 0.2990990990990991, "grad_norm": 15.82733154296875, "learning_rate": 1.6451852535365872e-06, "loss": 0.7240886688232422, "step": 498 }, { "ce_loss": 0.00011932018242077902, "cls_loss": 0.053955078125, "epoch": 0.2990990990990991, "mask_bce_loss": 0.7288092970848083, "mask_dice_loss": 0.13813379406929016, "mask_loss": 0.8669431209564209, "step": 498 }, { "epoch": 0.2996996996996997, "grad_norm": 16.227025985717773, "learning_rate": 1.643697805288835e-06, "loss": 0.8374532461166382, "step": 499 }, { "ce_loss": 3.186220055795275e-05, "cls_loss": 0.03466796875, "epoch": 0.2996996996996997, "mask_bce_loss": 0.7428895235061646, "mask_dice_loss": 0.04796099290251732, "mask_loss": 0.7908505201339722, "step": 499 }, { "epoch": 0.3003003003003003, "grad_norm": 22.602970123291016, "learning_rate": 1.6422079212693651e-06, "loss": 0.8179476857185364, "step": 500 }, { "ce_loss": 0.0009011829970404506, "cls_loss": 0.04736328125, "epoch": 0.3003003003003003, "mask_bce_loss": 0.6484220027923584, "mask_dice_loss": 0.08701559156179428, "mask_loss": 0.7354375720024109, "step": 500 }, { "epoch": 0.3009009009009009, "grad_norm": 20.283037185668945, "learning_rate": 1.640715607115943e-06, "loss": 0.8728958368301392, "step": 501 }, { "ce_loss": 5.231178874964826e-05, "cls_loss": 0.06298828125, "epoch": 0.3009009009009009, "mask_bce_loss": 1.200227975845337, "mask_dice_loss": 0.09833282232284546, "mask_loss": 1.2985608577728271, "step": 501 }, { "epoch": 0.3015015015015015, "grad_norm": 22.172740936279297, "learning_rate": 1.6392208684755296e-06, "loss": 0.9666433334350586, "step": 502 }, { "ce_loss": 0.00011305549560347572, "cls_loss": 0.03662109375, "epoch": 0.3015015015015015, "mask_bce_loss": 0.18691249191761017, "mask_dice_loss": 0.031023407354950905, "mask_loss": 0.21793590486049652, "step": 502 }, { "epoch": 0.3021021021021021, "grad_norm": 23.620019912719727, "learning_rate": 1.6377237110042614e-06, "loss": 0.8783830404281616, "step": 503 }, { "ce_loss": 5.030901593272574e-05, "cls_loss": 0.04296875, "epoch": 0.3021021021021021, "mask_bce_loss": 0.501758337020874, "mask_dice_loss": 0.07926111668348312, "mask_loss": 0.5810194611549377, "step": 503 }, { "epoch": 0.3027027027027027, "grad_norm": 22.58134651184082, "learning_rate": 1.6362241403674264e-06, "loss": 0.9939875602722168, "step": 504 }, { "ce_loss": 0.01249963603913784, "cls_loss": 0.0546875, "epoch": 0.3027027027027027, "mask_bce_loss": 0.5465947985649109, "mask_dice_loss": 0.09672916680574417, "mask_loss": 0.6433239579200745, "step": 504 }, { "epoch": 0.3033033033033033, "grad_norm": 15.74026107788086, "learning_rate": 1.634722162239445e-06, "loss": 0.8042969703674316, "step": 505 }, { "ce_loss": 4.9705966375768185e-05, "cls_loss": 0.04150390625, "epoch": 0.3033033033033033, "mask_bce_loss": 0.6577520370483398, "mask_dice_loss": 0.034696221351623535, "mask_loss": 0.6924482583999634, "step": 505 }, { "epoch": 0.3039039039039039, "grad_norm": 28.59585952758789, "learning_rate": 1.6332177823038468e-06, "loss": 0.9217078685760498, "step": 506 }, { "ce_loss": 3.908454891643487e-05, "cls_loss": 0.061279296875, "epoch": 0.3039039039039039, "mask_bce_loss": 0.8928335309028625, "mask_dice_loss": 0.11524175852537155, "mask_loss": 1.00807523727417, "step": 506 }, { "epoch": 0.3045045045045045, "grad_norm": 19.601985931396484, "learning_rate": 1.6317110062532509e-06, "loss": 0.8549994230270386, "step": 507 }, { "ce_loss": 9.548009256832302e-05, "cls_loss": 0.05224609375, "epoch": 0.3045045045045045, "mask_bce_loss": 0.7851627469062805, "mask_dice_loss": 0.09087537974119186, "mask_loss": 0.876038134098053, "step": 507 }, { "epoch": 0.3051051051051051, "grad_norm": 14.7960786819458, "learning_rate": 1.6302018397893427e-06, "loss": 0.8452387452125549, "step": 508 }, { "ce_loss": 0.00022477433958556503, "cls_loss": 0.050048828125, "epoch": 0.3051051051051051, "mask_bce_loss": 0.41131392121315, "mask_dice_loss": 0.05595485121011734, "mask_loss": 0.46726876497268677, "step": 508 }, { "epoch": 0.3057057057057057, "grad_norm": 23.251522064208984, "learning_rate": 1.6286902886228533e-06, "loss": 0.8324176073074341, "step": 509 }, { "ce_loss": 6.917075370438397e-05, "cls_loss": 0.046142578125, "epoch": 0.3057057057057057, "mask_bce_loss": 0.5295606255531311, "mask_dice_loss": 0.04781407117843628, "mask_loss": 0.5773746967315674, "step": 509 }, { "epoch": 0.3063063063063063, "grad_norm": 35.199012756347656, "learning_rate": 1.627176358473537e-06, "loss": 0.8777340054512024, "step": 510 }, { "ce_loss": 0.000150578809552826, "cls_loss": 0.05078125, "epoch": 0.3063063063063063, "mask_bce_loss": 0.4730374813079834, "mask_dice_loss": 0.07556948810815811, "mask_loss": 0.5486069917678833, "step": 510 }, { "epoch": 0.3069069069069069, "grad_norm": 28.51448631286621, "learning_rate": 1.6256600550701508e-06, "loss": 0.8022491931915283, "step": 511 }, { "ce_loss": 0.00013779501023236662, "cls_loss": 0.0380859375, "epoch": 0.3069069069069069, "mask_bce_loss": 0.35656413435935974, "mask_dice_loss": 0.033050913363695145, "mask_loss": 0.3896150588989258, "step": 511 }, { "epoch": 0.3075075075075075, "grad_norm": 37.455955505371094, "learning_rate": 1.624141384150433e-06, "loss": 0.7887868881225586, "step": 512 }, { "ce_loss": 3.302454570075497e-05, "cls_loss": 0.0380859375, "epoch": 0.3075075075075075, "mask_bce_loss": 0.24099834263324738, "mask_dice_loss": 0.032934047281742096, "mask_loss": 0.27393239736557007, "step": 512 }, { "epoch": 0.3081081081081081, "grad_norm": 54.52518081665039, "learning_rate": 1.6226203514610783e-06, "loss": 0.833780825138092, "step": 513 }, { "ce_loss": 0.0012753470800817013, "cls_loss": 0.0673828125, "epoch": 0.3081081081081081, "mask_bce_loss": 0.8541211485862732, "mask_dice_loss": 0.13918213546276093, "mask_loss": 0.9933032989501953, "step": 513 }, { "epoch": 0.3087087087087087, "grad_norm": 22.415584564208984, "learning_rate": 1.6210969627577212e-06, "loss": 0.8287939429283142, "step": 514 }, { "ce_loss": 4.3949610699201e-05, "cls_loss": 0.048095703125, "epoch": 0.3087087087087087, "mask_bce_loss": 0.9722272157669067, "mask_dice_loss": 0.050862859934568405, "mask_loss": 1.023090124130249, "step": 514 }, { "epoch": 0.30930930930930933, "grad_norm": 49.13716506958008, "learning_rate": 1.6195712238049088e-06, "loss": 0.8283877968788147, "step": 515 }, { "ce_loss": 0.00013087934348732233, "cls_loss": 0.0673828125, "epoch": 0.30930930930930933, "mask_bce_loss": 1.0608863830566406, "mask_dice_loss": 0.13421666622161865, "mask_loss": 1.1951030492782593, "step": 515 }, { "epoch": 0.3099099099099099, "grad_norm": 14.483683586120605, "learning_rate": 1.6180431403760835e-06, "loss": 0.667913556098938, "step": 516 }, { "ce_loss": 2.9561706469394267e-05, "cls_loss": 0.06103515625, "epoch": 0.3099099099099099, "mask_bce_loss": 1.125489354133606, "mask_dice_loss": 0.11364485323429108, "mask_loss": 1.2391341924667358, "step": 516 }, { "epoch": 0.3105105105105105, "grad_norm": 20.083290100097656, "learning_rate": 1.6165127182535585e-06, "loss": 0.750308632850647, "step": 517 }, { "ce_loss": 0.000681667122989893, "cls_loss": 0.06396484375, "epoch": 0.3105105105105105, "mask_bce_loss": 0.9884328246116638, "mask_dice_loss": 0.10293982177972794, "mask_loss": 1.0913726091384888, "step": 517 }, { "epoch": 0.3111111111111111, "grad_norm": 28.65480613708496, "learning_rate": 1.614979963228497e-06, "loss": 0.88535076379776, "step": 518 }, { "ce_loss": 9.443172893952578e-05, "cls_loss": 0.05224609375, "epoch": 0.3111111111111111, "mask_bce_loss": 1.0099300146102905, "mask_dice_loss": 0.1223694235086441, "mask_loss": 1.1322994232177734, "step": 518 }, { "epoch": 0.3117117117117117, "grad_norm": 19.77787971496582, "learning_rate": 1.6134448811008894e-06, "loss": 0.8709837794303894, "step": 519 }, { "ce_loss": 2.3590557248098776e-05, "cls_loss": 0.05859375, "epoch": 0.3117117117117117, "mask_bce_loss": 0.42565950751304626, "mask_dice_loss": 0.08028631657361984, "mask_loss": 0.5059458017349243, "step": 519 }, { "epoch": 0.3123123123123123, "grad_norm": 16.2099552154541, "learning_rate": 1.6119074776795327e-06, "loss": 0.6989619135856628, "step": 520 }, { "ce_loss": 4.447246828931384e-05, "cls_loss": 0.027587890625, "epoch": 0.3123123123123123, "mask_bce_loss": 0.17713889479637146, "mask_dice_loss": 0.019737547263503075, "mask_loss": 0.1968764364719391, "step": 520 }, { "epoch": 0.3129129129129129, "grad_norm": 30.828807830810547, "learning_rate": 1.6103677587820075e-06, "loss": 0.8340659141540527, "step": 521 }, { "ce_loss": 6.175487214932218e-05, "cls_loss": 0.052001953125, "epoch": 0.3129129129129129, "mask_bce_loss": 2.022521495819092, "mask_dice_loss": 0.14306288957595825, "mask_loss": 2.1655843257904053, "step": 521 }, { "epoch": 0.31351351351351353, "grad_norm": 23.36058235168457, "learning_rate": 1.6088257302346553e-06, "loss": 0.72088623046875, "step": 522 }, { "ce_loss": 4.7122881369432434e-05, "cls_loss": 0.0546875, "epoch": 0.31351351351351353, "mask_bce_loss": 0.7327452898025513, "mask_dice_loss": 0.12130896002054214, "mask_loss": 0.8540542721748352, "step": 522 }, { "epoch": 0.3141141141141141, "grad_norm": 39.8271598815918, "learning_rate": 1.6072813978725587e-06, "loss": 0.8982866406440735, "step": 523 }, { "ce_loss": 5.1119535783072934e-05, "cls_loss": 0.047607421875, "epoch": 0.3141141141141141, "mask_bce_loss": 0.8188932538032532, "mask_dice_loss": 0.11781438440084457, "mask_loss": 0.936707615852356, "step": 523 }, { "epoch": 0.31471471471471474, "grad_norm": 28.64552116394043, "learning_rate": 1.6057347675395176e-06, "loss": 0.823167085647583, "step": 524 }, { "ce_loss": 7.694781379541382e-05, "cls_loss": 0.055419921875, "epoch": 0.31471471471471474, "mask_bce_loss": 1.2085202932357788, "mask_dice_loss": 0.06982219964265823, "mask_loss": 1.2783424854278564, "step": 524 }, { "epoch": 0.3153153153153153, "grad_norm": 36.44322204589844, "learning_rate": 1.6041858450880273e-06, "loss": 0.8816688656806946, "step": 525 }, { "ce_loss": 0.00017286518414039165, "cls_loss": 0.040283203125, "epoch": 0.3153153153153153, "mask_bce_loss": 0.35223492980003357, "mask_dice_loss": 0.06567788124084473, "mask_loss": 0.4179128110408783, "step": 525 }, { "epoch": 0.3159159159159159, "grad_norm": 31.66288948059082, "learning_rate": 1.6026346363792564e-06, "loss": 1.00554621219635, "step": 526 }, { "ce_loss": 0.00013005358050577343, "cls_loss": 0.056640625, "epoch": 0.3159159159159159, "mask_bce_loss": 1.332563042640686, "mask_dice_loss": 0.0652877613902092, "mask_loss": 1.397850751876831, "step": 526 }, { "epoch": 0.3165165165165165, "grad_norm": 29.056819915771484, "learning_rate": 1.601081147283025e-06, "loss": 0.7246330976486206, "step": 527 }, { "ce_loss": 2.4316463168361224e-05, "cls_loss": 0.03564453125, "epoch": 0.3165165165165165, "mask_bce_loss": 0.4281015396118164, "mask_dice_loss": 0.033351488411426544, "mask_loss": 0.46145302057266235, "step": 527 }, { "epoch": 0.3171171171171171, "grad_norm": 22.429649353027344, "learning_rate": 1.5995253836777826e-06, "loss": 0.6525899767875671, "step": 528 }, { "ce_loss": 0.00015399084077216685, "cls_loss": 0.05224609375, "epoch": 0.3171171171171171, "mask_bce_loss": 0.7535882592201233, "mask_dice_loss": 0.05327742174267769, "mask_loss": 0.8068656921386719, "step": 528 }, { "epoch": 0.31771771771771773, "grad_norm": 58.0653076171875, "learning_rate": 1.5979673514505847e-06, "loss": 1.0608080625534058, "step": 529 }, { "ce_loss": 2.4594939532107674e-05, "cls_loss": 0.05029296875, "epoch": 0.31771771771771773, "mask_bce_loss": 0.6916816234588623, "mask_dice_loss": 0.08622698485851288, "mask_loss": 0.7779086232185364, "step": 529 }, { "epoch": 0.3183183183183183, "grad_norm": 15.51457405090332, "learning_rate": 1.5964070564970718e-06, "loss": 0.7749120593070984, "step": 530 }, { "ce_loss": 4.434380025486462e-05, "cls_loss": 0.04541015625, "epoch": 0.3183183183183183, "mask_bce_loss": 0.6104758977890015, "mask_dice_loss": 0.05146367475390434, "mask_loss": 0.6619395613670349, "step": 530 }, { "epoch": 0.31891891891891894, "grad_norm": 20.18614959716797, "learning_rate": 1.594844504721447e-06, "loss": 0.782856285572052, "step": 531 }, { "ce_loss": 8.1610371125862e-05, "cls_loss": 0.05419921875, "epoch": 0.31891891891891894, "mask_bce_loss": 0.8204337954521179, "mask_dice_loss": 0.07389296591281891, "mask_loss": 0.8943267464637756, "step": 531 }, { "epoch": 0.3195195195195195, "grad_norm": 18.895639419555664, "learning_rate": 1.5932797020364524e-06, "loss": 0.7883816957473755, "step": 532 }, { "ce_loss": 0.00023534357023891062, "cls_loss": 0.036376953125, "epoch": 0.3195195195195195, "mask_bce_loss": 0.5604948401451111, "mask_dice_loss": 0.05469219759106636, "mask_loss": 0.6151870489120483, "step": 532 }, { "epoch": 0.32012012012012014, "grad_norm": 52.45560836791992, "learning_rate": 1.5917126543633484e-06, "loss": 0.9187787771224976, "step": 533 }, { "ce_loss": 7.44555436540395e-05, "cls_loss": 0.05078125, "epoch": 0.32012012012012014, "mask_bce_loss": 0.34372222423553467, "mask_dice_loss": 0.08159079402685165, "mask_loss": 0.4253130257129669, "step": 533 }, { "epoch": 0.3207207207207207, "grad_norm": 28.63956642150879, "learning_rate": 1.59014336763189e-06, "loss": 0.7906644344329834, "step": 534 }, { "ce_loss": 9.692905587144196e-05, "cls_loss": 0.037353515625, "epoch": 0.3207207207207207, "mask_bce_loss": 0.6857088208198547, "mask_dice_loss": 0.07716932147741318, "mask_loss": 0.7628781199455261, "step": 534 }, { "epoch": 0.3213213213213213, "grad_norm": 23.49483299255371, "learning_rate": 1.5885718477803054e-06, "loss": 0.7150397300720215, "step": 535 }, { "ce_loss": 3.775547884288244e-05, "cls_loss": 0.04150390625, "epoch": 0.3213213213213213, "mask_bce_loss": 0.4999710023403168, "mask_dice_loss": 0.043448254466056824, "mask_loss": 0.5434192419052124, "step": 535 }, { "epoch": 0.3219219219219219, "grad_norm": 22.933866500854492, "learning_rate": 1.5869981007552725e-06, "loss": 0.7319062948226929, "step": 536 }, { "ce_loss": 4.299514330341481e-05, "cls_loss": 0.04052734375, "epoch": 0.3219219219219219, "mask_bce_loss": 0.8358597159385681, "mask_dice_loss": 0.06742936372756958, "mask_loss": 0.9032890796661377, "step": 536 }, { "epoch": 0.3225225225225225, "grad_norm": 44.977149963378906, "learning_rate": 1.5854221325118967e-06, "loss": 0.8334075212478638, "step": 537 }, { "ce_loss": 9.571358532411978e-05, "cls_loss": 0.0625, "epoch": 0.3225225225225225, "mask_bce_loss": 1.6577433347702026, "mask_dice_loss": 0.09037186205387115, "mask_loss": 1.7481151819229126, "step": 537 }, { "epoch": 0.32312312312312313, "grad_norm": 20.6094913482666, "learning_rate": 1.5838439490136892e-06, "loss": 0.8482764363288879, "step": 538 }, { "ce_loss": 8.773932495387271e-05, "cls_loss": 0.055419921875, "epoch": 0.32312312312312313, "mask_bce_loss": 0.5521535277366638, "mask_dice_loss": 0.08199631422758102, "mask_loss": 0.6341498494148254, "step": 538 }, { "epoch": 0.3237237237237237, "grad_norm": 15.827376365661621, "learning_rate": 1.5822635562325432e-06, "loss": 0.7979594469070435, "step": 539 }, { "ce_loss": 0.00017127511091530323, "cls_loss": 0.055419921875, "epoch": 0.3237237237237237, "mask_bce_loss": 0.40757033228874207, "mask_dice_loss": 0.051550425589084625, "mask_loss": 0.4591207504272461, "step": 539 }, { "epoch": 0.32432432432432434, "grad_norm": 22.22112464904785, "learning_rate": 1.5806809601487127e-06, "loss": 0.7169085741043091, "step": 540 }, { "ce_loss": 2.7018322725780308e-05, "cls_loss": 0.05224609375, "epoch": 0.32432432432432434, "mask_bce_loss": 0.43051862716674805, "mask_dice_loss": 0.1361866593360901, "mask_loss": 0.5667052865028381, "step": 540 }, { "epoch": 0.3249249249249249, "grad_norm": 48.711830139160156, "learning_rate": 1.579096166750788e-06, "loss": 0.7901633977890015, "step": 541 }, { "ce_loss": 5.842104656039737e-05, "cls_loss": 0.06689453125, "epoch": 0.3249249249249249, "mask_bce_loss": 0.722732424736023, "mask_dice_loss": 0.055180106312036514, "mask_loss": 0.7779125571250916, "step": 541 }, { "epoch": 0.32552552552552555, "grad_norm": 36.07096481323242, "learning_rate": 1.5775091820356751e-06, "loss": 0.875673234462738, "step": 542 }, { "ce_loss": 4.502920637605712e-05, "cls_loss": 0.055908203125, "epoch": 0.32552552552552555, "mask_bce_loss": 0.978080689907074, "mask_dice_loss": 0.08528228104114532, "mask_loss": 1.063362956047058, "step": 542 }, { "epoch": 0.3261261261261261, "grad_norm": 18.016130447387695, "learning_rate": 1.5759200120085713e-06, "loss": 0.8714385032653809, "step": 543 }, { "ce_loss": 2.3278013031813316e-05, "cls_loss": 0.052734375, "epoch": 0.3261261261261261, "mask_bce_loss": 0.495339959859848, "mask_dice_loss": 0.1283424347639084, "mask_loss": 0.6236823797225952, "step": 543 }, { "epoch": 0.3267267267267267, "grad_norm": 40.45136642456055, "learning_rate": 1.5743286626829435e-06, "loss": 0.8934125900268555, "step": 544 }, { "ce_loss": 0.0008080240804702044, "cls_loss": 0.03759765625, "epoch": 0.3267267267267267, "mask_bce_loss": 0.811870276927948, "mask_dice_loss": 0.11062993854284286, "mask_loss": 0.9225001931190491, "step": 544 }, { "epoch": 0.32732732732732733, "grad_norm": 31.17477798461914, "learning_rate": 1.5727351400805052e-06, "loss": 0.9171512722969055, "step": 545 }, { "ce_loss": 2.1406865926110186e-05, "cls_loss": 0.03173828125, "epoch": 0.32732732732732733, "mask_bce_loss": 0.25417816638946533, "mask_dice_loss": 0.02475813589990139, "mask_loss": 0.2789362967014313, "step": 545 }, { "epoch": 0.3279279279279279, "grad_norm": 24.847379684448242, "learning_rate": 1.5711394502311932e-06, "loss": 0.8047362565994263, "step": 546 }, { "ce_loss": 3.208657290088013e-05, "cls_loss": 0.06591796875, "epoch": 0.3279279279279279, "mask_bce_loss": 0.4535464942455292, "mask_dice_loss": 0.14490272104740143, "mask_loss": 0.5984492301940918, "step": 546 }, { "epoch": 0.32852852852852854, "grad_norm": 32.37005615234375, "learning_rate": 1.5695415991731458e-06, "loss": 0.9479286670684814, "step": 547 }, { "ce_loss": 2.991813926200848e-05, "cls_loss": 0.060546875, "epoch": 0.32852852852852854, "mask_bce_loss": 0.646368145942688, "mask_dice_loss": 0.07009104639291763, "mask_loss": 0.7164592146873474, "step": 547 }, { "epoch": 0.3291291291291291, "grad_norm": 22.499399185180664, "learning_rate": 1.5679415929526787e-06, "loss": 0.753691554069519, "step": 548 }, { "ce_loss": 5.053954737377353e-05, "cls_loss": 0.053466796875, "epoch": 0.3291291291291291, "mask_bce_loss": 0.5082824230194092, "mask_dice_loss": 0.11607612669467926, "mask_loss": 0.6243585348129272, "step": 548 }, { "epoch": 0.32972972972972975, "grad_norm": 33.80604934692383, "learning_rate": 1.566339437624264e-06, "loss": 0.8036681413650513, "step": 549 }, { "ce_loss": 2.244541610707529e-05, "cls_loss": 0.05712890625, "epoch": 0.32972972972972975, "mask_bce_loss": 0.6045295596122742, "mask_dice_loss": 0.13037538528442383, "mask_loss": 0.734904944896698, "step": 549 }, { "epoch": 0.3303303303303303, "grad_norm": 32.22311782836914, "learning_rate": 1.5647351392505046e-06, "loss": 0.9381461143493652, "step": 550 }, { "ce_loss": 6.315539212664589e-05, "cls_loss": 0.043701171875, "epoch": 0.3303303303303303, "mask_bce_loss": 0.915301501750946, "mask_dice_loss": 0.12201489508152008, "mask_loss": 1.0373164415359497, "step": 550 }, { "epoch": 0.33093093093093096, "grad_norm": 62.94209289550781, "learning_rate": 1.563128703902114e-06, "loss": 0.7398525476455688, "step": 551 }, { "ce_loss": 3.6465167795540765e-05, "cls_loss": 0.09130859375, "epoch": 0.33093093093093096, "mask_bce_loss": 0.708172619342804, "mask_dice_loss": 0.0785866230726242, "mask_loss": 0.7867592573165894, "step": 551 }, { "epoch": 0.33153153153153153, "grad_norm": 39.24970626831055, "learning_rate": 1.561520137657891e-06, "loss": 0.9635295867919922, "step": 552 }, { "ce_loss": 3.910356099368073e-05, "cls_loss": 0.06201171875, "epoch": 0.33153153153153153, "mask_bce_loss": 0.7241570949554443, "mask_dice_loss": 0.08152692019939423, "mask_loss": 0.8056840300559998, "step": 552 }, { "epoch": 0.3321321321321321, "grad_norm": 26.98660659790039, "learning_rate": 1.5599094466046986e-06, "loss": 0.9000906944274902, "step": 553 }, { "ce_loss": 0.00012501700257416815, "cls_loss": 0.0625, "epoch": 0.3321321321321321, "mask_bce_loss": 1.2460083961486816, "mask_dice_loss": 0.07381515204906464, "mask_loss": 1.3198235034942627, "step": 553 }, { "epoch": 0.33273273273273274, "grad_norm": 23.131526947021484, "learning_rate": 1.5582966368374397e-06, "loss": 0.8063367009162903, "step": 554 }, { "ce_loss": 3.647449193522334e-05, "cls_loss": 0.05517578125, "epoch": 0.33273273273273274, "mask_bce_loss": 0.3051700294017792, "mask_dice_loss": 0.06720183044672012, "mask_loss": 0.3723718523979187, "step": 554 }, { "epoch": 0.3333333333333333, "grad_norm": 32.01723861694336, "learning_rate": 1.5566817144590347e-06, "loss": 0.9075714349746704, "step": 555 }, { "ce_loss": 5.8104302297579125e-05, "cls_loss": 0.07470703125, "epoch": 0.3333333333333333, "mask_bce_loss": 1.317726492881775, "mask_dice_loss": 0.11700525134801865, "mask_loss": 1.4347317218780518, "step": 555 }, { "epoch": 0.33393393393393395, "grad_norm": 26.561880111694336, "learning_rate": 1.5550646855803985e-06, "loss": 0.7910491228103638, "step": 556 }, { "ce_loss": 6.650367140537128e-05, "cls_loss": 0.03369140625, "epoch": 0.33393393393393395, "mask_bce_loss": 0.23410344123840332, "mask_dice_loss": 0.029768425971269608, "mask_loss": 0.2638718783855438, "step": 556 }, { "epoch": 0.3345345345345345, "grad_norm": 27.9754581451416, "learning_rate": 1.553445556320416e-06, "loss": 0.8037314414978027, "step": 557 }, { "ce_loss": 5.151627919985913e-05, "cls_loss": 0.06396484375, "epoch": 0.3345345345345345, "mask_bce_loss": 0.6564624905586243, "mask_dice_loss": 0.10156196355819702, "mask_loss": 0.7580244541168213, "step": 557 }, { "epoch": 0.33513513513513515, "grad_norm": 39.61222839355469, "learning_rate": 1.5518243328059208e-06, "loss": 0.8189167976379395, "step": 558 }, { "ce_loss": 4.678596815210767e-05, "cls_loss": 0.0634765625, "epoch": 0.33513513513513515, "mask_bce_loss": 0.5623964667320251, "mask_dice_loss": 0.06052042171359062, "mask_loss": 0.6229168772697449, "step": 558 }, { "epoch": 0.33573573573573573, "grad_norm": 44.371849060058594, "learning_rate": 1.5502010211716713e-06, "loss": 0.8231154680252075, "step": 559 }, { "ce_loss": 0.00014395875041373074, "cls_loss": 0.06396484375, "epoch": 0.33573573573573573, "mask_bce_loss": 1.3364671468734741, "mask_dice_loss": 0.1556628942489624, "mask_loss": 1.4921300411224365, "step": 559 }, { "epoch": 0.33633633633633636, "grad_norm": 19.849637985229492, "learning_rate": 1.5485756275603276e-06, "loss": 0.7426413893699646, "step": 560 }, { "ce_loss": 3.518857556628063e-05, "cls_loss": 0.061767578125, "epoch": 0.33633633633633636, "mask_bce_loss": 0.9486652612686157, "mask_dice_loss": 0.08121062815189362, "mask_loss": 1.0298758745193481, "step": 560 }, { "epoch": 0.33693693693693694, "grad_norm": 28.363271713256836, "learning_rate": 1.5469481581224271e-06, "loss": 0.9865392446517944, "step": 561 }, { "ce_loss": 5.2225332183297724e-05, "cls_loss": 0.0546875, "epoch": 0.33693693693693694, "mask_bce_loss": 1.3714176416397095, "mask_dice_loss": 0.0715736374258995, "mask_loss": 1.4429912567138672, "step": 561 }, { "epoch": 0.3375375375375375, "grad_norm": 20.251466751098633, "learning_rate": 1.5453186190163625e-06, "loss": 0.8451699018478394, "step": 562 }, { "ce_loss": 5.573339876718819e-05, "cls_loss": 0.04296875, "epoch": 0.3375375375375375, "mask_bce_loss": 1.1940631866455078, "mask_dice_loss": 0.0713348314166069, "mask_loss": 1.2653980255126953, "step": 562 }, { "epoch": 0.33813813813813814, "grad_norm": 35.34019470214844, "learning_rate": 1.5436870164083589e-06, "loss": 0.8818879127502441, "step": 563 }, { "ce_loss": 3.336633017170243e-05, "cls_loss": 0.0216064453125, "epoch": 0.33813813813813814, "mask_bce_loss": 0.18212689459323883, "mask_dice_loss": 0.013439424335956573, "mask_loss": 0.195566326379776, "step": 563 }, { "epoch": 0.3387387387387387, "grad_norm": 28.41703987121582, "learning_rate": 1.5420533564724494e-06, "loss": 0.8602814078330994, "step": 564 }, { "ce_loss": 2.435508940834552e-05, "cls_loss": 0.037109375, "epoch": 0.3387387387387387, "mask_bce_loss": 0.4005964696407318, "mask_dice_loss": 0.05410044267773628, "mask_loss": 0.454696923494339, "step": 564 }, { "epoch": 0.33933933933933935, "grad_norm": 17.935232162475586, "learning_rate": 1.5404176453904515e-06, "loss": 0.8504316806793213, "step": 565 }, { "ce_loss": 3.5231591027695686e-05, "cls_loss": 0.11962890625, "epoch": 0.33933933933933935, "mask_bce_loss": 0.903509259223938, "mask_dice_loss": 0.07803443819284439, "mask_loss": 0.9815437197685242, "step": 565 }, { "epoch": 0.33993993993993993, "grad_norm": 30.60167121887207, "learning_rate": 1.5387798893519454e-06, "loss": 0.9825742244720459, "step": 566 }, { "ce_loss": 0.00014112949429545552, "cls_loss": 0.06591796875, "epoch": 0.33993993993993993, "mask_bce_loss": 1.014258623123169, "mask_dice_loss": 0.1392832100391388, "mask_loss": 1.1535418033599854, "step": 566 }, { "epoch": 0.34054054054054056, "grad_norm": 24.505001068115234, "learning_rate": 1.5371400945542482e-06, "loss": 0.6717413067817688, "step": 567 }, { "ce_loss": 0.00014255956921260804, "cls_loss": 0.051513671875, "epoch": 0.34054054054054056, "mask_bce_loss": 0.6934412717819214, "mask_dice_loss": 0.0711129680275917, "mask_loss": 0.7645542621612549, "step": 567 }, { "epoch": 0.34114114114114114, "grad_norm": 30.478618621826172, "learning_rate": 1.5354982672023931e-06, "loss": 0.8695802092552185, "step": 568 }, { "ce_loss": 0.00036557583371177316, "cls_loss": 0.06103515625, "epoch": 0.34114114114114114, "mask_bce_loss": 0.8601900935173035, "mask_dice_loss": 0.12298436462879181, "mask_loss": 0.9831744432449341, "step": 568 }, { "epoch": 0.34174174174174177, "grad_norm": 76.24569702148438, "learning_rate": 1.5338544135091032e-06, "loss": 0.9841409921646118, "step": 569 }, { "ce_loss": 4.990880552213639e-05, "cls_loss": 0.06005859375, "epoch": 0.34174174174174177, "mask_bce_loss": 1.061835765838623, "mask_dice_loss": 0.1183672770857811, "mask_loss": 1.1802030801773071, "step": 569 }, { "epoch": 0.34234234234234234, "grad_norm": 20.06853485107422, "learning_rate": 1.5322085396947704e-06, "loss": 0.8263274431228638, "step": 570 }, { "ce_loss": 2.964820851047989e-05, "cls_loss": 0.04736328125, "epoch": 0.34234234234234234, "mask_bce_loss": 0.4188990294933319, "mask_dice_loss": 0.05144960805773735, "mask_loss": 0.47034862637519836, "step": 570 }, { "epoch": 0.3429429429429429, "grad_norm": 25.97765350341797, "learning_rate": 1.5305606519874296e-06, "loss": 0.8566622734069824, "step": 571 }, { "ce_loss": 5.992890874040313e-05, "cls_loss": 0.0634765625, "epoch": 0.3429429429429429, "mask_bce_loss": 1.239444375038147, "mask_dice_loss": 0.10433147102594376, "mask_loss": 1.3437758684158325, "step": 571 }, { "epoch": 0.34354354354354355, "grad_norm": 20.61661148071289, "learning_rate": 1.5289107566227381e-06, "loss": 0.8480437397956848, "step": 572 }, { "ce_loss": 3.2125280995387584e-05, "cls_loss": 0.041748046875, "epoch": 0.34354354354354355, "mask_bce_loss": 0.43066149950027466, "mask_dice_loss": 0.05946092680096626, "mask_loss": 0.4901224374771118, "step": 572 }, { "epoch": 0.3441441441441441, "grad_norm": 18.027719497680664, "learning_rate": 1.527258859843949e-06, "loss": 0.7659430503845215, "step": 573 }, { "ce_loss": 0.000583730754442513, "cls_loss": 0.059326171875, "epoch": 0.3441441441441441, "mask_bce_loss": 2.8152570724487305, "mask_dice_loss": 0.09665381163358688, "mask_loss": 2.9119107723236084, "step": 573 }, { "epoch": 0.34474474474474476, "grad_norm": 60.10403823852539, "learning_rate": 1.5256049679018885e-06, "loss": 0.9042614698410034, "step": 574 }, { "ce_loss": 5.4832009482197464e-05, "cls_loss": 0.05419921875, "epoch": 0.34474474474474476, "mask_bce_loss": 1.2410593032836914, "mask_dice_loss": 0.12301065027713776, "mask_loss": 1.364069938659668, "step": 574 }, { "epoch": 0.34534534534534533, "grad_norm": 20.457094192504883, "learning_rate": 1.5239490870549335e-06, "loss": 0.7808780670166016, "step": 575 }, { "ce_loss": 0.00016079841589089483, "cls_loss": 0.0390625, "epoch": 0.34534534534534533, "mask_bce_loss": 0.6214520931243896, "mask_dice_loss": 0.06278517097234726, "mask_loss": 0.6842372417449951, "step": 575 }, { "epoch": 0.34594594594594597, "grad_norm": 23.873374938964844, "learning_rate": 1.5222912235689866e-06, "loss": 0.7947945594787598, "step": 576 }, { "ce_loss": 0.00020750670228153467, "cls_loss": 0.0419921875, "epoch": 0.34594594594594597, "mask_bce_loss": 0.24903996288776398, "mask_dice_loss": 0.052793752402067184, "mask_loss": 0.30183371901512146, "step": 576 }, { "epoch": 0.34654654654654654, "grad_norm": 45.52773666381836, "learning_rate": 1.5206313837174527e-06, "loss": 0.7889147400856018, "step": 577 }, { "ce_loss": 8.287528180517256e-05, "cls_loss": 0.045654296875, "epoch": 0.34654654654654654, "mask_bce_loss": 0.5101816654205322, "mask_dice_loss": 0.11780857294797897, "mask_loss": 0.6279902458190918, "step": 577 }, { "epoch": 0.3471471471471472, "grad_norm": 25.7515811920166, "learning_rate": 1.5189695737812151e-06, "loss": 0.9355249404907227, "step": 578 }, { "ce_loss": 4.884295049123466e-05, "cls_loss": 0.06689453125, "epoch": 0.3471471471471472, "mask_bce_loss": 1.660616159439087, "mask_dice_loss": 0.10819008201360703, "mask_loss": 1.7688062191009521, "step": 578 }, { "epoch": 0.34774774774774775, "grad_norm": 19.94342613220215, "learning_rate": 1.5173058000486124e-06, "loss": 0.7883056402206421, "step": 579 }, { "ce_loss": 2.890488576667849e-05, "cls_loss": 0.05517578125, "epoch": 0.34774774774774775, "mask_bce_loss": 0.8360010981559753, "mask_dice_loss": 0.09619271010160446, "mask_loss": 0.9321938157081604, "step": 579 }, { "epoch": 0.3483483483483483, "grad_norm": 33.682411193847656, "learning_rate": 1.5156400688154145e-06, "loss": 0.8990561366081238, "step": 580 }, { "ce_loss": 0.00012437847908586264, "cls_loss": 0.05810546875, "epoch": 0.3483483483483483, "mask_bce_loss": 0.6034532189369202, "mask_dice_loss": 0.07489269226789474, "mask_loss": 0.6783459186553955, "step": 580 }, { "epoch": 0.34894894894894896, "grad_norm": 21.77681541442871, "learning_rate": 1.5139723863847976e-06, "loss": 0.8008059859275818, "step": 581 }, { "ce_loss": 3.0454599254881032e-05, "cls_loss": 0.061279296875, "epoch": 0.34894894894894896, "mask_bce_loss": 0.3432963192462921, "mask_dice_loss": 0.08609116077423096, "mask_loss": 0.42938748002052307, "step": 581 }, { "epoch": 0.34954954954954953, "grad_norm": 21.207996368408203, "learning_rate": 1.5123027590673218e-06, "loss": 0.7945948839187622, "step": 582 }, { "ce_loss": 0.03083653748035431, "cls_loss": 0.047607421875, "epoch": 0.34954954954954953, "mask_bce_loss": 0.5646308660507202, "mask_dice_loss": 0.06910844892263412, "mask_loss": 0.6337392926216125, "step": 582 }, { "epoch": 0.35015015015015016, "grad_norm": 27.079397201538086, "learning_rate": 1.5106311931809067e-06, "loss": 0.8366100192070007, "step": 583 }, { "ce_loss": 0.00034614416654221714, "cls_loss": 0.03173828125, "epoch": 0.35015015015015016, "mask_bce_loss": 0.3084556460380554, "mask_dice_loss": 0.024442952126264572, "mask_loss": 0.3328985869884491, "step": 583 }, { "epoch": 0.35075075075075074, "grad_norm": 31.359416961669922, "learning_rate": 1.508957695050808e-06, "loss": 0.8752367496490479, "step": 584 }, { "ce_loss": 5.2749470341950655e-05, "cls_loss": 0.0693359375, "epoch": 0.35075075075075074, "mask_bce_loss": 0.45071926712989807, "mask_dice_loss": 0.08632278442382812, "mask_loss": 0.5370420217514038, "step": 584 }, { "epoch": 0.35135135135135137, "grad_norm": 49.839359283447266, "learning_rate": 1.5072822710095928e-06, "loss": 0.9422351121902466, "step": 585 }, { "ce_loss": 5.650360253639519e-05, "cls_loss": 0.03564453125, "epoch": 0.35135135135135137, "mask_bce_loss": 0.41680532693862915, "mask_dice_loss": 0.0753629207611084, "mask_loss": 0.49216824769973755, "step": 585 }, { "epoch": 0.35195195195195195, "grad_norm": 66.70286560058594, "learning_rate": 1.5056049273971147e-06, "loss": 0.9321378469467163, "step": 586 }, { "ce_loss": 7.287271728273481e-05, "cls_loss": 0.039794921875, "epoch": 0.35195195195195195, "mask_bce_loss": 0.9398218393325806, "mask_dice_loss": 0.11026713997125626, "mask_loss": 1.0500890016555786, "step": 586 }, { "epoch": 0.3525525525525526, "grad_norm": 15.844221115112305, "learning_rate": 1.5039256705604926e-06, "loss": 0.8095748424530029, "step": 587 }, { "ce_loss": 3.175854726578109e-05, "cls_loss": 0.04833984375, "epoch": 0.3525525525525526, "mask_bce_loss": 0.33358439803123474, "mask_dice_loss": 0.06990190595388412, "mask_loss": 0.40348631143569946, "step": 587 }, { "epoch": 0.35315315315315315, "grad_norm": 54.14293670654297, "learning_rate": 1.5022445068540846e-06, "loss": 0.9232004284858704, "step": 588 }, { "ce_loss": 3.009987085533794e-05, "cls_loss": 0.055419921875, "epoch": 0.35315315315315315, "mask_bce_loss": 0.23579159379005432, "mask_dice_loss": 0.06434424221515656, "mask_loss": 0.30013585090637207, "step": 588 }, { "epoch": 0.35375375375375373, "grad_norm": 35.74336624145508, "learning_rate": 1.5005614426394641e-06, "loss": 0.9442768096923828, "step": 589 }, { "ce_loss": 3.210664726793766e-05, "cls_loss": 0.05029296875, "epoch": 0.35375375375375373, "mask_bce_loss": 0.5763837099075317, "mask_dice_loss": 0.11992885172367096, "mask_loss": 0.6963125467300415, "step": 589 }, { "epoch": 0.35435435435435436, "grad_norm": 19.511537551879883, "learning_rate": 1.4988764842853967e-06, "loss": 0.7869775295257568, "step": 590 }, { "ce_loss": 2.5705172447487712e-05, "cls_loss": 0.10009765625, "epoch": 0.35435435435435436, "mask_bce_loss": 0.46286606788635254, "mask_dice_loss": 0.0845421776175499, "mask_loss": 0.5474082231521606, "step": 590 }, { "epoch": 0.35495495495495494, "grad_norm": 42.36270523071289, "learning_rate": 1.4971896381678152e-06, "loss": 0.7830350399017334, "step": 591 }, { "ce_loss": 6.521039176732302e-05, "cls_loss": 0.053466796875, "epoch": 0.35495495495495494, "mask_bce_loss": 0.6398840546607971, "mask_dice_loss": 0.08113150298595428, "mask_loss": 0.7210155725479126, "step": 591 }, { "epoch": 0.35555555555555557, "grad_norm": 35.62019348144531, "learning_rate": 1.4955009106697947e-06, "loss": 0.8448269367218018, "step": 592 }, { "ce_loss": 8.922501729102805e-05, "cls_loss": 0.0546875, "epoch": 0.35555555555555557, "mask_bce_loss": 0.597217857837677, "mask_dice_loss": 0.08298342674970627, "mask_loss": 0.6802012920379639, "step": 592 }, { "epoch": 0.35615615615615615, "grad_norm": 57.26443099975586, "learning_rate": 1.4938103081815313e-06, "loss": 0.9229112863540649, "step": 593 }, { "ce_loss": 2.3393697119900025e-05, "cls_loss": 0.0546875, "epoch": 0.35615615615615615, "mask_bce_loss": 0.5882713198661804, "mask_dice_loss": 0.10506217926740646, "mask_loss": 0.6933335065841675, "step": 593 }, { "epoch": 0.3567567567567568, "grad_norm": 22.661645889282227, "learning_rate": 1.4921178371003155e-06, "loss": 0.8568211793899536, "step": 594 }, { "ce_loss": 9.353577479487285e-05, "cls_loss": 0.0308837890625, "epoch": 0.3567567567567568, "mask_bce_loss": 0.21446184813976288, "mask_dice_loss": 0.021959995850920677, "mask_loss": 0.2364218384027481, "step": 594 }, { "epoch": 0.35735735735735735, "grad_norm": 39.526363372802734, "learning_rate": 1.4904235038305082e-06, "loss": 0.8324247598648071, "step": 595 }, { "ce_loss": 3.395287421881221e-05, "cls_loss": 0.06787109375, "epoch": 0.35735735735735735, "mask_bce_loss": 0.5026091933250427, "mask_dice_loss": 0.13306894898414612, "mask_loss": 0.6356781721115112, "step": 595 }, { "epoch": 0.357957957957958, "grad_norm": 26.252588272094727, "learning_rate": 1.4887273147835167e-06, "loss": 0.8966396450996399, "step": 596 }, { "ce_loss": 6.696295167785138e-05, "cls_loss": 0.0673828125, "epoch": 0.357957957957958, "mask_bce_loss": 0.9164884686470032, "mask_dice_loss": 0.10949063301086426, "mask_loss": 1.0259790420532227, "step": 596 }, { "epoch": 0.35855855855855856, "grad_norm": 23.743574142456055, "learning_rate": 1.4870292763777714e-06, "loss": 0.7546733021736145, "step": 597 }, { "ce_loss": 3.512360854074359e-05, "cls_loss": 0.052001953125, "epoch": 0.35855855855855856, "mask_bce_loss": 0.8482257127761841, "mask_dice_loss": 0.06730590015649796, "mask_loss": 0.9155316352844238, "step": 597 }, { "epoch": 0.35915915915915914, "grad_norm": 29.851579666137695, "learning_rate": 1.4853293950387003e-06, "loss": 0.8384482264518738, "step": 598 }, { "ce_loss": 0.00018822273705154657, "cls_loss": 0.0390625, "epoch": 0.35915915915915914, "mask_bce_loss": 0.33209341764450073, "mask_dice_loss": 0.031412526965141296, "mask_loss": 0.3635059595108032, "step": 598 }, { "epoch": 0.35975975975975977, "grad_norm": 27.38129997253418, "learning_rate": 1.483627677198705e-06, "loss": 0.7614625692367554, "step": 599 }, { "ce_loss": 6.924058106960729e-05, "cls_loss": 0.05517578125, "epoch": 0.35975975975975977, "mask_bce_loss": 0.9049690365791321, "mask_dice_loss": 0.10696699470281601, "mask_loss": 1.011936068534851, "step": 599 }, { "epoch": 0.36036036036036034, "grad_norm": 20.91666603088379, "learning_rate": 1.481924129297137e-06, "loss": 0.7956207990646362, "step": 600 }, { "ce_loss": 3.075775020988658e-05, "cls_loss": 0.03173828125, "epoch": 0.36036036036036034, "mask_bce_loss": 0.6928605437278748, "mask_dice_loss": 0.04349662736058235, "mask_loss": 0.7363571524620056, "step": 600 }, { "epoch": 0.360960960960961, "grad_norm": 45.585269927978516, "learning_rate": 1.480218757780272e-06, "loss": 0.8441762328147888, "step": 601 }, { "ce_loss": 5.136221807333641e-05, "cls_loss": 0.03466796875, "epoch": 0.360960960960961, "mask_bce_loss": 0.35477861762046814, "mask_dice_loss": 0.05158982425928116, "mask_loss": 0.4063684344291687, "step": 601 }, { "epoch": 0.36156156156156155, "grad_norm": 62.65250778198242, "learning_rate": 1.4785115691012863e-06, "loss": 0.8700617551803589, "step": 602 }, { "ce_loss": 4.963417086401023e-05, "cls_loss": 0.0673828125, "epoch": 0.36156156156156155, "mask_bce_loss": 0.38667672872543335, "mask_dice_loss": 0.043595146387815475, "mask_loss": 0.43027186393737793, "step": 602 }, { "epoch": 0.3621621621621622, "grad_norm": 45.0212287902832, "learning_rate": 1.4768025697202341e-06, "loss": 0.9240483045578003, "step": 603 }, { "ce_loss": 0.0002441323595121503, "cls_loss": 0.055419921875, "epoch": 0.3621621621621622, "mask_bce_loss": 0.9649333953857422, "mask_dice_loss": 0.09385309368371964, "mask_loss": 1.0587865114212036, "step": 603 }, { "epoch": 0.36276276276276276, "grad_norm": 38.026885986328125, "learning_rate": 1.4750917661040195e-06, "loss": 0.9369242787361145, "step": 604 }, { "ce_loss": 5.402108945418149e-05, "cls_loss": 0.041259765625, "epoch": 0.36276276276276276, "mask_bce_loss": 0.25101980566978455, "mask_dice_loss": 0.05871669203042984, "mask_loss": 0.3097364902496338, "step": 604 }, { "epoch": 0.3633633633633634, "grad_norm": 26.45413589477539, "learning_rate": 1.4733791647263741e-06, "loss": 0.7831135988235474, "step": 605 }, { "ce_loss": 0.00018124523921869695, "cls_loss": 0.06005859375, "epoch": 0.3633633633633634, "mask_bce_loss": 0.5662277340888977, "mask_dice_loss": 0.07388270646333694, "mask_loss": 0.640110433101654, "step": 605 }, { "epoch": 0.36396396396396397, "grad_norm": 22.27405548095703, "learning_rate": 1.4716647720678329e-06, "loss": 0.8387880325317383, "step": 606 }, { "ce_loss": 8.226490899687633e-05, "cls_loss": 0.057373046875, "epoch": 0.36396396396396397, "mask_bce_loss": 0.7156448364257812, "mask_dice_loss": 0.10186602920293808, "mask_loss": 0.8175108432769775, "step": 606 }, { "epoch": 0.36456456456456454, "grad_norm": 15.445514678955078, "learning_rate": 1.469948594615709e-06, "loss": 0.6809194087982178, "step": 607 }, { "ce_loss": 0.00016597218927927315, "cls_loss": 0.06201171875, "epoch": 0.36456456456456454, "mask_bce_loss": 0.7771168351173401, "mask_dice_loss": 0.1062227413058281, "mask_loss": 0.8833395838737488, "step": 607 }, { "epoch": 0.3651651651651652, "grad_norm": 29.867713928222656, "learning_rate": 1.4682306388640693e-06, "loss": 0.9324671030044556, "step": 608 }, { "ce_loss": 0.00013745055184699595, "cls_loss": 0.061767578125, "epoch": 0.3651651651651652, "mask_bce_loss": 1.9684990644454956, "mask_dice_loss": 0.08314909785985947, "mask_loss": 2.0516481399536133, "step": 608 }, { "epoch": 0.36576576576576575, "grad_norm": 34.15983963012695, "learning_rate": 1.4665109113137094e-06, "loss": 0.9520032405853271, "step": 609 }, { "ce_loss": 0.009687747806310654, "cls_loss": 0.03857421875, "epoch": 0.36576576576576575, "mask_bce_loss": 0.8390003442764282, "mask_dice_loss": 0.06872052699327469, "mask_loss": 0.9077208638191223, "step": 609 }, { "epoch": 0.3663663663663664, "grad_norm": 24.513500213623047, "learning_rate": 1.4647894184721297e-06, "loss": 0.8893556594848633, "step": 610 }, { "ce_loss": 3.896245834766887e-05, "cls_loss": 0.028076171875, "epoch": 0.3663663663663664, "mask_bce_loss": 0.4258343279361725, "mask_dice_loss": 0.023775490000844002, "mask_loss": 0.44960981607437134, "step": 610 }, { "epoch": 0.36696696696696696, "grad_norm": 19.2001953125, "learning_rate": 1.4630661668535105e-06, "loss": 0.7468442916870117, "step": 611 }, { "ce_loss": 3.9491173083661124e-05, "cls_loss": 0.0625, "epoch": 0.36696696696696696, "mask_bce_loss": 0.5749711394309998, "mask_dice_loss": 0.09497248381376266, "mask_loss": 0.669943630695343, "step": 611 }, { "epoch": 0.3675675675675676, "grad_norm": 22.527450561523438, "learning_rate": 1.4613411629786878e-06, "loss": 0.9352952241897583, "step": 612 }, { "ce_loss": 6.031653902027756e-05, "cls_loss": 0.05615234375, "epoch": 0.3675675675675676, "mask_bce_loss": 0.7466955184936523, "mask_dice_loss": 0.12274205684661865, "mask_loss": 0.869437575340271, "step": 612 }, { "epoch": 0.36816816816816816, "grad_norm": 20.530527114868164, "learning_rate": 1.4596144133751274e-06, "loss": 0.9181609749794006, "step": 613 }, { "ce_loss": 2.3103082639863715e-05, "cls_loss": 0.03955078125, "epoch": 0.36816816816816816, "mask_bce_loss": 0.4815174639225006, "mask_dice_loss": 0.05702577158808708, "mask_loss": 0.5385432243347168, "step": 613 }, { "epoch": 0.3687687687687688, "grad_norm": 22.912588119506836, "learning_rate": 1.457885924576901e-06, "loss": 0.7974179983139038, "step": 614 }, { "ce_loss": 0.0001662695431150496, "cls_loss": 0.055908203125, "epoch": 0.3687687687687688, "mask_bce_loss": 0.8044983148574829, "mask_dice_loss": 0.10489463806152344, "mask_loss": 0.9093929529190063, "step": 614 }, { "epoch": 0.36936936936936937, "grad_norm": 20.32954216003418, "learning_rate": 1.4561557031246628e-06, "loss": 0.8178210258483887, "step": 615 }, { "ce_loss": 0.040695466101169586, "cls_loss": 0.04638671875, "epoch": 0.36936936936936937, "mask_bce_loss": 0.5808338522911072, "mask_dice_loss": 0.08218862861394882, "mask_loss": 0.6630224585533142, "step": 615 }, { "epoch": 0.36996996996996995, "grad_norm": 10.452431678771973, "learning_rate": 1.4544237555656216e-06, "loss": 0.6936496496200562, "step": 616 }, { "ce_loss": 5.124543167767115e-05, "cls_loss": 0.03955078125, "epoch": 0.36996996996996995, "mask_bce_loss": 0.33994168043136597, "mask_dice_loss": 0.037756312638521194, "mask_loss": 0.37769800424575806, "step": 616 }, { "epoch": 0.3705705705705706, "grad_norm": 23.271610260009766, "learning_rate": 1.4526900884535191e-06, "loss": 0.705162763595581, "step": 617 }, { "ce_loss": 7.876088056946173e-05, "cls_loss": 0.06591796875, "epoch": 0.3705705705705706, "mask_bce_loss": 0.8235536813735962, "mask_dice_loss": 0.14020878076553345, "mask_loss": 0.9637624621391296, "step": 617 }, { "epoch": 0.37117117117117115, "grad_norm": 37.24272155761719, "learning_rate": 1.4509547083486032e-06, "loss": 0.9604480266571045, "step": 618 }, { "ce_loss": 3.399401975912042e-05, "cls_loss": 0.042724609375, "epoch": 0.37117117117117115, "mask_bce_loss": 0.549074113368988, "mask_dice_loss": 0.06046475097537041, "mask_loss": 0.6095388531684875, "step": 618 }, { "epoch": 0.3717717717717718, "grad_norm": 20.114852905273438, "learning_rate": 1.4492176218176042e-06, "loss": 0.8417680859565735, "step": 619 }, { "ce_loss": 3.322382326587103e-05, "cls_loss": 0.055908203125, "epoch": 0.3717717717717718, "mask_bce_loss": 0.860669732093811, "mask_dice_loss": 0.1199726089835167, "mask_loss": 0.9806423187255859, "step": 619 }, { "epoch": 0.37237237237237236, "grad_norm": 48.962371826171875, "learning_rate": 1.4474788354337093e-06, "loss": 0.8662930727005005, "step": 620 }, { "ce_loss": 0.0052436161786317825, "cls_loss": 0.068359375, "epoch": 0.37237237237237236, "mask_bce_loss": 0.29394546151161194, "mask_dice_loss": 0.061341892927885056, "mask_loss": 0.3552873432636261, "step": 620 }, { "epoch": 0.372972972972973, "grad_norm": 33.29944610595703, "learning_rate": 1.4457383557765383e-06, "loss": 0.8895025253295898, "step": 621 }, { "ce_loss": 5.275190051179379e-05, "cls_loss": 0.06591796875, "epoch": 0.372972972972973, "mask_bce_loss": 0.6531856656074524, "mask_dice_loss": 0.09297198057174683, "mask_loss": 0.7461576461791992, "step": 621 }, { "epoch": 0.37357357357357357, "grad_norm": 29.394424438476562, "learning_rate": 1.443996189432118e-06, "loss": 0.8787436485290527, "step": 622 }, { "ce_loss": 0.00011726390948751941, "cls_loss": 0.034912109375, "epoch": 0.37357357357357357, "mask_bce_loss": 0.41352424025535583, "mask_dice_loss": 0.028780508786439896, "mask_loss": 0.4423047602176666, "step": 622 }, { "epoch": 0.3741741741741742, "grad_norm": 45.819854736328125, "learning_rate": 1.442252342992858e-06, "loss": 0.8296586275100708, "step": 623 }, { "ce_loss": 4.956190605298616e-05, "cls_loss": 0.041748046875, "epoch": 0.3741741741741742, "mask_bce_loss": 0.7633671164512634, "mask_dice_loss": 0.07622366398572922, "mask_loss": 0.8395907878875732, "step": 623 }, { "epoch": 0.3747747747747748, "grad_norm": 34.77122116088867, "learning_rate": 1.4405068230575249e-06, "loss": 0.7861945629119873, "step": 624 }, { "ce_loss": 7.207097223727033e-05, "cls_loss": 0.06884765625, "epoch": 0.3747747747747748, "mask_bce_loss": 1.1764862537384033, "mask_dice_loss": 0.0882064700126648, "mask_loss": 1.264692783355713, "step": 624 }, { "epoch": 0.37537537537537535, "grad_norm": 19.245296478271484, "learning_rate": 1.4387596362312188e-06, "loss": 0.8743371367454529, "step": 625 }, { "ce_loss": 9.140096517512575e-05, "cls_loss": 0.05908203125, "epoch": 0.37537537537537535, "mask_bce_loss": 0.7275797128677368, "mask_dice_loss": 0.09536358714103699, "mask_loss": 0.8229433298110962, "step": 625 }, { "epoch": 0.375975975975976, "grad_norm": 20.099414825439453, "learning_rate": 1.4370107891253467e-06, "loss": 0.820283830165863, "step": 626 }, { "ce_loss": 8.120057464111596e-05, "cls_loss": 0.020751953125, "epoch": 0.375975975975976, "mask_bce_loss": 0.2483489066362381, "mask_dice_loss": 0.013496804051101208, "mask_loss": 0.2618457078933716, "step": 626 }, { "epoch": 0.37657657657657656, "grad_norm": 27.354248046875, "learning_rate": 1.4352602883575982e-06, "loss": 0.8507751822471619, "step": 627 }, { "ce_loss": 9.082026372198015e-05, "cls_loss": 0.05029296875, "epoch": 0.37657657657657656, "mask_bce_loss": 0.5301099419593811, "mask_dice_loss": 0.05547083541750908, "mask_loss": 0.5855807662010193, "step": 627 }, { "epoch": 0.3771771771771772, "grad_norm": 22.621721267700195, "learning_rate": 1.4335081405519205e-06, "loss": 0.8902187347412109, "step": 628 }, { "ce_loss": 5.045230864197947e-05, "cls_loss": 0.06689453125, "epoch": 0.3771771771771772, "mask_bce_loss": 0.8706046938896179, "mask_dice_loss": 0.15742073953151703, "mask_loss": 1.0280253887176514, "step": 628 }, { "epoch": 0.37777777777777777, "grad_norm": 22.760356903076172, "learning_rate": 1.4317543523384928e-06, "loss": 0.7649794816970825, "step": 629 }, { "ce_loss": 0.00018836092203855515, "cls_loss": 0.043701171875, "epoch": 0.37777777777777777, "mask_bce_loss": 0.29584261775016785, "mask_dice_loss": 0.03934386000037193, "mask_loss": 0.3351864814758301, "step": 629 }, { "epoch": 0.3783783783783784, "grad_norm": 16.67914581298828, "learning_rate": 1.4299989303537028e-06, "loss": 0.7836309671401978, "step": 630 }, { "ce_loss": 0.00030170142417773604, "cls_loss": 0.0625, "epoch": 0.3783783783783784, "mask_bce_loss": 0.9398748278617859, "mask_dice_loss": 0.09308742731809616, "mask_loss": 1.0329622030258179, "step": 630 }, { "epoch": 0.378978978978979, "grad_norm": 22.562557220458984, "learning_rate": 1.4282418812401196e-06, "loss": 0.8405314683914185, "step": 631 }, { "ce_loss": 0.00011845221160911024, "cls_loss": 0.03759765625, "epoch": 0.378978978978979, "mask_bce_loss": 0.24334990978240967, "mask_dice_loss": 0.02827414870262146, "mask_loss": 0.27162405848503113, "step": 631 }, { "epoch": 0.3795795795795796, "grad_norm": 45.943668365478516, "learning_rate": 1.4264832116464689e-06, "loss": 0.9226745367050171, "step": 632 }, { "ce_loss": 5.773522934759967e-05, "cls_loss": 0.031494140625, "epoch": 0.3795795795795796, "mask_bce_loss": 0.23466749489307404, "mask_dice_loss": 0.025730639696121216, "mask_loss": 0.26039814949035645, "step": 632 }, { "epoch": 0.3801801801801802, "grad_norm": 19.87589454650879, "learning_rate": 1.4247229282276094e-06, "loss": 0.9515233039855957, "step": 633 }, { "ce_loss": 5.8587298553902656e-05, "cls_loss": 0.046142578125, "epoch": 0.3801801801801802, "mask_bce_loss": 0.6502931714057922, "mask_dice_loss": 0.08397717028856277, "mask_loss": 0.7342703342437744, "step": 633 }, { "epoch": 0.38078078078078076, "grad_norm": 150.38575744628906, "learning_rate": 1.4229610376445063e-06, "loss": 0.9035167694091797, "step": 634 }, { "ce_loss": 0.0011220622109249234, "cls_loss": 0.04833984375, "epoch": 0.38078078078078076, "mask_bce_loss": 0.5812867283821106, "mask_dice_loss": 0.05407421663403511, "mask_loss": 0.6353609561920166, "step": 634 }, { "epoch": 0.3813813813813814, "grad_norm": 16.000652313232422, "learning_rate": 1.421197546564206e-06, "loss": 0.8066563606262207, "step": 635 }, { "ce_loss": 0.0001521986268926412, "cls_loss": 0.05078125, "epoch": 0.3813813813813814, "mask_bce_loss": 0.6753308176994324, "mask_dice_loss": 0.07204438745975494, "mask_loss": 0.7473751902580261, "step": 635 }, { "epoch": 0.38198198198198197, "grad_norm": 19.65711784362793, "learning_rate": 1.4194324616598107e-06, "loss": 0.6983388662338257, "step": 636 }, { "ce_loss": 5.342107033357024e-05, "cls_loss": 0.060546875, "epoch": 0.38198198198198197, "mask_bce_loss": 0.5528099536895752, "mask_dice_loss": 0.11496064811944962, "mask_loss": 0.6677706241607666, "step": 636 }, { "epoch": 0.3825825825825826, "grad_norm": 35.21226119995117, "learning_rate": 1.4176657896104559e-06, "loss": 1.037369966506958, "step": 637 }, { "ce_loss": 3.3491927752038464e-05, "cls_loss": 0.09423828125, "epoch": 0.3825825825825826, "mask_bce_loss": 1.7424348592758179, "mask_dice_loss": 0.10325178503990173, "mask_loss": 1.845686674118042, "step": 637 }, { "epoch": 0.3831831831831832, "grad_norm": 57.25870895385742, "learning_rate": 1.41589753710128e-06, "loss": 0.9320389628410339, "step": 638 }, { "ce_loss": 0.0001013615692500025, "cls_loss": 0.056640625, "epoch": 0.3831831831831832, "mask_bce_loss": 0.9130816459655762, "mask_dice_loss": 0.09910070151090622, "mask_loss": 1.012182354927063, "step": 638 }, { "epoch": 0.3837837837837838, "grad_norm": 15.844152450561523, "learning_rate": 1.4141277108234042e-06, "loss": 0.7892382740974426, "step": 639 }, { "ce_loss": 4.330981755629182e-05, "cls_loss": 0.06689453125, "epoch": 0.3837837837837838, "mask_bce_loss": 2.08388614654541, "mask_dice_loss": 0.10573076456785202, "mask_loss": 2.1896169185638428, "step": 639 }, { "epoch": 0.3843843843843844, "grad_norm": 26.581363677978516, "learning_rate": 1.4123563174739036e-06, "loss": 0.9505511522293091, "step": 640 }, { "ce_loss": 0.00011114902008557692, "cls_loss": 0.044189453125, "epoch": 0.3843843843843844, "mask_bce_loss": 0.43663135170936584, "mask_dice_loss": 0.21143662929534912, "mask_loss": 0.6480679512023926, "step": 640 }, { "epoch": 0.384984984984985, "grad_norm": 18.48186492919922, "learning_rate": 1.4105833637557834e-06, "loss": 0.7100874185562134, "step": 641 }, { "ce_loss": 3.929725426132791e-05, "cls_loss": 0.03662109375, "epoch": 0.384984984984985, "mask_bce_loss": 0.4911326467990875, "mask_dice_loss": 0.032477013766765594, "mask_loss": 0.5236096382141113, "step": 641 }, { "epoch": 0.3855855855855856, "grad_norm": 54.66416549682617, "learning_rate": 1.4088088563779534e-06, "loss": 1.03770112991333, "step": 642 }, { "ce_loss": 3.3892451028805226e-05, "cls_loss": 0.0654296875, "epoch": 0.3855855855855856, "mask_bce_loss": 0.6909699440002441, "mask_dice_loss": 0.12451481819152832, "mask_loss": 0.8154847621917725, "step": 642 }, { "epoch": 0.38618618618618616, "grad_norm": 22.227771759033203, "learning_rate": 1.407032802055203e-06, "loss": 0.8857964277267456, "step": 643 }, { "ce_loss": 0.00013546994887292385, "cls_loss": 0.060546875, "epoch": 0.38618618618618616, "mask_bce_loss": 0.606901228427887, "mask_dice_loss": 0.0786265954375267, "mask_loss": 0.6855278015136719, "step": 643 }, { "epoch": 0.3867867867867868, "grad_norm": 28.39056396484375, "learning_rate": 1.4052552075081748e-06, "loss": 0.78203284740448, "step": 644 }, { "ce_loss": 0.004080571699887514, "cls_loss": 0.05859375, "epoch": 0.3867867867867868, "mask_bce_loss": 0.7867510914802551, "mask_dice_loss": 0.07984519004821777, "mask_loss": 0.8665962815284729, "step": 644 }, { "epoch": 0.38738738738738737, "grad_norm": 30.25860595703125, "learning_rate": 1.4034760794633388e-06, "loss": 0.7823581099510193, "step": 645 }, { "ce_loss": 0.0003674349281936884, "cls_loss": 0.04345703125, "epoch": 0.38738738738738737, "mask_bce_loss": 0.7341260313987732, "mask_dice_loss": 0.10135593265295029, "mask_loss": 0.8354819416999817, "step": 645 }, { "epoch": 0.387987987987988, "grad_norm": 16.410608291625977, "learning_rate": 1.4016954246529694e-06, "loss": 0.8519115447998047, "step": 646 }, { "ce_loss": 0.0001225570886163041, "cls_loss": 0.05712890625, "epoch": 0.387987987987988, "mask_bce_loss": 0.6095561385154724, "mask_dice_loss": 0.07980161905288696, "mask_loss": 0.6893577575683594, "step": 646 }, { "epoch": 0.3885885885885886, "grad_norm": 38.42129898071289, "learning_rate": 1.3999132498151174e-06, "loss": 0.9298736453056335, "step": 647 }, { "ce_loss": 8.747675019549206e-05, "cls_loss": 0.04443359375, "epoch": 0.3885885885885886, "mask_bce_loss": 0.6166914105415344, "mask_dice_loss": 0.10435737669467926, "mask_loss": 0.7210487723350525, "step": 647 }, { "epoch": 0.3891891891891892, "grad_norm": 32.71314239501953, "learning_rate": 1.3981295616935855e-06, "loss": 0.8836300373077393, "step": 648 }, { "ce_loss": 4.271691068424843e-05, "cls_loss": 0.039306640625, "epoch": 0.3891891891891892, "mask_bce_loss": 0.23816852271556854, "mask_dice_loss": 0.033094581216573715, "mask_loss": 0.27126309275627136, "step": 648 }, { "epoch": 0.3897897897897898, "grad_norm": 23.028446197509766, "learning_rate": 1.3963443670379032e-06, "loss": 0.7641613483428955, "step": 649 }, { "ce_loss": 0.024370266124606133, "cls_loss": 0.06884765625, "epoch": 0.3897897897897898, "mask_bce_loss": 0.2774313986301422, "mask_dice_loss": 0.09052655845880508, "mask_loss": 0.3679579496383667, "step": 649 }, { "epoch": 0.39039039039039036, "grad_norm": 32.86122131347656, "learning_rate": 1.3945576726032996e-06, "loss": 0.8007315397262573, "step": 650 }, { "ce_loss": 5.816799966851249e-05, "cls_loss": 0.0673828125, "epoch": 0.39039039039039036, "mask_bce_loss": 0.883063018321991, "mask_dice_loss": 0.08844057470560074, "mask_loss": 0.9715036153793335, "step": 650 }, { "epoch": 0.390990990990991, "grad_norm": 34.96807098388672, "learning_rate": 1.3927694851506804e-06, "loss": 0.8336824774742126, "step": 651 }, { "ce_loss": 2.194360968132969e-05, "cls_loss": 0.059814453125, "epoch": 0.390990990990991, "mask_bce_loss": 0.6209900379180908, "mask_dice_loss": 0.10976483672857285, "mask_loss": 0.7307548522949219, "step": 651 }, { "epoch": 0.39159159159159157, "grad_norm": 28.217519760131836, "learning_rate": 1.3909798114466003e-06, "loss": 0.8787999153137207, "step": 652 }, { "ce_loss": 0.011365575715899467, "cls_loss": 0.068359375, "epoch": 0.39159159159159157, "mask_bce_loss": 1.9543758630752563, "mask_dice_loss": 0.06976785510778427, "mask_loss": 2.024143695831299, "step": 652 }, { "epoch": 0.3921921921921922, "grad_norm": 17.808908462524414, "learning_rate": 1.3891886582632383e-06, "loss": 0.7816166877746582, "step": 653 }, { "ce_loss": 0.0023650075308978558, "cls_loss": 0.042724609375, "epoch": 0.3921921921921922, "mask_bce_loss": 0.36210542917251587, "mask_dice_loss": 0.052419621497392654, "mask_loss": 0.4145250618457794, "step": 653 }, { "epoch": 0.3927927927927928, "grad_norm": 17.90109634399414, "learning_rate": 1.3873960323783705e-06, "loss": 0.7460364103317261, "step": 654 }, { "ce_loss": 5.7353016018169e-05, "cls_loss": 0.059814453125, "epoch": 0.3927927927927928, "mask_bce_loss": 0.711320161819458, "mask_dice_loss": 0.1262199729681015, "mask_loss": 0.8375401496887207, "step": 654 }, { "epoch": 0.3933933933933934, "grad_norm": 38.65781021118164, "learning_rate": 1.385601940575348e-06, "loss": 0.7586143612861633, "step": 655 }, { "ce_loss": 7.950923463795334e-05, "cls_loss": 0.03955078125, "epoch": 0.3933933933933934, "mask_bce_loss": 0.6100844740867615, "mask_dice_loss": 0.06262575089931488, "mask_loss": 0.6727102398872375, "step": 655 }, { "epoch": 0.393993993993994, "grad_norm": 38.05052185058594, "learning_rate": 1.3838063896430668e-06, "loss": 0.77317214012146, "step": 656 }, { "ce_loss": 8.504307334078476e-05, "cls_loss": 0.06640625, "epoch": 0.393993993993994, "mask_bce_loss": 1.3197829723358154, "mask_dice_loss": 0.12341997772455215, "mask_loss": 1.4432029724121094, "step": 656 }, { "epoch": 0.3945945945945946, "grad_norm": 25.547767639160156, "learning_rate": 1.3820093863759459e-06, "loss": 0.8675908446311951, "step": 657 }, { "ce_loss": 7.124288094928488e-05, "cls_loss": 0.05224609375, "epoch": 0.3945945945945946, "mask_bce_loss": 0.5593017935752869, "mask_dice_loss": 0.06874856352806091, "mask_loss": 0.6280503273010254, "step": 657 }, { "epoch": 0.3951951951951952, "grad_norm": 32.90003967285156, "learning_rate": 1.3802109375738992e-06, "loss": 0.7979005575180054, "step": 658 }, { "ce_loss": 0.00027747623971663415, "cls_loss": 0.0546875, "epoch": 0.3951951951951952, "mask_bce_loss": 0.35324516892433167, "mask_dice_loss": 0.054135192185640335, "mask_loss": 0.4073803722858429, "step": 658 }, { "epoch": 0.39579579579579577, "grad_norm": 13.611188888549805, "learning_rate": 1.3784110500423102e-06, "loss": 0.6698867082595825, "step": 659 }, { "ce_loss": 5.543512452277355e-05, "cls_loss": 0.053955078125, "epoch": 0.39579579579579577, "mask_bce_loss": 1.105817437171936, "mask_dice_loss": 0.08555898070335388, "mask_loss": 1.1913764476776123, "step": 659 }, { "epoch": 0.3963963963963964, "grad_norm": 20.798233032226562, "learning_rate": 1.3766097305920074e-06, "loss": 1.0064702033996582, "step": 660 }, { "ce_loss": 1.826211519073695e-05, "cls_loss": 0.036376953125, "epoch": 0.3963963963963964, "mask_bce_loss": 0.36630767583847046, "mask_dice_loss": 0.04717900976538658, "mask_loss": 0.41348668932914734, "step": 660 }, { "epoch": 0.396996996996997, "grad_norm": 43.33415222167969, "learning_rate": 1.374806986039238e-06, "loss": 1.0175743103027344, "step": 661 }, { "ce_loss": 0.0003070069942623377, "cls_loss": 0.04736328125, "epoch": 0.396996996996997, "mask_bce_loss": 0.41258174180984497, "mask_dice_loss": 0.04070276394486427, "mask_loss": 0.45328450202941895, "step": 661 }, { "epoch": 0.3975975975975976, "grad_norm": 36.14617919921875, "learning_rate": 1.3730028232056403e-06, "loss": 0.8964970111846924, "step": 662 }, { "ce_loss": 6.086334906285629e-05, "cls_loss": 0.033447265625, "epoch": 0.3975975975975976, "mask_bce_loss": 0.29743853211402893, "mask_dice_loss": 0.02548709511756897, "mask_loss": 0.3229256272315979, "step": 662 }, { "epoch": 0.3981981981981982, "grad_norm": 16.835447311401367, "learning_rate": 1.3711972489182206e-06, "loss": 0.7965752482414246, "step": 663 }, { "ce_loss": 5.258131932350807e-05, "cls_loss": 0.04443359375, "epoch": 0.3981981981981982, "mask_bce_loss": 0.37346887588500977, "mask_dice_loss": 0.047302503138780594, "mask_loss": 0.42077139019966125, "step": 663 }, { "epoch": 0.3987987987987988, "grad_norm": 23.599376678466797, "learning_rate": 1.3693902700093263e-06, "loss": 0.8789092302322388, "step": 664 }, { "ce_loss": 2.7127445719088428e-05, "cls_loss": 0.04736328125, "epoch": 0.3987987987987988, "mask_bce_loss": 0.7640061974525452, "mask_dice_loss": 0.04485667496919632, "mask_loss": 0.8088628649711609, "step": 664 }, { "epoch": 0.3993993993993994, "grad_norm": 21.799694061279297, "learning_rate": 1.367581893316619e-06, "loss": 0.802903413772583, "step": 665 }, { "ce_loss": 0.012440589256584644, "cls_loss": 0.052001953125, "epoch": 0.3993993993993994, "mask_bce_loss": 0.5911379456520081, "mask_dice_loss": 0.10761409997940063, "mask_loss": 0.6987520456314087, "step": 665 }, { "epoch": 0.4, "grad_norm": 20.673126220703125, "learning_rate": 1.3657721256830507e-06, "loss": 0.8719495534896851, "step": 666 }, { "ce_loss": 5.400109876063652e-05, "cls_loss": 0.048828125, "epoch": 0.4, "mask_bce_loss": 0.8949980139732361, "mask_dice_loss": 0.09506735950708389, "mask_loss": 0.9900653958320618, "step": 666 }, { "epoch": 0.4006006006006006, "grad_norm": 77.19306945800781, "learning_rate": 1.3639609739568357e-06, "loss": 0.853844404220581, "step": 667 }, { "ce_loss": 4.3809242924908176e-05, "cls_loss": 0.04443359375, "epoch": 0.4006006006006006, "mask_bce_loss": 0.9434391856193542, "mask_dice_loss": 0.09711930900812149, "mask_loss": 1.0405584573745728, "step": 667 }, { "epoch": 0.4012012012012012, "grad_norm": 17.71759605407715, "learning_rate": 1.362148444991426e-06, "loss": 0.8578077554702759, "step": 668 }, { "ce_loss": 0.0003000429423991591, "cls_loss": 0.057861328125, "epoch": 0.4012012012012012, "mask_bce_loss": 0.44990140199661255, "mask_dice_loss": 0.06841836869716644, "mask_loss": 0.5183197855949402, "step": 668 }, { "epoch": 0.4018018018018018, "grad_norm": 27.622529983520508, "learning_rate": 1.3603345456454859e-06, "loss": 0.731269121170044, "step": 669 }, { "ce_loss": 0.0010169566376134753, "cls_loss": 0.041015625, "epoch": 0.4018018018018018, "mask_bce_loss": 0.4434907138347626, "mask_dice_loss": 0.03671915829181671, "mask_loss": 0.4802098870277405, "step": 669 }, { "epoch": 0.4024024024024024, "grad_norm": 19.01423454284668, "learning_rate": 1.3585192827828648e-06, "loss": 0.8090158104896545, "step": 670 }, { "ce_loss": 4.2436931835254654e-05, "cls_loss": 0.05908203125, "epoch": 0.4024024024024024, "mask_bce_loss": 0.9115857481956482, "mask_dice_loss": 0.13571666181087494, "mask_loss": 1.0473023653030396, "step": 670 }, { "epoch": 0.403003003003003, "grad_norm": 45.40174865722656, "learning_rate": 1.3567026632725709e-06, "loss": 0.9234320521354675, "step": 671 }, { "ce_loss": 0.04696395620703697, "cls_loss": 0.04052734375, "epoch": 0.403003003003003, "mask_bce_loss": 0.5717089176177979, "mask_dice_loss": 0.06288313865661621, "mask_loss": 0.6345920562744141, "step": 671 }, { "epoch": 0.4036036036036036, "grad_norm": 15.16917896270752, "learning_rate": 1.354884693988747e-06, "loss": 0.7645335793495178, "step": 672 }, { "ce_loss": 2.5741821445990354e-05, "cls_loss": 0.049560546875, "epoch": 0.4036036036036036, "mask_bce_loss": 0.46549421548843384, "mask_dice_loss": 0.07148711383342743, "mask_loss": 0.5369813442230225, "step": 672 }, { "epoch": 0.4042042042042042, "grad_norm": 24.70475196838379, "learning_rate": 1.3530653818106435e-06, "loss": 0.7987892627716064, "step": 673 }, { "ce_loss": 0.000253895967034623, "cls_loss": 0.04736328125, "epoch": 0.4042042042042042, "mask_bce_loss": 0.8613687753677368, "mask_dice_loss": 0.09894856810569763, "mask_loss": 0.9603173732757568, "step": 673 }, { "epoch": 0.4048048048048048, "grad_norm": 27.780513763427734, "learning_rate": 1.3512447336225915e-06, "loss": 0.7944475412368774, "step": 674 }, { "ce_loss": 0.0016501829959452152, "cls_loss": 0.042236328125, "epoch": 0.4048048048048048, "mask_bce_loss": 1.0042482614517212, "mask_dice_loss": 0.08302532881498337, "mask_loss": 1.0872735977172852, "step": 674 }, { "epoch": 0.40540540540540543, "grad_norm": 30.531578063964844, "learning_rate": 1.3494227563139783e-06, "loss": 0.7464995384216309, "step": 675 }, { "ce_loss": 0.00010885352821787819, "cls_loss": 0.055908203125, "epoch": 0.40540540540540543, "mask_bce_loss": 0.41318631172180176, "mask_dice_loss": 0.06768190860748291, "mask_loss": 0.48086822032928467, "step": 675 }, { "epoch": 0.406006006006006, "grad_norm": 34.520301818847656, "learning_rate": 1.3475994567792205e-06, "loss": 0.7768267393112183, "step": 676 }, { "ce_loss": 6.921528256498277e-05, "cls_loss": 0.0712890625, "epoch": 0.406006006006006, "mask_bce_loss": 0.9067763686180115, "mask_dice_loss": 0.11422164738178253, "mask_loss": 1.0209980010986328, "step": 676 }, { "epoch": 0.4066066066066066, "grad_norm": 23.194612503051758, "learning_rate": 1.3457748419177382e-06, "loss": 0.7816750407218933, "step": 677 }, { "ce_loss": 0.020474081858992577, "cls_loss": 0.05712890625, "epoch": 0.4066066066066066, "mask_bce_loss": 0.7427220344543457, "mask_dice_loss": 0.20585155487060547, "mask_loss": 0.9485735893249512, "step": 677 }, { "epoch": 0.4072072072072072, "grad_norm": 17.82710838317871, "learning_rate": 1.343948918633928e-06, "loss": 0.8119887113571167, "step": 678 }, { "ce_loss": 4.1264400351792574e-05, "cls_loss": 0.03662109375, "epoch": 0.4072072072072072, "mask_bce_loss": 0.3108845055103302, "mask_dice_loss": 0.05265219137072563, "mask_loss": 0.36353668570518494, "step": 678 }, { "epoch": 0.4078078078078078, "grad_norm": 31.670717239379883, "learning_rate": 1.3421216938371386e-06, "loss": 0.9061055183410645, "step": 679 }, { "ce_loss": 0.00023606004833709449, "cls_loss": 0.034912109375, "epoch": 0.4078078078078078, "mask_bce_loss": 0.23001721501350403, "mask_dice_loss": 0.02692081406712532, "mask_loss": 0.25693804025650024, "step": 679 }, { "epoch": 0.4084084084084084, "grad_norm": 29.676834106445312, "learning_rate": 1.340293174441643e-06, "loss": 0.8856677412986755, "step": 680 }, { "ce_loss": 3.878131610690616e-05, "cls_loss": 0.057861328125, "epoch": 0.4084084084084084, "mask_bce_loss": 0.9771634936332703, "mask_dice_loss": 0.0815035030245781, "mask_loss": 1.0586669445037842, "step": 680 }, { "epoch": 0.409009009009009, "grad_norm": 33.42451858520508, "learning_rate": 1.338463367366613e-06, "loss": 0.9671831130981445, "step": 681 }, { "ce_loss": 2.8665604986599647e-05, "cls_loss": 0.0634765625, "epoch": 0.409009009009009, "mask_bce_loss": 0.7264576554298401, "mask_dice_loss": 0.1316148191690445, "mask_loss": 0.8580724596977234, "step": 681 }, { "epoch": 0.4096096096096096, "grad_norm": 26.680036544799805, "learning_rate": 1.3366322795360936e-06, "loss": 0.9706312417984009, "step": 682 }, { "ce_loss": 6.144901271909475e-05, "cls_loss": 0.0380859375, "epoch": 0.4096096096096096, "mask_bce_loss": 0.6911236643791199, "mask_dice_loss": 0.0638318583369255, "mask_loss": 0.754955530166626, "step": 682 }, { "epoch": 0.4102102102102102, "grad_norm": 30.09131622314453, "learning_rate": 1.3347999178789758e-06, "loss": 0.8348636627197266, "step": 683 }, { "ce_loss": 0.006877517327666283, "cls_loss": 0.049560546875, "epoch": 0.4102102102102102, "mask_bce_loss": 0.5834897756576538, "mask_dice_loss": 0.04717951640486717, "mask_loss": 0.6306692957878113, "step": 683 }, { "epoch": 0.41081081081081083, "grad_norm": 35.28017807006836, "learning_rate": 1.3329662893289706e-06, "loss": 0.9519654512405396, "step": 684 }, { "ce_loss": 4.8429083108203486e-05, "cls_loss": 0.046142578125, "epoch": 0.41081081081081083, "mask_bce_loss": 0.7338581681251526, "mask_dice_loss": 0.06179181858897209, "mask_loss": 0.7956500053405762, "step": 684 }, { "epoch": 0.4114114114114114, "grad_norm": 19.914661407470703, "learning_rate": 1.331131400824583e-06, "loss": 0.8739578723907471, "step": 685 }, { "ce_loss": 3.653981548268348e-05, "cls_loss": 0.03173828125, "epoch": 0.4114114114114114, "mask_bce_loss": 0.30564776062965393, "mask_dice_loss": 0.024851419031620026, "mask_loss": 0.33049917221069336, "step": 685 }, { "epoch": 0.412012012012012, "grad_norm": 22.44626235961914, "learning_rate": 1.3292952593090867e-06, "loss": 0.7115669250488281, "step": 686 }, { "ce_loss": 0.00010178783850278705, "cls_loss": 0.05078125, "epoch": 0.412012012012012, "mask_bce_loss": 0.30047130584716797, "mask_dice_loss": 0.06275427341461182, "mask_loss": 0.3632255792617798, "step": 686 }, { "epoch": 0.4126126126126126, "grad_norm": 21.345571517944336, "learning_rate": 1.327457871730495e-06, "loss": 0.8455881476402283, "step": 687 }, { "ce_loss": 2.609602415759582e-05, "cls_loss": 0.0625, "epoch": 0.4126126126126126, "mask_bce_loss": 1.536672592163086, "mask_dice_loss": 0.08162641525268555, "mask_loss": 1.6182990074157715, "step": 687 }, { "epoch": 0.4132132132132132, "grad_norm": 25.89703369140625, "learning_rate": 1.3256192450415382e-06, "loss": 0.8915035128593445, "step": 688 }, { "ce_loss": 5.984916060697287e-05, "cls_loss": 0.04248046875, "epoch": 0.4132132132132132, "mask_bce_loss": 0.4272187352180481, "mask_dice_loss": 0.05263537913560867, "mask_loss": 0.47985410690307617, "step": 688 }, { "epoch": 0.4138138138138138, "grad_norm": 18.208263397216797, "learning_rate": 1.3237793861996343e-06, "loss": 0.7597731351852417, "step": 689 }, { "ce_loss": 2.8749445846187882e-05, "cls_loss": 0.05712890625, "epoch": 0.4138138138138138, "mask_bce_loss": 0.6032981276512146, "mask_dice_loss": 0.14449255168437958, "mask_loss": 0.7477906942367554, "step": 689 }, { "epoch": 0.4144144144144144, "grad_norm": 40.771583557128906, "learning_rate": 1.3219383021668638e-06, "loss": 1.0006046295166016, "step": 690 }, { "ce_loss": 4.997090945835225e-05, "cls_loss": 0.0654296875, "epoch": 0.4144144144144144, "mask_bce_loss": 0.335639089345932, "mask_dice_loss": 0.16398519277572632, "mask_loss": 0.4996242821216583, "step": 690 }, { "epoch": 0.41501501501501503, "grad_norm": 25.070205688476562, "learning_rate": 1.3200959999099438e-06, "loss": 0.9145486354827881, "step": 691 }, { "ce_loss": 7.169512537075207e-05, "cls_loss": 0.021240234375, "epoch": 0.41501501501501503, "mask_bce_loss": 0.7509756088256836, "mask_dice_loss": 0.030966198071837425, "mask_loss": 0.781941831111908, "step": 691 }, { "epoch": 0.4156156156156156, "grad_norm": 21.29458236694336, "learning_rate": 1.3182524864002016e-06, "loss": 0.7873988151550293, "step": 692 }, { "ce_loss": 6.82744721416384e-05, "cls_loss": 0.04736328125, "epoch": 0.4156156156156156, "mask_bce_loss": 0.3840591609477997, "mask_dice_loss": 0.08740327507257462, "mask_loss": 0.4714624285697937, "step": 692 }, { "epoch": 0.41621621621621624, "grad_norm": 38.586273193359375, "learning_rate": 1.3164077686135469e-06, "loss": 0.9188232421875, "step": 693 }, { "ce_loss": 0.0003307429433334619, "cls_loss": 0.05029296875, "epoch": 0.41621621621621624, "mask_bce_loss": 0.6305398344993591, "mask_dice_loss": 0.07878061383962631, "mask_loss": 0.7093204259872437, "step": 693 }, { "epoch": 0.4168168168168168, "grad_norm": 29.805967330932617, "learning_rate": 1.3145618535304466e-06, "loss": 0.934280514717102, "step": 694 }, { "ce_loss": 0.00023459619842469692, "cls_loss": 0.04296875, "epoch": 0.4168168168168168, "mask_bce_loss": 0.30361801385879517, "mask_dice_loss": 0.0403742790222168, "mask_loss": 0.34399229288101196, "step": 694 }, { "epoch": 0.4174174174174174, "grad_norm": 29.25069236755371, "learning_rate": 1.312714748135899e-06, "loss": 0.8314527869224548, "step": 695 }, { "ce_loss": 0.00010200426186202094, "cls_loss": 0.037353515625, "epoch": 0.4174174174174174, "mask_bce_loss": 0.3982764184474945, "mask_dice_loss": 0.05933786556124687, "mask_loss": 0.4576142728328705, "step": 695 }, { "epoch": 0.418018018018018, "grad_norm": 31.17323112487793, "learning_rate": 1.310866459419406e-06, "loss": 0.8462741374969482, "step": 696 }, { "ce_loss": 5.266728112474084e-05, "cls_loss": 0.0299072265625, "epoch": 0.418018018018018, "mask_bce_loss": 0.20351532101631165, "mask_dice_loss": 0.022809624671936035, "mask_loss": 0.22632494568824768, "step": 696 }, { "epoch": 0.4186186186186186, "grad_norm": 19.61150360107422, "learning_rate": 1.3090169943749473e-06, "loss": 0.8922959566116333, "step": 697 }, { "ce_loss": 2.907255111495033e-05, "cls_loss": 0.06005859375, "epoch": 0.4186186186186186, "mask_bce_loss": 0.9147437214851379, "mask_dice_loss": 0.12279651314020157, "mask_loss": 1.0375401973724365, "step": 697 }, { "epoch": 0.41921921921921923, "grad_norm": 45.722232818603516, "learning_rate": 1.3071663600009539e-06, "loss": 0.9217414855957031, "step": 698 }, { "ce_loss": 7.626471779076383e-05, "cls_loss": 0.02099609375, "epoch": 0.41921921921921923, "mask_bce_loss": 0.3937850892543793, "mask_dice_loss": 0.03477032855153084, "mask_loss": 0.428555428981781, "step": 698 }, { "epoch": 0.4198198198198198, "grad_norm": 37.55644226074219, "learning_rate": 1.305314563300281e-06, "loss": 1.0455043315887451, "step": 699 }, { "ce_loss": 2.0235889678588137e-05, "cls_loss": 0.033447265625, "epoch": 0.4198198198198198, "mask_bce_loss": 0.4096870422363281, "mask_dice_loss": 0.024685321375727654, "mask_loss": 0.4343723654747009, "step": 699 }, { "epoch": 0.42042042042042044, "grad_norm": 23.737939834594727, "learning_rate": 1.3034616112801833e-06, "loss": 0.7662962675094604, "step": 700 }, { "ce_loss": 0.0013831807300448418, "cls_loss": 0.0289306640625, "epoch": 0.42042042042042044, "mask_bce_loss": 0.49487391114234924, "mask_dice_loss": 0.041649315506219864, "mask_loss": 0.5365232229232788, "step": 700 }, { "epoch": 0.421021021021021, "grad_norm": 55.09552764892578, "learning_rate": 1.3016075109522864e-06, "loss": 0.8615195751190186, "step": 701 }, { "ce_loss": 5.4317537433234975e-05, "cls_loss": 0.05810546875, "epoch": 0.421021021021021, "mask_bce_loss": 0.7596988081932068, "mask_dice_loss": 0.16080139577388763, "mask_loss": 0.9205002188682556, "step": 701 }, { "epoch": 0.42162162162162165, "grad_norm": 24.346372604370117, "learning_rate": 1.299752269332561e-06, "loss": 0.855843186378479, "step": 702 }, { "ce_loss": 2.4749915610300377e-05, "cls_loss": 0.05712890625, "epoch": 0.42162162162162165, "mask_bce_loss": 0.9643392562866211, "mask_dice_loss": 0.08707060664892197, "mask_loss": 1.0514098405838013, "step": 702 }, { "epoch": 0.4222222222222222, "grad_norm": 26.02216148376465, "learning_rate": 1.2978958934412969e-06, "loss": 0.9056000113487244, "step": 703 }, { "ce_loss": 6.37176344753243e-05, "cls_loss": 0.020751953125, "epoch": 0.4222222222222222, "mask_bce_loss": 0.20719514787197113, "mask_dice_loss": 0.012920704670250416, "mask_loss": 0.22011585533618927, "step": 703 }, { "epoch": 0.4228228228228228, "grad_norm": 46.48820495605469, "learning_rate": 1.2960383903030763e-06, "loss": 0.9154869914054871, "step": 704 }, { "ce_loss": 0.00013088817649986595, "cls_loss": 0.05419921875, "epoch": 0.4228228228228228, "mask_bce_loss": 0.44603556394577026, "mask_dice_loss": 0.0459110364317894, "mask_loss": 0.49194660782814026, "step": 704 }, { "epoch": 0.42342342342342343, "grad_norm": 17.21481704711914, "learning_rate": 1.2941797669467462e-06, "loss": 0.7746362090110779, "step": 705 }, { "ce_loss": 0.0016536925686523318, "cls_loss": 0.045166015625, "epoch": 0.42342342342342343, "mask_bce_loss": 1.1866850852966309, "mask_dice_loss": 0.1444983333349228, "mask_loss": 1.3311834335327148, "step": 705 }, { "epoch": 0.424024024024024, "grad_norm": 25.865028381347656, "learning_rate": 1.2923200304053929e-06, "loss": 0.8344086408615112, "step": 706 }, { "ce_loss": 4.0138376789400354e-05, "cls_loss": 0.039306640625, "epoch": 0.424024024024024, "mask_bce_loss": 0.23534810543060303, "mask_dice_loss": 0.03504751995205879, "mask_loss": 0.2703956365585327, "step": 706 }, { "epoch": 0.42462462462462464, "grad_norm": 33.62355041503906, "learning_rate": 1.290459187716315e-06, "loss": 0.7062844038009644, "step": 707 }, { "ce_loss": 0.00015266292029991746, "cls_loss": 0.0654296875, "epoch": 0.42462462462462464, "mask_bce_loss": 1.1257941722869873, "mask_dice_loss": 0.11867628246545792, "mask_loss": 1.244470477104187, "step": 707 }, { "epoch": 0.4252252252252252, "grad_norm": 26.833040237426758, "learning_rate": 1.288597245920997e-06, "loss": 0.8167043924331665, "step": 708 }, { "ce_loss": 3.8690821384079754e-05, "cls_loss": 0.026123046875, "epoch": 0.4252252252252252, "mask_bce_loss": 0.3300594985485077, "mask_dice_loss": 0.020137032493948936, "mask_loss": 0.3501965403556824, "step": 708 }, { "epoch": 0.42582582582582584, "grad_norm": 20.334991455078125, "learning_rate": 1.286734212065082e-06, "loss": 0.8028894662857056, "step": 709 }, { "ce_loss": 9.605544619262218e-05, "cls_loss": 0.055908203125, "epoch": 0.42582582582582584, "mask_bce_loss": 0.6992489099502563, "mask_dice_loss": 0.07553290575742722, "mask_loss": 0.7747818231582642, "step": 709 }, { "epoch": 0.4264264264264264, "grad_norm": 34.45218276977539, "learning_rate": 1.2848700931983461e-06, "loss": 0.8515400886535645, "step": 710 }, { "ce_loss": 0.000268796255113557, "cls_loss": 0.032470703125, "epoch": 0.4264264264264264, "mask_bce_loss": 0.2083667516708374, "mask_dice_loss": 0.02751585841178894, "mask_loss": 0.23588261008262634, "step": 710 }, { "epoch": 0.42702702702702705, "grad_norm": 32.45187759399414, "learning_rate": 1.2830048963746702e-06, "loss": 0.9037745594978333, "step": 711 }, { "ce_loss": 2.0374991436256096e-05, "cls_loss": 0.061279296875, "epoch": 0.42702702702702705, "mask_bce_loss": 0.7400842308998108, "mask_dice_loss": 0.09464634954929352, "mask_loss": 0.8347305655479431, "step": 711 }, { "epoch": 0.4276276276276276, "grad_norm": 43.045841217041016, "learning_rate": 1.281138628652015e-06, "loss": 0.8684367537498474, "step": 712 }, { "ce_loss": 3.985965668107383e-05, "cls_loss": 0.05517578125, "epoch": 0.4276276276276276, "mask_bce_loss": 0.5633795261383057, "mask_dice_loss": 0.13131342828273773, "mask_loss": 0.6946929693222046, "step": 712 }, { "epoch": 0.4282282282282282, "grad_norm": 54.40781021118164, "learning_rate": 1.279271297092394e-06, "loss": 0.9212600588798523, "step": 713 }, { "ce_loss": 0.0001460024795960635, "cls_loss": 0.05419921875, "epoch": 0.4282282282282282, "mask_bce_loss": 0.726154625415802, "mask_dice_loss": 0.07789307832717896, "mask_loss": 0.804047703742981, "step": 713 }, { "epoch": 0.42882882882882883, "grad_norm": 46.84645080566406, "learning_rate": 1.2774029087618445e-06, "loss": 0.9031703472137451, "step": 714 }, { "ce_loss": 8.88924187165685e-05, "cls_loss": 0.049560546875, "epoch": 0.42882882882882883, "mask_bce_loss": 0.390556275844574, "mask_dice_loss": 0.04889143258333206, "mask_loss": 0.43944770097732544, "step": 714 }, { "epoch": 0.4294294294294294, "grad_norm": 32.36003875732422, "learning_rate": 1.2755334707304043e-06, "loss": 0.9471696615219116, "step": 715 }, { "ce_loss": 3.0080102078500204e-05, "cls_loss": 0.0279541015625, "epoch": 0.4294294294294294, "mask_bce_loss": 0.4170903265476227, "mask_dice_loss": 0.04429075866937637, "mask_loss": 0.46138107776641846, "step": 715 }, { "epoch": 0.43003003003003004, "grad_norm": 17.51795768737793, "learning_rate": 1.273662990072083e-06, "loss": 0.8111399412155151, "step": 716 }, { "ce_loss": 3.4331194910919294e-05, "cls_loss": 0.052734375, "epoch": 0.43003003003003004, "mask_bce_loss": 0.5318140983581543, "mask_dice_loss": 0.06647977977991104, "mask_loss": 0.5982939004898071, "step": 716 }, { "epoch": 0.4306306306306306, "grad_norm": 31.093360900878906, "learning_rate": 1.2717914738648345e-06, "loss": 0.7883923649787903, "step": 717 }, { "ce_loss": 4.7026936954353005e-05, "cls_loss": 0.02392578125, "epoch": 0.4306306306306306, "mask_bce_loss": 0.18473750352859497, "mask_dice_loss": 0.017037685960531235, "mask_loss": 0.2017751932144165, "step": 717 }, { "epoch": 0.43123123123123125, "grad_norm": 32.83554458618164, "learning_rate": 1.2699189291905325e-06, "loss": 0.8028268218040466, "step": 718 }, { "ce_loss": 0.0002573633973952383, "cls_loss": 0.051513671875, "epoch": 0.43123123123123125, "mask_bce_loss": 0.723812460899353, "mask_dice_loss": 0.10931290686130524, "mask_loss": 0.8331253528594971, "step": 718 }, { "epoch": 0.4318318318318318, "grad_norm": 50.80091857910156, "learning_rate": 1.268045363134942e-06, "loss": 0.9369886517524719, "step": 719 }, { "ce_loss": 8.671639807289466e-05, "cls_loss": 0.05078125, "epoch": 0.4318318318318318, "mask_bce_loss": 0.8266593813896179, "mask_dice_loss": 0.1358793079853058, "mask_loss": 0.9625387191772461, "step": 719 }, { "epoch": 0.43243243243243246, "grad_norm": 23.76282501220703, "learning_rate": 1.2661707827876931e-06, "loss": 0.7946118116378784, "step": 720 }, { "ce_loss": 4.620182517101057e-05, "cls_loss": 0.05615234375, "epoch": 0.43243243243243246, "mask_bce_loss": 0.9020969271659851, "mask_dice_loss": 0.08450745791196823, "mask_loss": 0.9866043925285339, "step": 720 }, { "epoch": 0.43303303303303303, "grad_norm": 30.887197494506836, "learning_rate": 1.264295195242253e-06, "loss": 0.8652093410491943, "step": 721 }, { "ce_loss": 4.36209375038743e-05, "cls_loss": 0.03857421875, "epoch": 0.43303303303303303, "mask_bce_loss": 0.44667848944664, "mask_dice_loss": 0.039584603160619736, "mask_loss": 0.48626309633255005, "step": 721 }, { "epoch": 0.4336336336336336, "grad_norm": 40.940860748291016, "learning_rate": 1.2624186075959019e-06, "loss": 0.9123404622077942, "step": 722 }, { "ce_loss": 4.955933036399074e-05, "cls_loss": 0.0712890625, "epoch": 0.4336336336336336, "mask_bce_loss": 0.4960578382015228, "mask_dice_loss": 0.14174306392669678, "mask_loss": 0.637800931930542, "step": 722 }, { "epoch": 0.43423423423423424, "grad_norm": 23.42190933227539, "learning_rate": 1.2605410269497029e-06, "loss": 0.764150083065033, "step": 723 }, { "ce_loss": 8.047857409110293e-05, "cls_loss": 0.040771484375, "epoch": 0.43423423423423424, "mask_bce_loss": 0.2694847285747528, "mask_dice_loss": 0.03748595714569092, "mask_loss": 0.3069706857204437, "step": 723 }, { "epoch": 0.4348348348348348, "grad_norm": 20.948558807373047, "learning_rate": 1.2586624604084777e-06, "loss": 0.7977696657180786, "step": 724 }, { "ce_loss": 4.707293919636868e-05, "cls_loss": 0.04736328125, "epoch": 0.4348348348348348, "mask_bce_loss": 0.286089152097702, "mask_dice_loss": 0.06070629879832268, "mask_loss": 0.3467954397201538, "step": 724 }, { "epoch": 0.43543543543543545, "grad_norm": 25.86458396911621, "learning_rate": 1.2567829150807777e-06, "loss": 0.704802930355072, "step": 725 }, { "ce_loss": 3.91429421142675e-05, "cls_loss": 0.050048828125, "epoch": 0.43543543543543545, "mask_bce_loss": 0.7112008929252625, "mask_dice_loss": 0.10622876137495041, "mask_loss": 0.8174296617507935, "step": 725 }, { "epoch": 0.436036036036036, "grad_norm": 55.35181427001953, "learning_rate": 1.2549023980788586e-06, "loss": 0.9843809008598328, "step": 726 }, { "ce_loss": 6.99916563462466e-05, "cls_loss": 0.05078125, "epoch": 0.436036036036036, "mask_bce_loss": 0.6138972043991089, "mask_dice_loss": 0.047861497849226, "mask_loss": 0.6617587208747864, "step": 726 }, { "epoch": 0.43663663663663665, "grad_norm": 37.364994049072266, "learning_rate": 1.253020916518653e-06, "loss": 0.9387848973274231, "step": 727 }, { "ce_loss": 0.00018784106941893697, "cls_loss": 0.05517578125, "epoch": 0.43663663663663665, "mask_bce_loss": 0.9686793684959412, "mask_dice_loss": 0.11199551820755005, "mask_loss": 1.0806748867034912, "step": 727 }, { "epoch": 0.43723723723723723, "grad_norm": 47.002933502197266, "learning_rate": 1.2511384775197442e-06, "loss": 0.7028055191040039, "step": 728 }, { "ce_loss": 0.0064775533974170685, "cls_loss": 0.048828125, "epoch": 0.43723723723723723, "mask_bce_loss": 0.7268842458724976, "mask_dice_loss": 0.04377306252717972, "mask_loss": 0.7706573009490967, "step": 728 }, { "epoch": 0.43783783783783786, "grad_norm": 27.517623901367188, "learning_rate": 1.2492550882053362e-06, "loss": 0.8751168847084045, "step": 729 }, { "ce_loss": 5.3863674111198634e-05, "cls_loss": 0.055908203125, "epoch": 0.43783783783783786, "mask_bce_loss": 0.4598563313484192, "mask_dice_loss": 0.07103096693754196, "mask_loss": 0.5308873057365417, "step": 729 }, { "epoch": 0.43843843843843844, "grad_norm": 21.554981231689453, "learning_rate": 1.247370755702231e-06, "loss": 0.8450319170951843, "step": 730 }, { "ce_loss": 3.853996531688608e-05, "cls_loss": 0.0546875, "epoch": 0.43843843843843844, "mask_bce_loss": 0.5853770971298218, "mask_dice_loss": 0.10963030904531479, "mask_loss": 0.6950073838233948, "step": 730 }, { "epoch": 0.439039039039039, "grad_norm": 24.651071548461914, "learning_rate": 1.245485487140799e-06, "loss": 0.8465476632118225, "step": 731 }, { "ce_loss": 3.701842433656566e-05, "cls_loss": 0.061767578125, "epoch": 0.439039039039039, "mask_bce_loss": 0.42844653129577637, "mask_dice_loss": 0.12327047437429428, "mask_loss": 0.5517169833183289, "step": 731 }, { "epoch": 0.43963963963963965, "grad_norm": 22.509695053100586, "learning_rate": 1.2435992896549529e-06, "loss": 0.8318625688552856, "step": 732 }, { "ce_loss": 0.0034339199773967266, "cls_loss": 0.046875, "epoch": 0.43963963963963965, "mask_bce_loss": 0.9952413439750671, "mask_dice_loss": 0.16135196387767792, "mask_loss": 1.1565933227539062, "step": 732 }, { "epoch": 0.4402402402402402, "grad_norm": 24.523160934448242, "learning_rate": 1.2417121703821203e-06, "loss": 0.8749771118164062, "step": 733 }, { "ce_loss": 0.00010060727072414011, "cls_loss": 0.057861328125, "epoch": 0.4402402402402402, "mask_bce_loss": 0.4405632019042969, "mask_dice_loss": 0.16380171477794647, "mask_loss": 0.6043649315834045, "step": 733 }, { "epoch": 0.44084084084084085, "grad_norm": 19.6577091217041, "learning_rate": 1.2398241364632168e-06, "loss": 0.8699910640716553, "step": 734 }, { "ce_loss": 0.004165238700807095, "cls_loss": 0.053955078125, "epoch": 0.44084084084084085, "mask_bce_loss": 0.9284629821777344, "mask_dice_loss": 0.06135142594575882, "mask_loss": 0.9898144006729126, "step": 734 }, { "epoch": 0.44144144144144143, "grad_norm": 32.81895446777344, "learning_rate": 1.2379351950426186e-06, "loss": 1.0055938959121704, "step": 735 }, { "ce_loss": 7.298049604287371e-05, "cls_loss": 0.0277099609375, "epoch": 0.44144144144144143, "mask_bce_loss": 0.5407683849334717, "mask_dice_loss": 0.032533127814531326, "mask_loss": 0.5733014941215515, "step": 735 }, { "epoch": 0.44204204204204206, "grad_norm": 28.844696044921875, "learning_rate": 1.2360453532681374e-06, "loss": 0.8948678970336914, "step": 736 }, { "ce_loss": 4.105049083591439e-05, "cls_loss": 0.06005859375, "epoch": 0.44204204204204206, "mask_bce_loss": 1.034642219543457, "mask_dice_loss": 0.10465886443853378, "mask_loss": 1.139301061630249, "step": 736 }, { "epoch": 0.44264264264264264, "grad_norm": 40.665000915527344, "learning_rate": 1.2341546182909906e-06, "loss": 0.9736939668655396, "step": 737 }, { "ce_loss": 3.84890845452901e-05, "cls_loss": 0.0439453125, "epoch": 0.44264264264264264, "mask_bce_loss": 0.5850664377212524, "mask_dice_loss": 0.04584527015686035, "mask_loss": 0.6309117078781128, "step": 737 }, { "epoch": 0.44324324324324327, "grad_norm": 18.273101806640625, "learning_rate": 1.2322629972657754e-06, "loss": 0.7030397653579712, "step": 738 }, { "ce_loss": 4.744448960991576e-05, "cls_loss": 0.056640625, "epoch": 0.44324324324324327, "mask_bce_loss": 0.7612690329551697, "mask_dice_loss": 0.11116867512464523, "mask_loss": 0.8724377155303955, "step": 738 }, { "epoch": 0.44384384384384384, "grad_norm": 68.77681732177734, "learning_rate": 1.2303704973504424e-06, "loss": 1.084376573562622, "step": 739 }, { "ce_loss": 4.415192597662099e-05, "cls_loss": 0.06298828125, "epoch": 0.44384384384384384, "mask_bce_loss": 0.4127757251262665, "mask_dice_loss": 0.09420590847730637, "mask_loss": 0.506981611251831, "step": 739 }, { "epoch": 0.4444444444444444, "grad_norm": 39.634971618652344, "learning_rate": 1.2284771257062682e-06, "loss": 0.8967198729515076, "step": 740 }, { "ce_loss": 0.00014790985733270645, "cls_loss": 0.045166015625, "epoch": 0.4444444444444444, "mask_bce_loss": 1.0982950925827026, "mask_dice_loss": 0.08324893563985825, "mask_loss": 1.1815440654754639, "step": 740 }, { "epoch": 0.44504504504504505, "grad_norm": 48.170738220214844, "learning_rate": 1.2265828894978272e-06, "loss": 0.9192883968353271, "step": 741 }, { "ce_loss": 2.3875667466199957e-05, "cls_loss": 0.044921875, "epoch": 0.44504504504504505, "mask_bce_loss": 0.6941110491752625, "mask_dice_loss": 0.17539918422698975, "mask_loss": 0.8695102334022522, "step": 741 }, { "epoch": 0.4456456456456456, "grad_norm": 24.61949348449707, "learning_rate": 1.224687795892966e-06, "loss": 0.6511399149894714, "step": 742 }, { "ce_loss": 0.0004325772461015731, "cls_loss": 0.033935546875, "epoch": 0.4456456456456456, "mask_bce_loss": 0.5287262797355652, "mask_dice_loss": 0.02889237366616726, "mask_loss": 0.5576186776161194, "step": 742 }, { "epoch": 0.44624624624624626, "grad_norm": 52.51433563232422, "learning_rate": 1.2227918520627756e-06, "loss": 0.7446731328964233, "step": 743 }, { "ce_loss": 5.044350473326631e-05, "cls_loss": 0.04638671875, "epoch": 0.44624624624624626, "mask_bce_loss": 0.6313285231590271, "mask_dice_loss": 0.06544611603021622, "mask_loss": 0.6967746615409851, "step": 743 }, { "epoch": 0.44684684684684683, "grad_norm": 33.26121520996094, "learning_rate": 1.2208950651815632e-06, "loss": 1.0357201099395752, "step": 744 }, { "ce_loss": 4.5698379835812375e-05, "cls_loss": 0.028076171875, "epoch": 0.44684684684684683, "mask_bce_loss": 0.13575845956802368, "mask_dice_loss": 0.01930338703095913, "mask_loss": 0.15506184101104736, "step": 744 }, { "epoch": 0.44744744744744747, "grad_norm": 12.753948211669922, "learning_rate": 1.2189974424268279e-06, "loss": 0.6349391937255859, "step": 745 }, { "ce_loss": 3.808852125075646e-05, "cls_loss": 0.045166015625, "epoch": 0.44744744744744747, "mask_bce_loss": 0.4281824231147766, "mask_dice_loss": 0.0442834310233593, "mask_loss": 0.472465842962265, "step": 745 }, { "epoch": 0.44804804804804804, "grad_norm": 21.904083251953125, "learning_rate": 1.2170989909792312e-06, "loss": 0.7589070796966553, "step": 746 }, { "ce_loss": 0.00026864054962061346, "cls_loss": 0.03369140625, "epoch": 0.44804804804804804, "mask_bce_loss": 0.34975701570510864, "mask_dice_loss": 0.030682802200317383, "mask_loss": 0.380439817905426, "step": 746 }, { "epoch": 0.4486486486486487, "grad_norm": 19.602947235107422, "learning_rate": 1.2151997180225686e-06, "loss": 0.707724928855896, "step": 747 }, { "ce_loss": 0.00012519651500042528, "cls_loss": 0.03857421875, "epoch": 0.4486486486486487, "mask_bce_loss": 0.4556143283843994, "mask_dice_loss": 0.053126879036426544, "mask_loss": 0.5087411999702454, "step": 747 }, { "epoch": 0.44924924924924925, "grad_norm": 64.68488311767578, "learning_rate": 1.2132996307437468e-06, "loss": 1.0412163734436035, "step": 748 }, { "ce_loss": 2.8462192858569324e-05, "cls_loss": 0.05859375, "epoch": 0.44924924924924925, "mask_bce_loss": 0.8048354387283325, "mask_dice_loss": 0.12951014935970306, "mask_loss": 0.9343456029891968, "step": 748 }, { "epoch": 0.4498498498498498, "grad_norm": 25.09249496459961, "learning_rate": 1.2113987363327525e-06, "loss": 0.8414925932884216, "step": 749 }, { "ce_loss": 8.714415889699012e-05, "cls_loss": 0.04443359375, "epoch": 0.4498498498498498, "mask_bce_loss": 0.29759809374809265, "mask_dice_loss": 0.046818166971206665, "mask_loss": 0.3444162607192993, "step": 749 }, { "epoch": 0.45045045045045046, "grad_norm": 39.70106506347656, "learning_rate": 1.2094970419826274e-06, "loss": 0.7806898355484009, "step": 750 }, { "ce_loss": 0.0001907302503241226, "cls_loss": 0.0302734375, "epoch": 0.45045045045045046, "mask_bce_loss": 0.4981144964694977, "mask_dice_loss": 0.05051423981785774, "mask_loss": 0.5486287474632263, "step": 750 }, { "epoch": 0.45105105105105103, "grad_norm": 33.0090217590332, "learning_rate": 1.2075945548894386e-06, "loss": 0.9152379035949707, "step": 751 }, { "ce_loss": 3.3082582376664504e-05, "cls_loss": 0.05224609375, "epoch": 0.45105105105105103, "mask_bce_loss": 0.8362002372741699, "mask_dice_loss": 0.12412180006504059, "mask_loss": 0.9603220224380493, "step": 751 }, { "epoch": 0.45165165165165166, "grad_norm": 20.209857940673828, "learning_rate": 1.2056912822522552e-06, "loss": 0.8083670139312744, "step": 752 }, { "ce_loss": 9.045261685969308e-05, "cls_loss": 0.04052734375, "epoch": 0.45165165165165166, "mask_bce_loss": 0.28221797943115234, "mask_dice_loss": 0.030483638867735863, "mask_loss": 0.31270161271095276, "step": 752 }, { "epoch": 0.45225225225225224, "grad_norm": 21.35057830810547, "learning_rate": 1.2037872312731173e-06, "loss": 0.8377780914306641, "step": 753 }, { "ce_loss": 0.0006674789474345744, "cls_loss": 0.06494140625, "epoch": 0.45225225225225224, "mask_bce_loss": 0.3774029612541199, "mask_dice_loss": 0.17242340743541718, "mask_loss": 0.5498263835906982, "step": 753 }, { "epoch": 0.45285285285285287, "grad_norm": 20.503631591796875, "learning_rate": 1.2018824091570102e-06, "loss": 0.6998971700668335, "step": 754 }, { "ce_loss": 6.620413478231058e-05, "cls_loss": 0.048095703125, "epoch": 0.45285285285285287, "mask_bce_loss": 0.5271311402320862, "mask_dice_loss": 0.06936124712228775, "mask_loss": 0.5964924097061157, "step": 754 }, { "epoch": 0.45345345345345345, "grad_norm": 31.53561019897461, "learning_rate": 1.1999768231118383e-06, "loss": 0.6809221506118774, "step": 755 }, { "ce_loss": 5.5407595937140286e-05, "cls_loss": 0.057861328125, "epoch": 0.45345345345345345, "mask_bce_loss": 0.34311220049858093, "mask_dice_loss": 0.08209652453660965, "mask_loss": 0.42520871758461, "step": 755 }, { "epoch": 0.4540540540540541, "grad_norm": 12.047144889831543, "learning_rate": 1.1980704803483962e-06, "loss": 0.7535558342933655, "step": 756 }, { "ce_loss": 0.0005006912397220731, "cls_loss": 0.048583984375, "epoch": 0.4540540540540541, "mask_bce_loss": 0.41218456625938416, "mask_dice_loss": 0.050140511244535446, "mask_loss": 0.4623250663280487, "step": 756 }, { "epoch": 0.45465465465465466, "grad_norm": 18.978151321411133, "learning_rate": 1.196163388080341e-06, "loss": 0.8865047097206116, "step": 757 }, { "ce_loss": 0.002379131969064474, "cls_loss": 0.03564453125, "epoch": 0.45465465465465466, "mask_bce_loss": 0.19096845388412476, "mask_dice_loss": 0.05166534706950188, "mask_loss": 0.24263380467891693, "step": 757 }, { "epoch": 0.45525525525525523, "grad_norm": 18.846839904785156, "learning_rate": 1.1942555535241675e-06, "loss": 0.7410646677017212, "step": 758 }, { "ce_loss": 3.4555228921817616e-05, "cls_loss": 0.041015625, "epoch": 0.45525525525525523, "mask_bce_loss": 0.5860627293586731, "mask_dice_loss": 0.04550382122397423, "mask_loss": 0.6315665245056152, "step": 758 }, { "epoch": 0.45585585585585586, "grad_norm": 36.2799072265625, "learning_rate": 1.1923469838991787e-06, "loss": 0.9321595430374146, "step": 759 }, { "ce_loss": 0.00014750845730304718, "cls_loss": 0.050048828125, "epoch": 0.45585585585585586, "mask_bce_loss": 0.33695870637893677, "mask_dice_loss": 0.09714536368846893, "mask_loss": 0.4341040849685669, "step": 759 }, { "epoch": 0.45645645645645644, "grad_norm": 20.02613639831543, "learning_rate": 1.1904376864274585e-06, "loss": 0.7632109522819519, "step": 760 }, { "ce_loss": 6.409629713743925e-05, "cls_loss": 0.044921875, "epoch": 0.45645645645645644, "mask_bce_loss": 0.42576318979263306, "mask_dice_loss": 0.03424011915922165, "mask_loss": 0.4600033164024353, "step": 760 }, { "epoch": 0.45705705705705707, "grad_norm": 15.553970336914062, "learning_rate": 1.1885276683338459e-06, "loss": 0.7179821729660034, "step": 761 }, { "ce_loss": 3.738198211067356e-05, "cls_loss": 0.05419921875, "epoch": 0.45705705705705707, "mask_bce_loss": 0.4764203131198883, "mask_dice_loss": 0.05734945461153984, "mask_loss": 0.5337697863578796, "step": 761 }, { "epoch": 0.45765765765765765, "grad_norm": 18.005653381347656, "learning_rate": 1.1866169368459064e-06, "loss": 0.8382155895233154, "step": 762 }, { "ce_loss": 7.826234650565311e-05, "cls_loss": 0.0927734375, "epoch": 0.45765765765765765, "mask_bce_loss": 2.3279097080230713, "mask_dice_loss": 0.12245813757181168, "mask_loss": 2.4503679275512695, "step": 762 }, { "epoch": 0.4582582582582583, "grad_norm": 15.620705604553223, "learning_rate": 1.184705499193905e-06, "loss": 0.8900195360183716, "step": 763 }, { "ce_loss": 5.0413062126608565e-05, "cls_loss": 0.044921875, "epoch": 0.4582582582582583, "mask_bce_loss": 0.5298343300819397, "mask_dice_loss": 0.08276106417179108, "mask_loss": 0.6125953793525696, "step": 763 }, { "epoch": 0.45885885885885885, "grad_norm": 32.365760803222656, "learning_rate": 1.182793362610779e-06, "loss": 0.8686375021934509, "step": 764 }, { "ce_loss": 1.9294244339107536e-05, "cls_loss": 0.05224609375, "epoch": 0.45885885885885885, "mask_bce_loss": 0.3820410966873169, "mask_dice_loss": 0.07674746960401535, "mask_loss": 0.45878857374191284, "step": 764 }, { "epoch": 0.4594594594594595, "grad_norm": 41.39151382446289, "learning_rate": 1.18088053433211e-06, "loss": 0.9814169406890869, "step": 765 }, { "ce_loss": 3.0119803341221996e-05, "cls_loss": 0.055419921875, "epoch": 0.4594594594594595, "mask_bce_loss": 0.8183189630508423, "mask_dice_loss": 0.06182442232966423, "mask_loss": 0.880143404006958, "step": 765 }, { "epoch": 0.46006006006006006, "grad_norm": 69.63567352294922, "learning_rate": 1.1789670215960975e-06, "loss": 0.8541865348815918, "step": 766 }, { "ce_loss": 0.00012074831465724856, "cls_loss": 0.0390625, "epoch": 0.46006006006006006, "mask_bce_loss": 0.5559815764427185, "mask_dice_loss": 0.042456693947315216, "mask_loss": 0.5984382629394531, "step": 766 }, { "epoch": 0.46066066066066064, "grad_norm": 13.421156883239746, "learning_rate": 1.1770528316435303e-06, "loss": 0.7507292628288269, "step": 767 }, { "ce_loss": 4.1480252548353747e-05, "cls_loss": 0.03955078125, "epoch": 0.46066066066066064, "mask_bce_loss": 0.9270166754722595, "mask_dice_loss": 0.03371767699718475, "mask_loss": 0.9607343673706055, "step": 767 }, { "epoch": 0.46126126126126127, "grad_norm": 23.812971115112305, "learning_rate": 1.175137971717762e-06, "loss": 0.8604525327682495, "step": 768 }, { "ce_loss": 4.417677700985223e-05, "cls_loss": 0.057861328125, "epoch": 0.46126126126126127, "mask_bce_loss": 1.5268136262893677, "mask_dice_loss": 0.16988614201545715, "mask_loss": 1.6966997385025024, "step": 768 }, { "epoch": 0.46186186186186184, "grad_norm": 59.16618347167969, "learning_rate": 1.1732224490646778e-06, "loss": 1.0280795097351074, "step": 769 }, { "ce_loss": 0.0010744142346084118, "cls_loss": 0.0791015625, "epoch": 0.46186186186186184, "mask_bce_loss": 1.7684725522994995, "mask_dice_loss": 0.1424424797296524, "mask_loss": 1.9109150171279907, "step": 769 }, { "epoch": 0.4624624624624625, "grad_norm": 33.97134017944336, "learning_rate": 1.1713062709326735e-06, "loss": 0.8112738728523254, "step": 770 }, { "ce_loss": 0.0002553511003497988, "cls_loss": 0.06005859375, "epoch": 0.4624624624624625, "mask_bce_loss": 0.45216161012649536, "mask_dice_loss": 0.07553573697805405, "mask_loss": 0.5276973247528076, "step": 770 }, { "epoch": 0.46306306306306305, "grad_norm": 35.520721435546875, "learning_rate": 1.1693894445726244e-06, "loss": 0.674146294593811, "step": 771 }, { "ce_loss": 0.001724763074889779, "cls_loss": 0.04541015625, "epoch": 0.46306306306306305, "mask_bce_loss": 1.4157464504241943, "mask_dice_loss": 0.06844448298215866, "mask_loss": 1.4841909408569336, "step": 771 }, { "epoch": 0.4636636636636637, "grad_norm": 16.752565383911133, "learning_rate": 1.1674719772378589e-06, "loss": 0.734950065612793, "step": 772 }, { "ce_loss": 0.00022284936858341098, "cls_loss": 0.053466796875, "epoch": 0.4636636636636637, "mask_bce_loss": 0.3841927945613861, "mask_dice_loss": 0.08596944808959961, "mask_loss": 0.4701622426509857, "step": 772 }, { "epoch": 0.46426426426426426, "grad_norm": 30.74295997619629, "learning_rate": 1.16555387618413e-06, "loss": 0.9245489239692688, "step": 773 }, { "ce_loss": 7.718567212577909e-05, "cls_loss": 0.05712890625, "epoch": 0.46426426426426426, "mask_bce_loss": 0.44771456718444824, "mask_dice_loss": 0.12093529850244522, "mask_loss": 0.5686498880386353, "step": 773 }, { "epoch": 0.4648648648648649, "grad_norm": 28.203369140625, "learning_rate": 1.16363514866959e-06, "loss": 0.9318421483039856, "step": 774 }, { "ce_loss": 7.865494262659922e-05, "cls_loss": 0.05908203125, "epoch": 0.4648648648648649, "mask_bce_loss": 1.3686926364898682, "mask_dice_loss": 0.08741045743227005, "mask_loss": 1.4561030864715576, "step": 774 }, { "epoch": 0.46546546546546547, "grad_norm": 20.515161514282227, "learning_rate": 1.1617158019547604e-06, "loss": 0.894646406173706, "step": 775 }, { "ce_loss": 9.76385927060619e-05, "cls_loss": 0.057861328125, "epoch": 0.46546546546546547, "mask_bce_loss": 0.7837337851524353, "mask_dice_loss": 0.15711013972759247, "mask_loss": 0.940843939781189, "step": 775 }, { "epoch": 0.46606606606606604, "grad_norm": 18.83812713623047, "learning_rate": 1.159795843302507e-06, "loss": 0.7231367826461792, "step": 776 }, { "ce_loss": 2.9361892302404158e-05, "cls_loss": 0.039306640625, "epoch": 0.46606606606606604, "mask_bce_loss": 0.29221463203430176, "mask_dice_loss": 0.04446506127715111, "mask_loss": 0.33667969703674316, "step": 776 }, { "epoch": 0.4666666666666667, "grad_norm": 41.52780532836914, "learning_rate": 1.157875279978011e-06, "loss": 0.8107903003692627, "step": 777 }, { "ce_loss": 3.335772635182366e-05, "cls_loss": 0.05078125, "epoch": 0.4666666666666667, "mask_bce_loss": 0.5710607767105103, "mask_dice_loss": 0.06477047502994537, "mask_loss": 0.6358312368392944, "step": 777 }, { "epoch": 0.46726726726726725, "grad_norm": 15.032249450683594, "learning_rate": 1.1559541192487407e-06, "loss": 0.7233729958534241, "step": 778 }, { "ce_loss": 0.0002472899795975536, "cls_loss": 0.03125, "epoch": 0.46726726726726725, "mask_bce_loss": 0.22455044090747833, "mask_dice_loss": 0.026368802413344383, "mask_loss": 0.25091925263404846, "step": 778 }, { "epoch": 0.4678678678678679, "grad_norm": 24.84003448486328, "learning_rate": 1.1540323683844259e-06, "loss": 0.8726996779441833, "step": 779 }, { "ce_loss": 5.4381158406613395e-05, "cls_loss": 0.046875, "epoch": 0.4678678678678679, "mask_bce_loss": 0.761923611164093, "mask_dice_loss": 0.08348046243190765, "mask_loss": 0.8454040884971619, "step": 779 }, { "epoch": 0.46846846846846846, "grad_norm": 12.648919105529785, "learning_rate": 1.1521100346570294e-06, "loss": 0.7743401527404785, "step": 780 }, { "ce_loss": 3.053646651096642e-05, "cls_loss": 0.0302734375, "epoch": 0.46846846846846846, "mask_bce_loss": 0.2535361349582672, "mask_dice_loss": 0.022797729820013046, "mask_loss": 0.27633386850357056, "step": 780 }, { "epoch": 0.4690690690690691, "grad_norm": 33.59210968017578, "learning_rate": 1.1501871253407194e-06, "loss": 0.7219409346580505, "step": 781 }, { "ce_loss": 0.0004376496362965554, "cls_loss": 0.05615234375, "epoch": 0.4690690690690691, "mask_bce_loss": 0.7252289652824402, "mask_dice_loss": 0.0820203348994255, "mask_loss": 0.8072493076324463, "step": 781 }, { "epoch": 0.46966966966966966, "grad_norm": 21.5989933013916, "learning_rate": 1.1482636477118419e-06, "loss": 0.7326957583427429, "step": 782 }, { "ce_loss": 0.00010954835306620225, "cls_loss": 0.053466796875, "epoch": 0.46966966966966966, "mask_bce_loss": 0.8342569470405579, "mask_dice_loss": 0.0718466266989708, "mask_loss": 0.9061035513877869, "step": 782 }, { "epoch": 0.4702702702702703, "grad_norm": 47.5162353515625, "learning_rate": 1.146339609048894e-06, "loss": 0.7739704847335815, "step": 783 }, { "ce_loss": 3.950054451706819e-05, "cls_loss": 0.0238037109375, "epoch": 0.4702702702702703, "mask_bce_loss": 0.23379600048065186, "mask_dice_loss": 0.01589047908782959, "mask_loss": 0.24968647956848145, "step": 783 }, { "epoch": 0.4708708708708709, "grad_norm": 54.3586311340332, "learning_rate": 1.1444150166324952e-06, "loss": 0.8745772838592529, "step": 784 }, { "ce_loss": 0.001282057841308415, "cls_loss": 0.0693359375, "epoch": 0.4708708708708709, "mask_bce_loss": 0.3005494177341461, "mask_dice_loss": 0.15749457478523254, "mask_loss": 0.45804399251937866, "step": 784 }, { "epoch": 0.47147147147147145, "grad_norm": 53.53141784667969, "learning_rate": 1.1424898777453613e-06, "loss": 1.0015840530395508, "step": 785 }, { "ce_loss": 0.0029192697256803513, "cls_loss": 0.049072265625, "epoch": 0.47147147147147145, "mask_bce_loss": 0.6828365921974182, "mask_dice_loss": 0.061243314296007156, "mask_loss": 0.7440798878669739, "step": 785 }, { "epoch": 0.4720720720720721, "grad_norm": 24.401979446411133, "learning_rate": 1.1405641996722746e-06, "loss": 0.7264619469642639, "step": 786 }, { "ce_loss": 4.87210854771547e-05, "cls_loss": 0.0458984375, "epoch": 0.4720720720720721, "mask_bce_loss": 0.6727069020271301, "mask_dice_loss": 0.061792004853487015, "mask_loss": 0.734498918056488, "step": 786 }, { "epoch": 0.47267267267267266, "grad_norm": 41.48252868652344, "learning_rate": 1.1386379897000584e-06, "loss": 0.8123598098754883, "step": 787 }, { "ce_loss": 0.0001050160572049208, "cls_loss": 0.055419921875, "epoch": 0.47267267267267266, "mask_bce_loss": 1.5401427745819092, "mask_dice_loss": 0.11212978512048721, "mask_loss": 1.6522725820541382, "step": 787 }, { "epoch": 0.4732732732732733, "grad_norm": 18.166563034057617, "learning_rate": 1.1367112551175492e-06, "loss": 0.824908971786499, "step": 788 }, { "ce_loss": 4.310266740503721e-05, "cls_loss": 0.06884765625, "epoch": 0.4732732732732733, "mask_bce_loss": 0.941645085811615, "mask_dice_loss": 0.1384483426809311, "mask_loss": 1.0800933837890625, "step": 788 }, { "epoch": 0.47387387387387386, "grad_norm": 28.856969833374023, "learning_rate": 1.1347840032155684e-06, "loss": 0.7620365023612976, "step": 789 }, { "ce_loss": 0.00013675216177944094, "cls_loss": 0.031005859375, "epoch": 0.47387387387387386, "mask_bce_loss": 0.6052498817443848, "mask_dice_loss": 0.01994192972779274, "mask_loss": 0.6251918077468872, "step": 789 }, { "epoch": 0.4744744744744745, "grad_norm": 42.16941452026367, "learning_rate": 1.1328562412868943e-06, "loss": 0.8153557181358337, "step": 790 }, { "ce_loss": 3.169375122524798e-05, "cls_loss": 0.061279296875, "epoch": 0.4744744744744745, "mask_bce_loss": 1.0183852910995483, "mask_dice_loss": 0.08754737675189972, "mask_loss": 1.1059327125549316, "step": 790 }, { "epoch": 0.47507507507507507, "grad_norm": 20.065536499023438, "learning_rate": 1.1309279766262355e-06, "loss": 0.643365740776062, "step": 791 }, { "ce_loss": 2.031452277151402e-05, "cls_loss": 0.053466796875, "epoch": 0.47507507507507507, "mask_bce_loss": 0.6775416731834412, "mask_dice_loss": 0.0645073652267456, "mask_loss": 0.7420490384101868, "step": 791 }, { "epoch": 0.4756756756756757, "grad_norm": 45.621334075927734, "learning_rate": 1.1289992165302033e-06, "loss": 0.9445548057556152, "step": 792 }, { "ce_loss": 3.7138728657737374e-05, "cls_loss": 0.03564453125, "epoch": 0.4756756756756757, "mask_bce_loss": 0.34899744391441345, "mask_dice_loss": 0.03574555739760399, "mask_loss": 0.38474300503730774, "step": 792 }, { "epoch": 0.4762762762762763, "grad_norm": 29.7088680267334, "learning_rate": 1.1270699682972835e-06, "loss": 0.7457238435745239, "step": 793 }, { "ce_loss": 6.408818444469944e-05, "cls_loss": 0.039794921875, "epoch": 0.4762762762762763, "mask_bce_loss": 0.893303394317627, "mask_dice_loss": 0.04404358193278313, "mask_loss": 0.9373469948768616, "step": 793 }, { "epoch": 0.47687687687687685, "grad_norm": 25.68824577331543, "learning_rate": 1.125140239227809e-06, "loss": 0.8292553424835205, "step": 794 }, { "ce_loss": 4.4854721636511385e-05, "cls_loss": 0.053955078125, "epoch": 0.47687687687687685, "mask_bce_loss": 0.6365068554878235, "mask_dice_loss": 0.06704061478376389, "mask_loss": 0.703547477722168, "step": 794 }, { "epoch": 0.4774774774774775, "grad_norm": 33.620018005371094, "learning_rate": 1.1232100366239326e-06, "loss": 0.8450911045074463, "step": 795 }, { "ce_loss": 4.9493362894281745e-05, "cls_loss": 0.032958984375, "epoch": 0.4774774774774775, "mask_bce_loss": 0.33040574193000793, "mask_dice_loss": 0.026415526866912842, "mask_loss": 0.3568212687969208, "step": 795 }, { "epoch": 0.47807807807807806, "grad_norm": 23.65365219116211, "learning_rate": 1.1212793677895975e-06, "loss": 0.7713664770126343, "step": 796 }, { "ce_loss": 7.351317617576569e-05, "cls_loss": 0.046630859375, "epoch": 0.47807807807807806, "mask_bce_loss": 0.9094882011413574, "mask_dice_loss": 0.09017223864793777, "mask_loss": 0.9996604323387146, "step": 796 }, { "epoch": 0.4786786786786787, "grad_norm": 49.99652862548828, "learning_rate": 1.1193482400305134e-06, "loss": 0.8590223789215088, "step": 797 }, { "ce_loss": 2.575235703261569e-05, "cls_loss": 0.048583984375, "epoch": 0.4786786786786787, "mask_bce_loss": 1.1824079751968384, "mask_dice_loss": 0.09022747725248337, "mask_loss": 1.2726354598999023, "step": 797 }, { "epoch": 0.47927927927927927, "grad_norm": 30.687095642089844, "learning_rate": 1.117416660654125e-06, "loss": 0.7846364974975586, "step": 798 }, { "ce_loss": 7.861846825107932e-05, "cls_loss": 0.04150390625, "epoch": 0.47927927927927927, "mask_bce_loss": 0.23870837688446045, "mask_dice_loss": 0.04392021521925926, "mask_loss": 0.28262859582901, "step": 798 }, { "epoch": 0.4798798798798799, "grad_norm": 40.69140625, "learning_rate": 1.1154846369695862e-06, "loss": 0.808833122253418, "step": 799 }, { "ce_loss": 3.264175393269397e-05, "cls_loss": 0.0986328125, "epoch": 0.4798798798798799, "mask_bce_loss": 0.6993020176887512, "mask_dice_loss": 0.06332577764987946, "mask_loss": 0.7626277804374695, "step": 799 }, { "epoch": 0.4804804804804805, "grad_norm": 29.702817916870117, "learning_rate": 1.1135521762877322e-06, "loss": 0.7330321669578552, "step": 800 }, { "ce_loss": 4.578227526508272e-05, "cls_loss": 0.052734375, "epoch": 0.4804804804804805, "mask_bce_loss": 0.5111784338951111, "mask_dice_loss": 0.0997832641005516, "mask_loss": 0.6109616756439209, "step": 800 }, { "epoch": 0.4810810810810811, "grad_norm": 16.351655960083008, "learning_rate": 1.1116192859210525e-06, "loss": 0.7561138272285461, "step": 801 }, { "ce_loss": 0.0010739369317889214, "cls_loss": 0.06103515625, "epoch": 0.4810810810810811, "mask_bce_loss": 0.7623136639595032, "mask_dice_loss": 0.137268528342247, "mask_loss": 0.8995822072029114, "step": 801 }, { "epoch": 0.4816816816816817, "grad_norm": 28.648569107055664, "learning_rate": 1.1096859731836615e-06, "loss": 0.7874044179916382, "step": 802 }, { "ce_loss": 4.645987064577639e-05, "cls_loss": 0.04541015625, "epoch": 0.4816816816816817, "mask_bce_loss": 0.6858086585998535, "mask_dice_loss": 0.05135823413729668, "mask_loss": 0.7371668815612793, "step": 802 }, { "epoch": 0.48228228228228226, "grad_norm": 20.846172332763672, "learning_rate": 1.1077522453912721e-06, "loss": 0.6742606163024902, "step": 803 }, { "ce_loss": 8.299527689814568e-05, "cls_loss": 0.06396484375, "epoch": 0.48228228228228226, "mask_bce_loss": 1.2779457569122314, "mask_dice_loss": 0.139058917760849, "mask_loss": 1.4170047044754028, "step": 803 }, { "epoch": 0.4828828828828829, "grad_norm": 33.26679992675781, "learning_rate": 1.1058181098611689e-06, "loss": 0.8405102491378784, "step": 804 }, { "ce_loss": 0.015127797611057758, "cls_loss": 0.048095703125, "epoch": 0.4828828828828829, "mask_bce_loss": 0.5311204791069031, "mask_dice_loss": 0.04572918638586998, "mask_loss": 0.576849639415741, "step": 804 }, { "epoch": 0.48348348348348347, "grad_norm": 23.105728149414062, "learning_rate": 1.1038835739121778e-06, "loss": 0.889872670173645, "step": 805 }, { "ce_loss": 5.984168092254549e-05, "cls_loss": 0.053955078125, "epoch": 0.48348348348348347, "mask_bce_loss": 0.7188407778739929, "mask_dice_loss": 0.08653140068054199, "mask_loss": 0.8053721785545349, "step": 805 }, { "epoch": 0.4840840840840841, "grad_norm": 25.420442581176758, "learning_rate": 1.1019486448646405e-06, "loss": 0.8336936235427856, "step": 806 }, { "ce_loss": 0.0024577155709266663, "cls_loss": 0.059326171875, "epoch": 0.4840840840840841, "mask_bce_loss": 1.0133968591690063, "mask_dice_loss": 0.07518794387578964, "mask_loss": 1.0885847806930542, "step": 806 }, { "epoch": 0.4846846846846847, "grad_norm": 18.583560943603516, "learning_rate": 1.1000133300403867e-06, "loss": 0.6857542991638184, "step": 807 }, { "ce_loss": 7.961076335050166e-05, "cls_loss": 0.036376953125, "epoch": 0.4846846846846847, "mask_bce_loss": 0.3660178482532501, "mask_dice_loss": 0.05253511667251587, "mask_loss": 0.418552964925766, "step": 807 }, { "epoch": 0.4852852852852853, "grad_norm": 72.29716491699219, "learning_rate": 1.0980776367627056e-06, "loss": 0.9819505214691162, "step": 808 }, { "ce_loss": 0.00015127782535273582, "cls_loss": 0.035400390625, "epoch": 0.4852852852852853, "mask_bce_loss": 0.4201321601867676, "mask_dice_loss": 0.05526277422904968, "mask_loss": 0.47539493441581726, "step": 808 }, { "epoch": 0.4858858858858859, "grad_norm": 34.43510818481445, "learning_rate": 1.0961415723563178e-06, "loss": 0.9630829095840454, "step": 809 }, { "ce_loss": 0.00027617631712928414, "cls_loss": 0.05322265625, "epoch": 0.4858858858858859, "mask_bce_loss": 0.49477291107177734, "mask_dice_loss": 0.08986993879079819, "mask_loss": 0.5846428275108337, "step": 809 }, { "epoch": 0.4864864864864865, "grad_norm": 23.825572967529297, "learning_rate": 1.0942051441473493e-06, "loss": 0.846076250076294, "step": 810 }, { "ce_loss": 2.7389873139327392e-05, "cls_loss": 0.053955078125, "epoch": 0.4864864864864865, "mask_bce_loss": 0.3346635401248932, "mask_dice_loss": 0.08804071694612503, "mask_loss": 0.4227042496204376, "step": 810 }, { "epoch": 0.4870870870870871, "grad_norm": 23.30673599243164, "learning_rate": 1.092268359463302e-06, "loss": 0.765606164932251, "step": 811 }, { "ce_loss": 7.697183173149824e-05, "cls_loss": 0.061279296875, "epoch": 0.4870870870870871, "mask_bce_loss": 1.076846957206726, "mask_dice_loss": 0.08629264682531357, "mask_loss": 1.1631395816802979, "step": 811 }, { "epoch": 0.48768768768768767, "grad_norm": 51.49946975708008, "learning_rate": 1.0903312256330274e-06, "loss": 0.7586376667022705, "step": 812 }, { "ce_loss": 1.517203236289788e-05, "cls_loss": 0.050048828125, "epoch": 0.48768768768768767, "mask_bce_loss": 0.44607096910476685, "mask_dice_loss": 0.10598894208669662, "mask_loss": 0.5520598888397217, "step": 812 }, { "epoch": 0.4882882882882883, "grad_norm": 16.42375373840332, "learning_rate": 1.0883937499866972e-06, "loss": 0.73396897315979, "step": 813 }, { "ce_loss": 9.101603063754737e-05, "cls_loss": 0.053466796875, "epoch": 0.4882882882882883, "mask_bce_loss": 0.9086505770683289, "mask_dice_loss": 0.13263703882694244, "mask_loss": 1.0412876605987549, "step": 813 }, { "epoch": 0.4888888888888889, "grad_norm": 17.24394416809082, "learning_rate": 1.0864559398557773e-06, "loss": 0.7681161761283875, "step": 814 }, { "ce_loss": 0.0001320239098276943, "cls_loss": 0.060546875, "epoch": 0.4888888888888889, "mask_bce_loss": 0.8342634439468384, "mask_dice_loss": 0.18101303279399872, "mask_loss": 1.0152764320373535, "step": 814 }, { "epoch": 0.4894894894894895, "grad_norm": 24.023134231567383, "learning_rate": 1.084517802572999e-06, "loss": 0.7699315547943115, "step": 815 }, { "ce_loss": 0.00022083923977334052, "cls_loss": 0.031005859375, "epoch": 0.4894894894894895, "mask_bce_loss": 0.2680649161338806, "mask_dice_loss": 0.04522082954645157, "mask_loss": 0.3132857382297516, "step": 815 }, { "epoch": 0.4900900900900901, "grad_norm": 22.0269775390625, "learning_rate": 1.0825793454723324e-06, "loss": 0.6620349884033203, "step": 816 }, { "ce_loss": 9.046141349244863e-05, "cls_loss": 0.052734375, "epoch": 0.4900900900900901, "mask_bce_loss": 0.4612824618816376, "mask_dice_loss": 0.08929475396871567, "mask_loss": 0.5505772233009338, "step": 816 }, { "epoch": 0.4906906906906907, "grad_norm": 36.44974899291992, "learning_rate": 1.0806405758889562e-06, "loss": 0.8591538667678833, "step": 817 }, { "ce_loss": 3.420137727516703e-05, "cls_loss": 0.0240478515625, "epoch": 0.4906906906906907, "mask_bce_loss": 0.35755082964897156, "mask_dice_loss": 0.016160478815436363, "mask_loss": 0.37371131777763367, "step": 817 }, { "epoch": 0.4912912912912913, "grad_norm": 23.086807250976562, "learning_rate": 1.0787015011592326e-06, "loss": 0.7888658046722412, "step": 818 }, { "ce_loss": 9.054747351910919e-05, "cls_loss": 0.046875, "epoch": 0.4912912912912913, "mask_bce_loss": 0.4755048453807831, "mask_dice_loss": 0.07540776580572128, "mask_loss": 0.550912618637085, "step": 818 }, { "epoch": 0.4918918918918919, "grad_norm": 48.46745681762695, "learning_rate": 1.0767621286206785e-06, "loss": 0.7826238870620728, "step": 819 }, { "ce_loss": 0.007651366759091616, "cls_loss": 0.06591796875, "epoch": 0.4918918918918919, "mask_bce_loss": 1.2769485712051392, "mask_dice_loss": 0.08367618173360825, "mask_loss": 1.3606247901916504, "step": 819 }, { "epoch": 0.4924924924924925, "grad_norm": 61.026180267333984, "learning_rate": 1.0748224656119372e-06, "loss": 0.9205540418624878, "step": 820 }, { "ce_loss": 4.928872294840403e-05, "cls_loss": 0.044677734375, "epoch": 0.4924924924924925, "mask_bce_loss": 0.6923760771751404, "mask_dice_loss": 0.056154824793338776, "mask_loss": 0.748530924320221, "step": 820 }, { "epoch": 0.49309309309309307, "grad_norm": 25.99658966064453, "learning_rate": 1.0728825194727517e-06, "loss": 0.8347252607345581, "step": 821 }, { "ce_loss": 0.003266335232183337, "cls_loss": 0.06298828125, "epoch": 0.49309309309309307, "mask_bce_loss": 1.3022626638412476, "mask_dice_loss": 0.07045460492372513, "mask_loss": 1.372717261314392, "step": 821 }, { "epoch": 0.4936936936936937, "grad_norm": 37.18352508544922, "learning_rate": 1.0709422975439362e-06, "loss": 0.798374354839325, "step": 822 }, { "ce_loss": 7.640524563612416e-05, "cls_loss": 0.0296630859375, "epoch": 0.4936936936936937, "mask_bce_loss": 0.37478309869766235, "mask_dice_loss": 0.024952849373221397, "mask_loss": 0.3997359573841095, "step": 822 }, { "epoch": 0.4942942942942943, "grad_norm": 32.55410385131836, "learning_rate": 1.069001807167348e-06, "loss": 0.777398943901062, "step": 823 }, { "ce_loss": 4.090913353138603e-05, "cls_loss": 0.060546875, "epoch": 0.4942942942942943, "mask_bce_loss": 1.1008210182189941, "mask_dice_loss": 0.10613689571619034, "mask_loss": 1.2069579362869263, "step": 823 }, { "epoch": 0.4948948948948949, "grad_norm": 27.894868850708008, "learning_rate": 1.067061055685861e-06, "loss": 0.8599697351455688, "step": 824 }, { "ce_loss": 0.021070312708616257, "cls_loss": 0.04296875, "epoch": 0.4948948948948949, "mask_bce_loss": 0.6686858534812927, "mask_dice_loss": 0.04835732281208038, "mask_loss": 0.7170431613922119, "step": 824 }, { "epoch": 0.4954954954954955, "grad_norm": 34.317962646484375, "learning_rate": 1.065120050443337e-06, "loss": 0.7513887286186218, "step": 825 }, { "ce_loss": 4.792252730112523e-05, "cls_loss": 0.05712890625, "epoch": 0.4954954954954955, "mask_bce_loss": 0.5513919591903687, "mask_dice_loss": 0.0988515242934227, "mask_loss": 0.6502434611320496, "step": 825 }, { "epoch": 0.4960960960960961, "grad_norm": 36.30386734008789, "learning_rate": 1.063178798784597e-06, "loss": 0.8532986640930176, "step": 826 }, { "ce_loss": 0.00033206294756382704, "cls_loss": 0.0546875, "epoch": 0.4960960960960961, "mask_bce_loss": 1.1104106903076172, "mask_dice_loss": 0.12525390088558197, "mask_loss": 1.2356646060943604, "step": 826 }, { "epoch": 0.4966966966966967, "grad_norm": 27.135360717773438, "learning_rate": 1.0612373080553959e-06, "loss": 0.8372294902801514, "step": 827 }, { "ce_loss": 2.1776648281957023e-05, "cls_loss": 0.06787109375, "epoch": 0.4966966966966967, "mask_bce_loss": 1.7981970310211182, "mask_dice_loss": 0.08752012997865677, "mask_loss": 1.8857171535491943, "step": 827 }, { "epoch": 0.4972972972972973, "grad_norm": 24.449710845947266, "learning_rate": 1.0592955856023922e-06, "loss": 0.8330257534980774, "step": 828 }, { "ce_loss": 4.1549417801434174e-05, "cls_loss": 0.034423828125, "epoch": 0.4972972972972973, "mask_bce_loss": 0.35621607303619385, "mask_dice_loss": 0.03500312939286232, "mask_loss": 0.39121919870376587, "step": 828 }, { "epoch": 0.4978978978978979, "grad_norm": 27.23814582824707, "learning_rate": 1.057353638773122e-06, "loss": 0.805952250957489, "step": 829 }, { "ce_loss": 5.241751568973996e-05, "cls_loss": 0.06396484375, "epoch": 0.4978978978978979, "mask_bce_loss": 1.4691792726516724, "mask_dice_loss": 0.1293041855096817, "mask_loss": 1.5984834432601929, "step": 829 }, { "epoch": 0.4984984984984985, "grad_norm": 20.07718276977539, "learning_rate": 1.05541147491597e-06, "loss": 0.8039594888687134, "step": 830 }, { "ce_loss": 3.565500446711667e-05, "cls_loss": 0.0419921875, "epoch": 0.4984984984984985, "mask_bce_loss": 0.7708569765090942, "mask_dice_loss": 0.03916285187005997, "mask_loss": 0.810019850730896, "step": 830 }, { "epoch": 0.4990990990990991, "grad_norm": 24.385271072387695, "learning_rate": 1.053469101380142e-06, "loss": 0.9259171485900879, "step": 831 }, { "ce_loss": 3.008789281011559e-05, "cls_loss": 0.0242919921875, "epoch": 0.4990990990990991, "mask_bce_loss": 0.24328212440013885, "mask_dice_loss": 0.01731085032224655, "mask_loss": 0.2605929672718048, "step": 831 }, { "epoch": 0.4996996996996997, "grad_norm": 23.481670379638672, "learning_rate": 1.0515265255156379e-06, "loss": 0.8372352719306946, "step": 832 }, { "ce_loss": 3.4203378163510934e-05, "cls_loss": 0.05615234375, "epoch": 0.4996996996996997, "mask_bce_loss": 0.9171655774116516, "mask_dice_loss": 0.18676108121871948, "mask_loss": 1.103926658630371, "step": 832 }, { "epoch": 0.5003003003003003, "grad_norm": 21.127241134643555, "learning_rate": 1.0495837546732222e-06, "loss": 0.8362260460853577, "step": 833 }, { "ce_loss": 0.00019880120817106217, "cls_loss": 0.031982421875, "epoch": 0.5003003003003003, "mask_bce_loss": 0.2560975253582001, "mask_dice_loss": 0.022803567349910736, "mask_loss": 0.2789011001586914, "step": 833 }, { "epoch": 0.5009009009009009, "grad_norm": 25.80179786682129, "learning_rate": 1.0476407962043985e-06, "loss": 0.8243281841278076, "step": 834 }, { "ce_loss": 2.51341061812127e-05, "cls_loss": 0.03955078125, "epoch": 0.5009009009009009, "mask_bce_loss": 0.3138217628002167, "mask_dice_loss": 0.03652619197964668, "mask_loss": 0.35034796595573425, "step": 834 }, { "epoch": 0.5015015015015015, "grad_norm": 14.655965805053711, "learning_rate": 1.0456976574613791e-06, "loss": 0.7640138864517212, "step": 835 }, { "ce_loss": 0.00036194169661030173, "cls_loss": 0.05615234375, "epoch": 0.5015015015015015, "mask_bce_loss": 0.9116897583007812, "mask_dice_loss": 0.09332060813903809, "mask_loss": 1.0050103664398193, "step": 835 }, { "epoch": 0.5021021021021022, "grad_norm": 24.927522659301758, "learning_rate": 1.043754345797059e-06, "loss": 0.9112365245819092, "step": 836 }, { "ce_loss": 5.635679917759262e-05, "cls_loss": 0.06396484375, "epoch": 0.5021021021021022, "mask_bce_loss": 0.7748767137527466, "mask_dice_loss": 0.12301137298345566, "mask_loss": 0.8978880643844604, "step": 836 }, { "epoch": 0.5027027027027027, "grad_norm": 18.311420440673828, "learning_rate": 1.0418108685649879e-06, "loss": 0.7012025117874146, "step": 837 }, { "ce_loss": 1.9962340957135893e-05, "cls_loss": 0.048828125, "epoch": 0.5027027027027027, "mask_bce_loss": 1.642562747001648, "mask_dice_loss": 0.057381607592105865, "mask_loss": 1.6999443769454956, "step": 837 }, { "epoch": 0.5033033033033033, "grad_norm": 23.120893478393555, "learning_rate": 1.0398672331193417e-06, "loss": 0.8758591413497925, "step": 838 }, { "ce_loss": 4.088560672244057e-05, "cls_loss": 0.0673828125, "epoch": 0.5033033033033033, "mask_bce_loss": 0.6838771104812622, "mask_dice_loss": 0.16793254017829895, "mask_loss": 0.8518096208572388, "step": 838 }, { "epoch": 0.5039039039039039, "grad_norm": 18.135223388671875, "learning_rate": 1.0379234468148948e-06, "loss": 0.7669677138328552, "step": 839 }, { "ce_loss": 0.00011524977162480354, "cls_loss": 0.04150390625, "epoch": 0.5039039039039039, "mask_bce_loss": 0.31537678837776184, "mask_dice_loss": 0.0382174514234066, "mask_loss": 0.35359424352645874, "step": 839 }, { "epoch": 0.5045045045045045, "grad_norm": 29.842899322509766, "learning_rate": 1.0359795170069925e-06, "loss": 0.97507643699646, "step": 840 }, { "ce_loss": 0.00018334126798436046, "cls_loss": 0.0634765625, "epoch": 0.5045045045045045, "mask_bce_loss": 0.06934095174074173, "mask_dice_loss": 0.07943356037139893, "mask_loss": 0.14877450466156006, "step": 840 }, { "epoch": 0.5051051051051051, "grad_norm": 24.07546043395996, "learning_rate": 1.0340354510515235e-06, "loss": 0.7736411690711975, "step": 841 }, { "ce_loss": 3.6460089177126065e-05, "cls_loss": 0.0390625, "epoch": 0.5051051051051051, "mask_bce_loss": 0.29237720370292664, "mask_dice_loss": 0.03145582601428032, "mask_loss": 0.32383301854133606, "step": 841 }, { "epoch": 0.5057057057057057, "grad_norm": 53.01447677612305, "learning_rate": 1.0320912563048913e-06, "loss": 0.7868422269821167, "step": 842 }, { "ce_loss": 0.0001120223241741769, "cls_loss": 0.0546875, "epoch": 0.5057057057057057, "mask_bce_loss": 1.2184865474700928, "mask_dice_loss": 0.11057741940021515, "mask_loss": 1.3290640115737915, "step": 842 }, { "epoch": 0.5063063063063064, "grad_norm": 24.262845993041992, "learning_rate": 1.0301469401239871e-06, "loss": 0.776811957359314, "step": 843 }, { "ce_loss": 0.00038656589458696544, "cls_loss": 0.053955078125, "epoch": 0.5063063063063064, "mask_bce_loss": 0.46517810225486755, "mask_dice_loss": 0.05952518805861473, "mask_loss": 0.5247032642364502, "step": 843 }, { "epoch": 0.5069069069069069, "grad_norm": 22.717222213745117, "learning_rate": 1.0282025098661611e-06, "loss": 0.8604074120521545, "step": 844 }, { "ce_loss": 3.825258681899868e-05, "cls_loss": 0.04541015625, "epoch": 0.5069069069069069, "mask_bce_loss": 0.7393709421157837, "mask_dice_loss": 0.070508673787117, "mask_loss": 0.8098796010017395, "step": 844 }, { "epoch": 0.5075075075075075, "grad_norm": 24.601398468017578, "learning_rate": 1.0262579728891953e-06, "loss": 0.740152895450592, "step": 845 }, { "ce_loss": 4.585788701660931e-05, "cls_loss": 0.06591796875, "epoch": 0.5075075075075075, "mask_bce_loss": 1.7309919595718384, "mask_dice_loss": 0.10072014480829239, "mask_loss": 1.8317121267318726, "step": 845 }, { "epoch": 0.5081081081081081, "grad_norm": 33.53548812866211, "learning_rate": 1.024313336551276e-06, "loss": 1.0259268283843994, "step": 846 }, { "ce_loss": 3.290996392024681e-05, "cls_loss": 0.027587890625, "epoch": 0.5081081081081081, "mask_bce_loss": 0.22906695306301117, "mask_dice_loss": 0.01997736655175686, "mask_loss": 0.24904431402683258, "step": 846 }, { "epoch": 0.5087087087087087, "grad_norm": 21.10654067993164, "learning_rate": 1.0223686082109652e-06, "loss": 0.9758148789405823, "step": 847 }, { "ce_loss": 0.0002577652339823544, "cls_loss": 0.06201171875, "epoch": 0.5087087087087087, "mask_bce_loss": 1.2121329307556152, "mask_dice_loss": 0.14166788756847382, "mask_loss": 1.3538007736206055, "step": 847 }, { "epoch": 0.5093093093093093, "grad_norm": 50.63971710205078, "learning_rate": 1.0204237952271732e-06, "loss": 0.8563938140869141, "step": 848 }, { "ce_loss": 4.83345975226257e-05, "cls_loss": 0.05859375, "epoch": 0.5093093093093093, "mask_bce_loss": 0.31063181161880493, "mask_dice_loss": 0.10822346061468124, "mask_loss": 0.4188552796840668, "step": 848 }, { "epoch": 0.5099099099099099, "grad_norm": 20.21082305908203, "learning_rate": 1.0184789049591297e-06, "loss": 0.7596441507339478, "step": 849 }, { "ce_loss": 0.00011997533874819055, "cls_loss": 0.047607421875, "epoch": 0.5099099099099099, "mask_bce_loss": 0.4790947437286377, "mask_dice_loss": 0.06765373796224594, "mask_loss": 0.5467484593391418, "step": 849 }, { "epoch": 0.5105105105105106, "grad_norm": 29.093442916870117, "learning_rate": 1.0165339447663586e-06, "loss": 0.8075183629989624, "step": 850 }, { "ce_loss": 5.625987614621408e-05, "cls_loss": 0.048828125, "epoch": 0.5105105105105106, "mask_bce_loss": 0.8079559206962585, "mask_dice_loss": 0.15399998426437378, "mask_loss": 0.9619559049606323, "step": 850 }, { "epoch": 0.5111111111111111, "grad_norm": 24.153493881225586, "learning_rate": 1.0145889220086463e-06, "loss": 0.8071024417877197, "step": 851 }, { "ce_loss": 6.854793900856748e-05, "cls_loss": 0.06396484375, "epoch": 0.5111111111111111, "mask_bce_loss": 0.5674439072608948, "mask_dice_loss": 0.11439788341522217, "mask_loss": 0.6818417906761169, "step": 851 }, { "epoch": 0.5117117117117117, "grad_norm": 26.55138397216797, "learning_rate": 1.0126438440460177e-06, "loss": 0.7902085781097412, "step": 852 }, { "ce_loss": 0.00011961787822656333, "cls_loss": 0.0283203125, "epoch": 0.5117117117117117, "mask_bce_loss": 0.13563677668571472, "mask_dice_loss": 0.022001201286911964, "mask_loss": 0.15763798356056213, "step": 852 }, { "epoch": 0.5123123123123123, "grad_norm": 31.50015640258789, "learning_rate": 1.0106987182387055e-06, "loss": 0.8598842620849609, "step": 853 }, { "ce_loss": 3.445296169957146e-05, "cls_loss": 0.05615234375, "epoch": 0.5123123123123123, "mask_bce_loss": 0.7885751128196716, "mask_dice_loss": 0.12771563231945038, "mask_loss": 0.9162907600402832, "step": 853 }, { "epoch": 0.512912912912913, "grad_norm": 30.236703872680664, "learning_rate": 1.008753551947124e-06, "loss": 0.7368901371955872, "step": 854 }, { "ce_loss": 7.914243178674951e-05, "cls_loss": 0.03759765625, "epoch": 0.512912912912913, "mask_bce_loss": 0.32841187715530396, "mask_dice_loss": 0.06674570590257645, "mask_loss": 0.3951575756072998, "step": 854 }, { "epoch": 0.5135135135135135, "grad_norm": 22.714765548706055, "learning_rate": 1.0068083525318402e-06, "loss": 0.9026670455932617, "step": 855 }, { "ce_loss": 3.58599609171506e-05, "cls_loss": 0.053466796875, "epoch": 0.5135135135135135, "mask_bce_loss": 1.009846568107605, "mask_dice_loss": 0.09448233246803284, "mask_loss": 1.1043288707733154, "step": 855 }, { "epoch": 0.5141141141141141, "grad_norm": 24.400253295898438, "learning_rate": 1.0048631273535474e-06, "loss": 0.9129965305328369, "step": 856 }, { "ce_loss": 3.018168172275182e-05, "cls_loss": 0.056640625, "epoch": 0.5141141141141141, "mask_bce_loss": 1.000524878501892, "mask_dice_loss": 0.14277753233909607, "mask_loss": 1.1433024406433105, "step": 856 }, { "epoch": 0.5147147147147147, "grad_norm": 28.567977905273438, "learning_rate": 1.0029178837730359e-06, "loss": 0.7729479670524597, "step": 857 }, { "ce_loss": 0.00036843487760052085, "cls_loss": 0.06298828125, "epoch": 0.5147147147147147, "mask_bce_loss": 0.9652702212333679, "mask_dice_loss": 0.08814065158367157, "mask_loss": 1.0534108877182007, "step": 857 }, { "epoch": 0.5153153153153153, "grad_norm": 69.96674346923828, "learning_rate": 1.0009726291511644e-06, "loss": 0.9053327441215515, "step": 858 }, { "ce_loss": 4.502885349211283e-05, "cls_loss": 0.04150390625, "epoch": 0.5153153153153153, "mask_bce_loss": 0.2888145446777344, "mask_dice_loss": 0.035504501312971115, "mask_loss": 0.3243190348148346, "step": 858 }, { "epoch": 0.5159159159159159, "grad_norm": 29.184865951538086, "learning_rate": 9.990273708488355e-07, "loss": 0.8326783180236816, "step": 859 }, { "ce_loss": 4.3809282942675054e-05, "cls_loss": 0.04443359375, "epoch": 0.5159159159159159, "mask_bce_loss": 0.6319894790649414, "mask_dice_loss": 0.05622578784823418, "mask_loss": 0.6882152557373047, "step": 859 }, { "epoch": 0.5165165165165165, "grad_norm": 38.48680877685547, "learning_rate": 9.970821162269643e-07, "loss": 0.7885708808898926, "step": 860 }, { "ce_loss": 4.999475640943274e-05, "cls_loss": 0.04345703125, "epoch": 0.5165165165165165, "mask_bce_loss": 0.35371729731559753, "mask_dice_loss": 0.05080008506774902, "mask_loss": 0.40451738238334656, "step": 860 }, { "epoch": 0.5171171171171172, "grad_norm": 30.568920135498047, "learning_rate": 9.951368726464525e-07, "loss": 0.8281605839729309, "step": 861 }, { "ce_loss": 0.0004602769622579217, "cls_loss": 0.05859375, "epoch": 0.5171171171171172, "mask_bce_loss": 1.5024018287658691, "mask_dice_loss": 0.13912111520767212, "mask_loss": 1.6415228843688965, "step": 861 }, { "epoch": 0.5177177177177177, "grad_norm": 46.122703552246094, "learning_rate": 9.931916474681597e-07, "loss": 0.8309227824211121, "step": 862 }, { "ce_loss": 0.00019593507749959826, "cls_loss": 0.0263671875, "epoch": 0.5177177177177177, "mask_bce_loss": 0.17793813347816467, "mask_dice_loss": 0.021099640056490898, "mask_loss": 0.19903777539730072, "step": 862 }, { "epoch": 0.5183183183183183, "grad_norm": 17.957788467407227, "learning_rate": 9.912464480528759e-07, "loss": 0.6679586172103882, "step": 863 }, { "ce_loss": 8.048806193983182e-05, "cls_loss": 0.03662109375, "epoch": 0.5183183183183183, "mask_bce_loss": 0.2949976623058319, "mask_dice_loss": 0.029100889340043068, "mask_loss": 0.3240985572338104, "step": 863 }, { "epoch": 0.518918918918919, "grad_norm": 35.685977935791016, "learning_rate": 9.893012817612947e-07, "loss": 0.7123427987098694, "step": 864 }, { "ce_loss": 4.133959737373516e-05, "cls_loss": 0.06787109375, "epoch": 0.518918918918919, "mask_bce_loss": 1.0680338144302368, "mask_dice_loss": 0.13894735276699066, "mask_loss": 1.2069811820983887, "step": 864 }, { "epoch": 0.5195195195195195, "grad_norm": 21.143661499023438, "learning_rate": 9.873561559539824e-07, "loss": 0.7359479665756226, "step": 865 }, { "ce_loss": 0.0003631914150901139, "cls_loss": 0.060546875, "epoch": 0.5195195195195195, "mask_bce_loss": 0.5540327429771423, "mask_dice_loss": 0.09076425433158875, "mask_loss": 0.6447969675064087, "step": 865 }, { "epoch": 0.5201201201201201, "grad_norm": 50.42753219604492, "learning_rate": 9.854110779913536e-07, "loss": 0.8918430209159851, "step": 866 }, { "ce_loss": 8.37200423120521e-05, "cls_loss": 0.049560546875, "epoch": 0.5201201201201201, "mask_bce_loss": 0.28902769088745117, "mask_dice_loss": 0.0715661346912384, "mask_loss": 0.3605938255786896, "step": 866 }, { "epoch": 0.5207207207207207, "grad_norm": 67.13046264648438, "learning_rate": 9.834660552336415e-07, "loss": 0.910014271736145, "step": 867 }, { "ce_loss": 3.388363620615564e-05, "cls_loss": 0.039306640625, "epoch": 0.5207207207207207, "mask_bce_loss": 0.3816138803958893, "mask_dice_loss": 0.03662005811929703, "mask_loss": 0.4182339310646057, "step": 867 }, { "epoch": 0.5213213213213214, "grad_norm": 37.58573532104492, "learning_rate": 9.815210950408702e-07, "loss": 0.8345314264297485, "step": 868 }, { "ce_loss": 4.422701749717817e-05, "cls_loss": 0.05810546875, "epoch": 0.5213213213213214, "mask_bce_loss": 0.2415413111448288, "mask_dice_loss": 0.09483867883682251, "mask_loss": 0.3363800048828125, "step": 868 }, { "epoch": 0.5219219219219219, "grad_norm": 33.0797233581543, "learning_rate": 9.795762047728267e-07, "loss": 0.8032614588737488, "step": 869 }, { "ce_loss": 0.002892470918595791, "cls_loss": 0.06884765625, "epoch": 0.5219219219219219, "mask_bce_loss": 0.8100075125694275, "mask_dice_loss": 0.08270952850580215, "mask_loss": 0.8927170634269714, "step": 869 }, { "epoch": 0.5225225225225225, "grad_norm": 16.891427993774414, "learning_rate": 9.776313917890347e-07, "loss": 0.8266868591308594, "step": 870 }, { "ce_loss": 3.185729292454198e-05, "cls_loss": 0.047607421875, "epoch": 0.5225225225225225, "mask_bce_loss": 0.5573635697364807, "mask_dice_loss": 0.056487273424863815, "mask_loss": 0.6138508319854736, "step": 870 }, { "epoch": 0.5231231231231231, "grad_norm": 65.40937042236328, "learning_rate": 9.75686663448724e-07, "loss": 0.6817811727523804, "step": 871 }, { "ce_loss": 8.570592035539448e-05, "cls_loss": 0.0654296875, "epoch": 0.5231231231231231, "mask_bce_loss": 1.2537956237792969, "mask_dice_loss": 0.17503277957439423, "mask_loss": 1.4288283586502075, "step": 871 }, { "epoch": 0.5237237237237238, "grad_norm": 34.19195556640625, "learning_rate": 9.737420271108046e-07, "loss": 0.7517249584197998, "step": 872 }, { "ce_loss": 3.411685611354187e-05, "cls_loss": 0.03955078125, "epoch": 0.5237237237237238, "mask_bce_loss": 0.4639272689819336, "mask_dice_loss": 0.04532064124941826, "mask_loss": 0.509247899055481, "step": 872 }, { "epoch": 0.5243243243243243, "grad_norm": 19.106714248657227, "learning_rate": 9.71797490133839e-07, "loss": 0.9271371960639954, "step": 873 }, { "ce_loss": 4.682821599999443e-05, "cls_loss": 0.04541015625, "epoch": 0.5243243243243243, "mask_bce_loss": 0.7141562104225159, "mask_dice_loss": 0.059334564954042435, "mask_loss": 0.7734907865524292, "step": 873 }, { "epoch": 0.5249249249249249, "grad_norm": 20.855846405029297, "learning_rate": 9.69853059876013e-07, "loss": 0.7232174873352051, "step": 874 }, { "ce_loss": 4.496042674873024e-05, "cls_loss": 0.05419921875, "epoch": 0.5249249249249249, "mask_bce_loss": 0.6390623450279236, "mask_dice_loss": 0.07556455582380295, "mask_loss": 0.7146269083023071, "step": 874 }, { "epoch": 0.5255255255255256, "grad_norm": 25.69660186767578, "learning_rate": 9.679087436951086e-07, "loss": 0.901782751083374, "step": 875 }, { "ce_loss": 0.00015515563427470624, "cls_loss": 0.038818359375, "epoch": 0.5255255255255256, "mask_bce_loss": 0.33996152877807617, "mask_dice_loss": 0.03054916299879551, "mask_loss": 0.37051069736480713, "step": 875 }, { "epoch": 0.5261261261261261, "grad_norm": 18.20565414428711, "learning_rate": 9.659645489484765e-07, "loss": 0.7443608045578003, "step": 876 }, { "ce_loss": 2.052743366220966e-05, "cls_loss": 0.048095703125, "epoch": 0.5261261261261261, "mask_bce_loss": 0.5337938070297241, "mask_dice_loss": 0.04653644561767578, "mask_loss": 0.5803302526473999, "step": 876 }, { "epoch": 0.5267267267267267, "grad_norm": 21.892433166503906, "learning_rate": 9.640204829930076e-07, "loss": 0.802660346031189, "step": 877 }, { "ce_loss": 0.00032259392901323736, "cls_loss": 0.028076171875, "epoch": 0.5267267267267267, "mask_bce_loss": 0.4069390296936035, "mask_dice_loss": 0.020399419590830803, "mask_loss": 0.42733845114707947, "step": 877 }, { "epoch": 0.5273273273273273, "grad_norm": 37.58069610595703, "learning_rate": 9.620765531851053e-07, "loss": 1.0402790307998657, "step": 878 }, { "ce_loss": 7.384620403172448e-05, "cls_loss": 0.03564453125, "epoch": 0.5273273273273273, "mask_bce_loss": 0.23833179473876953, "mask_dice_loss": 0.07016872614622116, "mask_loss": 0.3085005283355713, "step": 878 }, { "epoch": 0.527927927927928, "grad_norm": 17.957300186157227, "learning_rate": 9.601327668806584e-07, "loss": 0.7356558442115784, "step": 879 }, { "ce_loss": 4.2931886127917096e-05, "cls_loss": 0.03759765625, "epoch": 0.527927927927928, "mask_bce_loss": 0.5976983308792114, "mask_dice_loss": 0.04060070961713791, "mask_loss": 0.6382990479469299, "step": 879 }, { "epoch": 0.5285285285285285, "grad_norm": 34.459598541259766, "learning_rate": 9.58189131435012e-07, "loss": 0.9860548973083496, "step": 880 }, { "ce_loss": 0.00016715661331545562, "cls_loss": 0.05078125, "epoch": 0.5285285285285285, "mask_bce_loss": 0.42197099328041077, "mask_dice_loss": 0.06574326008558273, "mask_loss": 0.4877142608165741, "step": 880 }, { "epoch": 0.5291291291291291, "grad_norm": 33.671993255615234, "learning_rate": 9.562456542029407e-07, "loss": 1.0475773811340332, "step": 881 }, { "ce_loss": 4.096004704479128e-05, "cls_loss": 0.0625, "epoch": 0.5291291291291291, "mask_bce_loss": 0.5140249729156494, "mask_dice_loss": 0.09774182736873627, "mask_loss": 0.6117668151855469, "step": 881 }, { "epoch": 0.5297297297297298, "grad_norm": 51.37704849243164, "learning_rate": 9.54302342538621e-07, "loss": 0.9324228763580322, "step": 882 }, { "ce_loss": 2.6885603801929392e-05, "cls_loss": 0.03271484375, "epoch": 0.5297297297297298, "mask_bce_loss": 0.5987259745597839, "mask_dice_loss": 0.043301939964294434, "mask_loss": 0.6420279145240784, "step": 882 }, { "epoch": 0.5303303303303303, "grad_norm": 15.649553298950195, "learning_rate": 9.523592037956016e-07, "loss": 0.7158737778663635, "step": 883 }, { "ce_loss": 7.987700519151986e-05, "cls_loss": 0.055419921875, "epoch": 0.5303303303303303, "mask_bce_loss": 0.8442670106887817, "mask_dice_loss": 0.104070283472538, "mask_loss": 0.9483373165130615, "step": 883 }, { "epoch": 0.5309309309309309, "grad_norm": 34.19274139404297, "learning_rate": 9.504162453267776e-07, "loss": 0.7928656339645386, "step": 884 }, { "ce_loss": 5.79817860852927e-05, "cls_loss": 0.055419921875, "epoch": 0.5309309309309309, "mask_bce_loss": 1.1700305938720703, "mask_dice_loss": 0.12229558080434799, "mask_loss": 1.2923262119293213, "step": 884 }, { "epoch": 0.5315315315315315, "grad_norm": 18.707578659057617, "learning_rate": 9.48473474484362e-07, "loss": 0.863745927810669, "step": 885 }, { "ce_loss": 2.7141868486069143e-05, "cls_loss": 0.05078125, "epoch": 0.5315315315315315, "mask_bce_loss": 0.6226107478141785, "mask_dice_loss": 0.07079429924488068, "mask_loss": 0.693405032157898, "step": 885 }, { "epoch": 0.5321321321321322, "grad_norm": 38.122798919677734, "learning_rate": 9.46530898619858e-07, "loss": 0.8771089911460876, "step": 886 }, { "ce_loss": 6.212649896042421e-05, "cls_loss": 0.158203125, "epoch": 0.5321321321321322, "mask_bce_loss": 0.8938608169555664, "mask_dice_loss": 0.17783810198307037, "mask_loss": 1.0716989040374756, "step": 886 }, { "epoch": 0.5327327327327327, "grad_norm": 29.575603485107422, "learning_rate": 9.4458852508403e-07, "loss": 0.7722306251525879, "step": 887 }, { "ce_loss": 6.57319906167686e-05, "cls_loss": 0.051513671875, "epoch": 0.5327327327327327, "mask_bce_loss": 1.2305792570114136, "mask_dice_loss": 0.07290336489677429, "mask_loss": 1.3034826517105103, "step": 887 }, { "epoch": 0.5333333333333333, "grad_norm": 22.287635803222656, "learning_rate": 9.42646361226878e-07, "loss": 0.7353468537330627, "step": 888 }, { "ce_loss": 0.00021696290059480816, "cls_loss": 0.057861328125, "epoch": 0.5333333333333333, "mask_bce_loss": 1.3101829290390015, "mask_dice_loss": 0.08148296177387238, "mask_loss": 1.3916659355163574, "step": 888 }, { "epoch": 0.533933933933934, "grad_norm": 20.934537887573242, "learning_rate": 9.407044143976079e-07, "loss": 0.7904185652732849, "step": 889 }, { "ce_loss": 6.11188734183088e-05, "cls_loss": 0.0546875, "epoch": 0.533933933933934, "mask_bce_loss": 1.1919405460357666, "mask_dice_loss": 0.10180097073316574, "mask_loss": 1.2937414646148682, "step": 889 }, { "epoch": 0.5345345345345346, "grad_norm": 38.99357604980469, "learning_rate": 9.38762691944604e-07, "loss": 0.8633276224136353, "step": 890 }, { "ce_loss": 0.00011346025712555274, "cls_loss": 0.06005859375, "epoch": 0.5345345345345346, "mask_bce_loss": 0.7475737929344177, "mask_dice_loss": 0.08053400367498398, "mask_loss": 0.8281077742576599, "step": 890 }, { "epoch": 0.5351351351351351, "grad_norm": 35.33418655395508, "learning_rate": 9.36821201215403e-07, "loss": 0.9271799921989441, "step": 891 }, { "ce_loss": 3.2297066354658455e-05, "cls_loss": 0.05322265625, "epoch": 0.5351351351351351, "mask_bce_loss": 0.6857191920280457, "mask_dice_loss": 0.11605153232812881, "mask_loss": 0.8017707467079163, "step": 891 }, { "epoch": 0.5357357357357357, "grad_norm": 20.757770538330078, "learning_rate": 9.348799495566631e-07, "loss": 0.7955480813980103, "step": 892 }, { "ce_loss": 0.00011101943528046831, "cls_loss": 0.0498046875, "epoch": 0.5357357357357357, "mask_bce_loss": 0.7438855171203613, "mask_dice_loss": 0.07970822602510452, "mask_loss": 0.8235937356948853, "step": 892 }, { "epoch": 0.5363363363363364, "grad_norm": 32.021446228027344, "learning_rate": 9.329389443141388e-07, "loss": 0.8614611625671387, "step": 893 }, { "ce_loss": 2.6801260901265778e-05, "cls_loss": 0.055419921875, "epoch": 0.5363363363363364, "mask_bce_loss": 0.7633022665977478, "mask_dice_loss": 0.11374493688344955, "mask_loss": 0.8770471811294556, "step": 893 }, { "epoch": 0.5369369369369369, "grad_norm": 40.91093063354492, "learning_rate": 9.309981928326519e-07, "loss": 0.8909924626350403, "step": 894 }, { "ce_loss": 0.0001678997214185074, "cls_loss": 0.025146484375, "epoch": 0.5369369369369369, "mask_bce_loss": 0.19393567740917206, "mask_dice_loss": 0.01687963493168354, "mask_loss": 0.21081531047821045, "step": 894 }, { "epoch": 0.5375375375375375, "grad_norm": 28.289548873901367, "learning_rate": 9.290577024560639e-07, "loss": 0.7989416718482971, "step": 895 }, { "ce_loss": 0.00027690024580806494, "cls_loss": 0.04345703125, "epoch": 0.5375375375375375, "mask_bce_loss": 0.4870321750640869, "mask_dice_loss": 0.06296073645353317, "mask_loss": 0.5499929189682007, "step": 895 }, { "epoch": 0.5381381381381382, "grad_norm": 42.6255989074707, "learning_rate": 9.271174805272481e-07, "loss": 0.9131446480751038, "step": 896 }, { "ce_loss": 7.051182183204219e-05, "cls_loss": 0.055419921875, "epoch": 0.5381381381381382, "mask_bce_loss": 0.5471059679985046, "mask_dice_loss": 0.10154356807470322, "mask_loss": 0.6486495137214661, "step": 896 }, { "epoch": 0.5387387387387388, "grad_norm": 24.936628341674805, "learning_rate": 9.251775343880628e-07, "loss": 0.8282215595245361, "step": 897 }, { "ce_loss": 0.00010809156810864806, "cls_loss": 0.05517578125, "epoch": 0.5387387387387388, "mask_bce_loss": 1.3488768339157104, "mask_dice_loss": 0.09263870120048523, "mask_loss": 1.441515564918518, "step": 897 }, { "epoch": 0.5393393393393393, "grad_norm": 24.751487731933594, "learning_rate": 9.232378713793217e-07, "loss": 0.814694344997406, "step": 898 }, { "ce_loss": 7.736702536931261e-05, "cls_loss": 0.0289306640625, "epoch": 0.5393393393393393, "mask_bce_loss": 0.16888320446014404, "mask_dice_loss": 0.02134360745549202, "mask_loss": 0.19022680819034576, "step": 898 }, { "epoch": 0.5399399399399399, "grad_norm": 17.848419189453125, "learning_rate": 9.212984988407673e-07, "loss": 0.6844253540039062, "step": 899 }, { "ce_loss": 5.926475205342285e-05, "cls_loss": 0.052734375, "epoch": 0.5399399399399399, "mask_bce_loss": 0.7134351134300232, "mask_dice_loss": 0.1284937560558319, "mask_loss": 0.8419288396835327, "step": 899 }, { "epoch": 0.5405405405405406, "grad_norm": 21.68507194519043, "learning_rate": 9.193594241110438e-07, "loss": 0.7717435956001282, "step": 900 }, { "ce_loss": 5.174529360374436e-05, "cls_loss": 0.04736328125, "epoch": 0.5405405405405406, "mask_bce_loss": 0.6150922775268555, "mask_dice_loss": 0.06294342130422592, "mask_loss": 0.6780356764793396, "step": 900 }, { "epoch": 0.5411411411411411, "grad_norm": 19.17519187927246, "learning_rate": 9.174206545276677e-07, "loss": 0.7414667010307312, "step": 901 }, { "ce_loss": 0.0003250775916967541, "cls_loss": 0.04345703125, "epoch": 0.5411411411411411, "mask_bce_loss": 0.5350953340530396, "mask_dice_loss": 0.043712317943573, "mask_loss": 0.5788076519966125, "step": 901 }, { "epoch": 0.5417417417417417, "grad_norm": 19.06415557861328, "learning_rate": 9.154821974270006e-07, "loss": 0.6710140705108643, "step": 902 }, { "ce_loss": 2.618205144244712e-05, "cls_loss": 0.05908203125, "epoch": 0.5417417417417417, "mask_bce_loss": 0.3206063210964203, "mask_dice_loss": 0.079197958111763, "mask_loss": 0.3998042941093445, "step": 902 }, { "epoch": 0.5423423423423424, "grad_norm": 27.676746368408203, "learning_rate": 9.135440601442228e-07, "loss": 0.924116849899292, "step": 903 }, { "ce_loss": 0.0002169852814404294, "cls_loss": 0.0654296875, "epoch": 0.5423423423423424, "mask_bce_loss": 1.9155393838882446, "mask_dice_loss": 0.11449974775314331, "mask_loss": 2.030039072036743, "step": 903 }, { "epoch": 0.542942942942943, "grad_norm": 56.80961608886719, "learning_rate": 9.116062500133031e-07, "loss": 0.8865700364112854, "step": 904 }, { "ce_loss": 0.0002304184454260394, "cls_loss": 0.064453125, "epoch": 0.542942942942943, "mask_bce_loss": 0.467092365026474, "mask_dice_loss": 0.09194651991128922, "mask_loss": 0.5590388774871826, "step": 904 }, { "epoch": 0.5435435435435435, "grad_norm": 67.80094909667969, "learning_rate": 9.096687743669728e-07, "loss": 0.8819120526313782, "step": 905 }, { "ce_loss": 6.516441499115899e-05, "cls_loss": 0.051513671875, "epoch": 0.5435435435435435, "mask_bce_loss": 0.6850938200950623, "mask_dice_loss": 0.06171974167227745, "mask_loss": 0.7468135356903076, "step": 905 }, { "epoch": 0.5441441441441441, "grad_norm": 48.01819610595703, "learning_rate": 9.077316405366981e-07, "loss": 0.7991788387298584, "step": 906 }, { "ce_loss": 2.5759745767572895e-05, "cls_loss": 0.032470703125, "epoch": 0.5441441441441441, "mask_bce_loss": 0.6784807443618774, "mask_dice_loss": 0.060548994690179825, "mask_loss": 0.7390297651290894, "step": 906 }, { "epoch": 0.5447447447447448, "grad_norm": 18.84556770324707, "learning_rate": 9.057948558526509e-07, "loss": 0.9305986762046814, "step": 907 }, { "ce_loss": 4.180889663984999e-05, "cls_loss": 0.0546875, "epoch": 0.5447447447447448, "mask_bce_loss": 0.7372807860374451, "mask_dice_loss": 0.2127833217382431, "mask_loss": 0.9500641226768494, "step": 907 }, { "epoch": 0.5453453453453454, "grad_norm": 21.736562728881836, "learning_rate": 9.03858427643682e-07, "loss": 0.8271975517272949, "step": 908 }, { "ce_loss": 0.010493166744709015, "cls_loss": 0.0712890625, "epoch": 0.5453453453453454, "mask_bce_loss": 2.48679256439209, "mask_dice_loss": 0.07201839238405228, "mask_loss": 2.5588109493255615, "step": 908 }, { "epoch": 0.5459459459459459, "grad_norm": 19.17778205871582, "learning_rate": 9.019223632372945e-07, "loss": 0.8454885482788086, "step": 909 }, { "ce_loss": 3.044412733288482e-05, "cls_loss": 0.033447265625, "epoch": 0.5459459459459459, "mask_bce_loss": 0.3153807520866394, "mask_dice_loss": 0.02546573616564274, "mask_loss": 0.3408464789390564, "step": 909 }, { "epoch": 0.5465465465465466, "grad_norm": 37.423126220703125, "learning_rate": 8.999866699596132e-07, "loss": 0.8072248697280884, "step": 910 }, { "ce_loss": 9.005561878439039e-05, "cls_loss": 0.05712890625, "epoch": 0.5465465465465466, "mask_bce_loss": 0.487324059009552, "mask_dice_loss": 0.14740577340126038, "mask_loss": 0.6347298622131348, "step": 910 }, { "epoch": 0.5471471471471472, "grad_norm": 51.99261474609375, "learning_rate": 8.980513551353593e-07, "loss": 0.9391722083091736, "step": 911 }, { "ce_loss": 6.665073306066915e-05, "cls_loss": 0.057861328125, "epoch": 0.5471471471471472, "mask_bce_loss": 1.1031583547592163, "mask_dice_loss": 0.142679825425148, "mask_loss": 1.2458381652832031, "step": 911 }, { "epoch": 0.5477477477477477, "grad_norm": 35.36312484741211, "learning_rate": 8.961164260878223e-07, "loss": 0.9439519643783569, "step": 912 }, { "ce_loss": 3.521580219967291e-05, "cls_loss": 0.033935546875, "epoch": 0.5477477477477477, "mask_bce_loss": 0.40288835763931274, "mask_dice_loss": 0.028218314051628113, "mask_loss": 0.43110668659210205, "step": 912 }, { "epoch": 0.5483483483483483, "grad_norm": 20.627573013305664, "learning_rate": 8.941818901388312e-07, "loss": 0.8436477184295654, "step": 913 }, { "ce_loss": 0.0012732446193695068, "cls_loss": 0.06103515625, "epoch": 0.5483483483483483, "mask_bce_loss": 0.5749018788337708, "mask_dice_loss": 0.06280239671468735, "mask_loss": 0.6377042531967163, "step": 913 }, { "epoch": 0.548948948948949, "grad_norm": 69.59999084472656, "learning_rate": 8.922477546087277e-07, "loss": 0.8879700899124146, "step": 914 }, { "ce_loss": 0.0001209136244142428, "cls_loss": 0.057861328125, "epoch": 0.548948948948949, "mask_bce_loss": 0.8868884444236755, "mask_dice_loss": 0.12265368551015854, "mask_loss": 1.0095421075820923, "step": 914 }, { "epoch": 0.5495495495495496, "grad_norm": 24.32650375366211, "learning_rate": 8.903140268163387e-07, "loss": 0.753529965877533, "step": 915 }, { "ce_loss": 4.6392568037845194e-05, "cls_loss": 0.024658203125, "epoch": 0.5495495495495496, "mask_bce_loss": 0.24971389770507812, "mask_dice_loss": 0.019900111481547356, "mask_loss": 0.26961401104927063, "step": 915 }, { "epoch": 0.5501501501501501, "grad_norm": 25.19508934020996, "learning_rate": 8.883807140789478e-07, "loss": 0.7016125321388245, "step": 916 }, { "ce_loss": 3.7049201637273654e-05, "cls_loss": 0.056640625, "epoch": 0.5501501501501501, "mask_bce_loss": 0.4390583038330078, "mask_dice_loss": 0.07165568321943283, "mask_loss": 0.5107139945030212, "step": 916 }, { "epoch": 0.5507507507507508, "grad_norm": 51.49658966064453, "learning_rate": 8.864478237122676e-07, "loss": 1.0338881015777588, "step": 917 }, { "ce_loss": 0.006705781910568476, "cls_loss": 0.0869140625, "epoch": 0.5507507507507508, "mask_bce_loss": 1.356727957725525, "mask_dice_loss": 0.144440159201622, "mask_loss": 1.501168131828308, "step": 917 }, { "epoch": 0.5513513513513514, "grad_norm": 14.359597206115723, "learning_rate": 8.845153630304139e-07, "loss": 0.7223151326179504, "step": 918 }, { "ce_loss": 8.565148891648278e-05, "cls_loss": 0.0458984375, "epoch": 0.5513513513513514, "mask_bce_loss": 0.9114193916320801, "mask_dice_loss": 0.12972868978977203, "mask_loss": 1.041148066520691, "step": 918 }, { "epoch": 0.5519519519519519, "grad_norm": 21.861080169677734, "learning_rate": 8.825833393458754e-07, "loss": 0.8439863324165344, "step": 919 }, { "ce_loss": 2.4343102268176153e-05, "cls_loss": 0.05810546875, "epoch": 0.5519519519519519, "mask_bce_loss": 0.6491317749023438, "mask_dice_loss": 0.10697569698095322, "mask_loss": 0.7561074495315552, "step": 919 }, { "epoch": 0.5525525525525525, "grad_norm": 21.76628303527832, "learning_rate": 8.806517599694865e-07, "loss": 0.8210608959197998, "step": 920 }, { "ce_loss": 3.326403384562582e-05, "cls_loss": 0.051513671875, "epoch": 0.5525525525525525, "mask_bce_loss": 0.38183537125587463, "mask_dice_loss": 0.11979057639837265, "mask_loss": 0.5016259551048279, "step": 920 }, { "epoch": 0.5531531531531532, "grad_norm": 24.60886001586914, "learning_rate": 8.787206322104024e-07, "loss": 0.9185322523117065, "step": 921 }, { "ce_loss": 3.303736957604997e-05, "cls_loss": 0.0673828125, "epoch": 0.5531531531531532, "mask_bce_loss": 1.2713521718978882, "mask_dice_loss": 0.11411335319280624, "mask_loss": 1.3854655027389526, "step": 921 }, { "epoch": 0.5537537537537538, "grad_norm": 31.23426628112793, "learning_rate": 8.767899633760678e-07, "loss": 0.9756548404693604, "step": 922 }, { "ce_loss": 4.1432296711718664e-05, "cls_loss": 0.0439453125, "epoch": 0.5537537537537538, "mask_bce_loss": 0.9450680017471313, "mask_dice_loss": 0.09594502300024033, "mask_loss": 1.0410130023956299, "step": 922 }, { "epoch": 0.5543543543543543, "grad_norm": 19.82107925415039, "learning_rate": 8.748597607721909e-07, "loss": 0.8300458192825317, "step": 923 }, { "ce_loss": 5.6094453611876816e-05, "cls_loss": 0.04833984375, "epoch": 0.5543543543543543, "mask_bce_loss": 0.570803165435791, "mask_dice_loss": 0.0777352973818779, "mask_loss": 0.6485384702682495, "step": 923 }, { "epoch": 0.554954954954955, "grad_norm": 26.0727481842041, "learning_rate": 8.729300317027167e-07, "loss": 0.8417010307312012, "step": 924 }, { "ce_loss": 0.00012605699885170907, "cls_loss": 0.04052734375, "epoch": 0.554954954954955, "mask_bce_loss": 0.5994707942008972, "mask_dice_loss": 0.04065682366490364, "mask_loss": 0.6401275992393494, "step": 924 }, { "epoch": 0.5555555555555556, "grad_norm": 26.373823165893555, "learning_rate": 8.710007834697969e-07, "loss": 0.8930571675300598, "step": 925 }, { "ce_loss": 0.00015274369798135012, "cls_loss": 0.03466796875, "epoch": 0.5555555555555556, "mask_bce_loss": 0.5022472143173218, "mask_dice_loss": 0.07255083322525024, "mask_loss": 0.574798047542572, "step": 925 }, { "epoch": 0.5561561561561561, "grad_norm": 24.228200912475586, "learning_rate": 8.690720233737644e-07, "loss": 0.9766445159912109, "step": 926 }, { "ce_loss": 5.565558603848331e-05, "cls_loss": 0.0546875, "epoch": 0.5561561561561561, "mask_bce_loss": 0.5473772287368774, "mask_dice_loss": 0.12186700105667114, "mask_loss": 0.6692442297935486, "step": 926 }, { "epoch": 0.5567567567567567, "grad_norm": 24.38997459411621, "learning_rate": 8.671437587131059e-07, "loss": 0.8554730415344238, "step": 927 }, { "ce_loss": 5.256813528831117e-05, "cls_loss": 0.03955078125, "epoch": 0.5567567567567567, "mask_bce_loss": 1.0381382703781128, "mask_dice_loss": 0.08376877754926682, "mask_loss": 1.1219069957733154, "step": 927 }, { "epoch": 0.5573573573573574, "grad_norm": 31.562734603881836, "learning_rate": 8.652159967844319e-07, "loss": 0.7831336259841919, "step": 928 }, { "ce_loss": 0.013235621154308319, "cls_loss": 0.040771484375, "epoch": 0.5573573573573574, "mask_bce_loss": 0.926792323589325, "mask_dice_loss": 0.03888244554400444, "mask_loss": 0.9656747579574585, "step": 928 }, { "epoch": 0.557957957957958, "grad_norm": 30.29644012451172, "learning_rate": 8.632887448824506e-07, "loss": 0.9319536685943604, "step": 929 }, { "ce_loss": 5.409381992649287e-05, "cls_loss": 0.255859375, "epoch": 0.557957957957958, "mask_bce_loss": 0.8440003991127014, "mask_dice_loss": 0.1218344047665596, "mask_loss": 0.9658347964286804, "step": 929 }, { "epoch": 0.5585585585585585, "grad_norm": 36.84908676147461, "learning_rate": 8.613620102999415e-07, "loss": 0.9124982357025146, "step": 930 }, { "ce_loss": 4.6401986764976755e-05, "cls_loss": 0.07275390625, "epoch": 0.5585585585585585, "mask_bce_loss": 2.272731065750122, "mask_dice_loss": 0.13726456463336945, "mask_loss": 2.4099955558776855, "step": 930 }, { "epoch": 0.5591591591591591, "grad_norm": 21.40240478515625, "learning_rate": 8.594358003277257e-07, "loss": 0.9126743078231812, "step": 931 }, { "ce_loss": 2.6205343601759523e-05, "cls_loss": 0.053955078125, "epoch": 0.5591591591591591, "mask_bce_loss": 1.0504041910171509, "mask_dice_loss": 0.06656774133443832, "mask_loss": 1.1169719696044922, "step": 931 }, { "epoch": 0.5597597597597598, "grad_norm": 37.32105255126953, "learning_rate": 8.575101222546388e-07, "loss": 0.8583958745002747, "step": 932 }, { "ce_loss": 6.552882405230775e-05, "cls_loss": 0.0546875, "epoch": 0.5597597597597598, "mask_bce_loss": 0.8798627257347107, "mask_dice_loss": 0.14450369775295258, "mask_loss": 1.0243663787841797, "step": 932 }, { "epoch": 0.5603603603603604, "grad_norm": 36.63002395629883, "learning_rate": 8.555849833675046e-07, "loss": 0.8702892065048218, "step": 933 }, { "ce_loss": 0.00014651156379841268, "cls_loss": 0.06298828125, "epoch": 0.5603603603603604, "mask_bce_loss": 1.0324498414993286, "mask_dice_loss": 0.11343803256750107, "mask_loss": 1.145887851715088, "step": 933 }, { "epoch": 0.5609609609609609, "grad_norm": 29.6329402923584, "learning_rate": 8.536603909511061e-07, "loss": 0.8130438923835754, "step": 934 }, { "ce_loss": 2.450186002533883e-05, "cls_loss": 0.04296875, "epoch": 0.5609609609609609, "mask_bce_loss": 0.5465928912162781, "mask_dice_loss": 0.0379069447517395, "mask_loss": 0.5844998359680176, "step": 934 }, { "epoch": 0.5615615615615616, "grad_norm": 28.558380126953125, "learning_rate": 8.517363522881578e-07, "loss": 0.9282448291778564, "step": 935 }, { "ce_loss": 3.155814192723483e-05, "cls_loss": 0.031005859375, "epoch": 0.5615615615615616, "mask_bce_loss": 0.3976843059062958, "mask_dice_loss": 0.05315232276916504, "mask_loss": 0.4508366286754608, "step": 935 }, { "epoch": 0.5621621621621622, "grad_norm": 42.10542678833008, "learning_rate": 8.498128746592806e-07, "loss": 0.8209486603736877, "step": 936 }, { "ce_loss": 5.248377055977471e-05, "cls_loss": 0.06005859375, "epoch": 0.5621621621621622, "mask_bce_loss": 1.430846929550171, "mask_dice_loss": 0.08486870676279068, "mask_loss": 1.5157155990600586, "step": 936 }, { "epoch": 0.5627627627627627, "grad_norm": 24.05954933166504, "learning_rate": 8.478899653429707e-07, "loss": 0.7106018662452698, "step": 937 }, { "ce_loss": 0.00031145778484642506, "cls_loss": 0.0341796875, "epoch": 0.5627627627627627, "mask_bce_loss": 0.34347909688949585, "mask_dice_loss": 0.025600118562579155, "mask_loss": 0.36907920241355896, "step": 937 }, { "epoch": 0.5633633633633633, "grad_norm": 23.373172760009766, "learning_rate": 8.459676316155739e-07, "loss": 0.8050915598869324, "step": 938 }, { "ce_loss": 5.5948985391296446e-05, "cls_loss": 0.033447265625, "epoch": 0.5633633633633633, "mask_bce_loss": 0.29558178782463074, "mask_dice_loss": 0.04663660749793053, "mask_loss": 0.34221839904785156, "step": 938 }, { "epoch": 0.563963963963964, "grad_norm": 42.776702880859375, "learning_rate": 8.440458807512592e-07, "loss": 0.937503457069397, "step": 939 }, { "ce_loss": 4.1226398025173694e-05, "cls_loss": 0.1328125, "epoch": 0.563963963963964, "mask_bce_loss": 0.7650752067565918, "mask_dice_loss": 0.12851549685001373, "mask_loss": 0.8935906887054443, "step": 939 }, { "epoch": 0.5645645645645646, "grad_norm": 44.14978790283203, "learning_rate": 8.421247200219892e-07, "loss": 0.7802547812461853, "step": 940 }, { "ce_loss": 2.9374417863436975e-05, "cls_loss": 0.041015625, "epoch": 0.5645645645645646, "mask_bce_loss": 0.6973505020141602, "mask_dice_loss": 0.0661781057715416, "mask_loss": 0.76352858543396, "step": 940 }, { "epoch": 0.5651651651651651, "grad_norm": 24.461841583251953, "learning_rate": 8.402041566974928e-07, "loss": 0.7710436582565308, "step": 941 }, { "ce_loss": 9.403523290529847e-05, "cls_loss": 0.0546875, "epoch": 0.5651651651651651, "mask_bce_loss": 0.7432957887649536, "mask_dice_loss": 0.057800889015197754, "mask_loss": 0.8010966777801514, "step": 941 }, { "epoch": 0.5657657657657658, "grad_norm": 50.16414260864258, "learning_rate": 8.382841980452397e-07, "loss": 0.9146883487701416, "step": 942 }, { "ce_loss": 6.957189179956913e-05, "cls_loss": 0.03955078125, "epoch": 0.5657657657657658, "mask_bce_loss": 0.5557332634925842, "mask_dice_loss": 0.038583673536777496, "mask_loss": 0.5943169593811035, "step": 942 }, { "epoch": 0.5663663663663664, "grad_norm": 52.51352310180664, "learning_rate": 8.363648513304104e-07, "loss": 0.8132705092430115, "step": 943 }, { "ce_loss": 0.00011099416587967426, "cls_loss": 0.053955078125, "epoch": 0.5663663663663664, "mask_bce_loss": 0.5553451180458069, "mask_dice_loss": 0.09072371572256088, "mask_loss": 0.646068811416626, "step": 943 }, { "epoch": 0.5669669669669669, "grad_norm": 29.020078659057617, "learning_rate": 8.344461238158699e-07, "loss": 0.9914638996124268, "step": 944 }, { "ce_loss": 0.00017343423678539693, "cls_loss": 0.050048828125, "epoch": 0.5669669669669669, "mask_bce_loss": 0.5320128202438354, "mask_dice_loss": 0.05186088755726814, "mask_loss": 0.5838736891746521, "step": 944 }, { "epoch": 0.5675675675675675, "grad_norm": 13.399406433105469, "learning_rate": 8.325280227621412e-07, "loss": 0.6448794007301331, "step": 945 }, { "ce_loss": 3.82110447390005e-05, "cls_loss": 0.039794921875, "epoch": 0.5675675675675675, "mask_bce_loss": 0.391617089509964, "mask_dice_loss": 0.05524887517094612, "mask_loss": 0.446865975856781, "step": 945 }, { "epoch": 0.5681681681681682, "grad_norm": 32.71002960205078, "learning_rate": 8.306105554273757e-07, "loss": 0.8656740188598633, "step": 946 }, { "ce_loss": 4.236535096424632e-05, "cls_loss": 0.03125, "epoch": 0.5681681681681682, "mask_bce_loss": 0.4704546630382538, "mask_dice_loss": 0.023756330832839012, "mask_loss": 0.49421098828315735, "step": 946 }, { "epoch": 0.5687687687687688, "grad_norm": 33.10431671142578, "learning_rate": 8.286937290673263e-07, "loss": 0.756088137626648, "step": 947 }, { "ce_loss": 0.401785671710968, "cls_loss": 0.05517578125, "epoch": 0.5687687687687688, "mask_bce_loss": 0.5238205194473267, "mask_dice_loss": 0.08640103787183762, "mask_loss": 0.6102215647697449, "step": 947 }, { "epoch": 0.5693693693693693, "grad_norm": 55.776222229003906, "learning_rate": 8.267775509353223e-07, "loss": 0.9278581142425537, "step": 948 }, { "ce_loss": 4.840605106437579e-05, "cls_loss": 0.03173828125, "epoch": 0.5693693693693693, "mask_bce_loss": 0.2592821419239044, "mask_dice_loss": 0.048910971730947495, "mask_loss": 0.3081931173801422, "step": 948 }, { "epoch": 0.56996996996997, "grad_norm": 28.0798397064209, "learning_rate": 8.248620282822384e-07, "loss": 0.7391360998153687, "step": 949 }, { "ce_loss": 0.009119877591729164, "cls_loss": 0.05419921875, "epoch": 0.56996996996997, "mask_bce_loss": 0.8525142669677734, "mask_dice_loss": 0.06989981979131699, "mask_loss": 0.9224140644073486, "step": 949 }, { "epoch": 0.5705705705705706, "grad_norm": 37.93343734741211, "learning_rate": 8.229471683564693e-07, "loss": 0.8598083257675171, "step": 950 }, { "ce_loss": 5.534239608095959e-05, "cls_loss": 0.048583984375, "epoch": 0.5705705705705706, "mask_bce_loss": 0.6956745982170105, "mask_dice_loss": 0.07476296275854111, "mask_loss": 0.7704375386238098, "step": 950 }, { "epoch": 0.5711711711711712, "grad_norm": 45.1746940612793, "learning_rate": 8.210329784039027e-07, "loss": 0.7971339821815491, "step": 951 }, { "ce_loss": 3.837201074929908e-05, "cls_loss": 0.04736328125, "epoch": 0.5711711711711712, "mask_bce_loss": 0.5655588507652283, "mask_dice_loss": 0.06710483878850937, "mask_loss": 0.6326636672019958, "step": 951 }, { "epoch": 0.5717717717717717, "grad_norm": 23.77022933959961, "learning_rate": 8.191194656678904e-07, "loss": 0.887195348739624, "step": 952 }, { "ce_loss": 2.802965718728956e-05, "cls_loss": 0.0634765625, "epoch": 0.5717717717717717, "mask_bce_loss": 0.6611394882202148, "mask_dice_loss": 0.10289537906646729, "mask_loss": 0.7640348672866821, "step": 952 }, { "epoch": 0.5723723723723724, "grad_norm": 40.967193603515625, "learning_rate": 8.17206637389221e-07, "loss": 0.8528561592102051, "step": 953 }, { "ce_loss": 5.139516360941343e-05, "cls_loss": 0.0771484375, "epoch": 0.5723723723723724, "mask_bce_loss": 0.5365808010101318, "mask_dice_loss": 0.05300154909491539, "mask_loss": 0.5895823240280151, "step": 953 }, { "epoch": 0.572972972972973, "grad_norm": 83.88041687011719, "learning_rate": 8.152945008060952e-07, "loss": 0.9168552160263062, "step": 954 }, { "ce_loss": 7.859268225729465e-05, "cls_loss": 0.033935546875, "epoch": 0.572972972972973, "mask_bce_loss": 0.47585469484329224, "mask_dice_loss": 0.0429689958691597, "mask_loss": 0.5188236832618713, "step": 954 }, { "epoch": 0.5735735735735735, "grad_norm": 41.93610382080078, "learning_rate": 8.133830631540939e-07, "loss": 0.8529818058013916, "step": 955 }, { "ce_loss": 2.3356149540632032e-05, "cls_loss": 0.07080078125, "epoch": 0.5735735735735735, "mask_bce_loss": 0.38414421677589417, "mask_dice_loss": 0.16267235577106476, "mask_loss": 0.5468165874481201, "step": 955 }, { "epoch": 0.5741741741741742, "grad_norm": 33.81376266479492, "learning_rate": 8.11472331666154e-07, "loss": 0.7434039115905762, "step": 956 }, { "ce_loss": 0.005444166250526905, "cls_loss": 0.05712890625, "epoch": 0.5741741741741742, "mask_bce_loss": 1.1530464887619019, "mask_dice_loss": 0.18662488460540771, "mask_loss": 1.3396713733673096, "step": 956 }, { "epoch": 0.5747747747747748, "grad_norm": 28.239988327026367, "learning_rate": 8.095623135725416e-07, "loss": 0.8617711067199707, "step": 957 }, { "ce_loss": 0.00319283758290112, "cls_loss": 0.06005859375, "epoch": 0.5747747747747748, "mask_bce_loss": 0.34955036640167236, "mask_dice_loss": 0.10969340801239014, "mask_loss": 0.4592437744140625, "step": 957 }, { "epoch": 0.5753753753753754, "grad_norm": 29.81365203857422, "learning_rate": 8.076530161008216e-07, "loss": 0.9186460375785828, "step": 958 }, { "ce_loss": 9.298163058701903e-05, "cls_loss": 0.0634765625, "epoch": 0.5753753753753754, "mask_bce_loss": 0.4473191797733307, "mask_dice_loss": 0.14231517910957336, "mask_loss": 0.589634358882904, "step": 958 }, { "epoch": 0.5759759759759759, "grad_norm": 44.396907806396484, "learning_rate": 8.057444464758326e-07, "loss": 0.7808941006660461, "step": 959 }, { "ce_loss": 5.253294148133136e-05, "cls_loss": 0.05859375, "epoch": 0.5759759759759759, "mask_bce_loss": 0.8991150856018066, "mask_dice_loss": 0.1371840387582779, "mask_loss": 1.0362991094589233, "step": 959 }, { "epoch": 0.5765765765765766, "grad_norm": 16.369590759277344, "learning_rate": 8.038366119196591e-07, "loss": 0.8458569049835205, "step": 960 }, { "ce_loss": 0.002407932188361883, "cls_loss": 0.06201171875, "epoch": 0.5765765765765766, "mask_bce_loss": 0.44779524207115173, "mask_dice_loss": 0.0858987495303154, "mask_loss": 0.5336939692497253, "step": 960 }, { "epoch": 0.5771771771771772, "grad_norm": 36.7558708190918, "learning_rate": 8.019295196516043e-07, "loss": 0.8408716320991516, "step": 961 }, { "ce_loss": 2.8339214622974396e-05, "cls_loss": 0.051513671875, "epoch": 0.5771771771771772, "mask_bce_loss": 0.4487581253051758, "mask_dice_loss": 0.05530276522040367, "mask_loss": 0.5040608644485474, "step": 961 }, { "epoch": 0.5777777777777777, "grad_norm": 33.04008865356445, "learning_rate": 8.000231768881614e-07, "loss": 0.9998862743377686, "step": 962 }, { "ce_loss": 0.00022204077686183155, "cls_loss": 0.0380859375, "epoch": 0.5777777777777777, "mask_bce_loss": 0.3523685038089752, "mask_dice_loss": 0.07443717122077942, "mask_loss": 0.42680567502975464, "step": 962 }, { "epoch": 0.5783783783783784, "grad_norm": 28.0270938873291, "learning_rate": 7.9811759084299e-07, "loss": 0.8918143510818481, "step": 963 }, { "ce_loss": 0.00015020565479062498, "cls_loss": 0.05126953125, "epoch": 0.5783783783783784, "mask_bce_loss": 1.024574875831604, "mask_dice_loss": 0.07899110019207001, "mask_loss": 1.1035659313201904, "step": 963 }, { "epoch": 0.578978978978979, "grad_norm": 15.442102432250977, "learning_rate": 7.962127687268832e-07, "loss": 0.7019041776657104, "step": 964 }, { "ce_loss": 0.0570349283516407, "cls_loss": 0.06689453125, "epoch": 0.578978978978979, "mask_bce_loss": 0.7567554116249084, "mask_dice_loss": 0.16394494473934174, "mask_loss": 0.9207003712654114, "step": 964 }, { "epoch": 0.5795795795795796, "grad_norm": 16.074575424194336, "learning_rate": 7.943087177477447e-07, "loss": 0.7734538316726685, "step": 965 }, { "ce_loss": 0.0011835531331598759, "cls_loss": 0.05517578125, "epoch": 0.5795795795795796, "mask_bce_loss": 0.6239342093467712, "mask_dice_loss": 0.08795776963233948, "mask_loss": 0.7118920087814331, "step": 965 }, { "epoch": 0.5801801801801801, "grad_norm": 37.17163848876953, "learning_rate": 7.924054451105613e-07, "loss": 1.0136288404464722, "step": 966 }, { "ce_loss": 0.0001144443522207439, "cls_loss": 0.0390625, "epoch": 0.5801801801801801, "mask_bce_loss": 0.6094140410423279, "mask_dice_loss": 0.06183898448944092, "mask_loss": 0.6712530255317688, "step": 966 }, { "epoch": 0.5807807807807808, "grad_norm": 37.79874801635742, "learning_rate": 7.90502958017373e-07, "loss": 0.8369027972221375, "step": 967 }, { "ce_loss": 0.0014333253493532538, "cls_loss": 0.0712890625, "epoch": 0.5807807807807808, "mask_bce_loss": 0.615972638130188, "mask_dice_loss": 0.13585910201072693, "mask_loss": 0.7518317699432373, "step": 967 }, { "epoch": 0.5813813813813814, "grad_norm": 35.506778717041016, "learning_rate": 7.886012636672473e-07, "loss": 0.8275099992752075, "step": 968 }, { "ce_loss": 0.0007559915538877249, "cls_loss": 0.04833984375, "epoch": 0.5813813813813814, "mask_bce_loss": 0.7724649310112, "mask_dice_loss": 0.06113723665475845, "mask_loss": 0.8336021900177002, "step": 968 }, { "epoch": 0.581981981981982, "grad_norm": 30.18828773498535, "learning_rate": 7.867003692562533e-07, "loss": 0.877679169178009, "step": 969 }, { "ce_loss": 4.599009480443783e-05, "cls_loss": 0.035400390625, "epoch": 0.581981981981982, "mask_bce_loss": 0.6610850691795349, "mask_dice_loss": 0.029004082083702087, "mask_loss": 0.6900891661643982, "step": 969 }, { "epoch": 0.5825825825825826, "grad_norm": 21.366785049438477, "learning_rate": 7.848002819774315e-07, "loss": 0.7493525743484497, "step": 970 }, { "ce_loss": 9.974176646210253e-05, "cls_loss": 0.05712890625, "epoch": 0.5825825825825826, "mask_bce_loss": 1.5302757024765015, "mask_dice_loss": 0.17791281640529633, "mask_loss": 1.708188533782959, "step": 970 }, { "epoch": 0.5831831831831832, "grad_norm": 20.77702522277832, "learning_rate": 7.829010090207689e-07, "loss": 0.8974621295928955, "step": 971 }, { "ce_loss": 3.237720375182107e-05, "cls_loss": 0.052001953125, "epoch": 0.5831831831831832, "mask_bce_loss": 0.44449755549430847, "mask_dice_loss": 0.05140373855829239, "mask_loss": 0.49590128660202026, "step": 971 }, { "epoch": 0.5837837837837838, "grad_norm": 36.91118240356445, "learning_rate": 7.81002557573172e-07, "loss": 0.89838045835495, "step": 972 }, { "ce_loss": 4.936398545396514e-05, "cls_loss": 0.0206298828125, "epoch": 0.5837837837837838, "mask_bce_loss": 0.1294286996126175, "mask_dice_loss": 0.014466339722275734, "mask_loss": 0.14389504492282867, "step": 972 }, { "epoch": 0.5843843843843843, "grad_norm": 21.36758804321289, "learning_rate": 7.791049348184369e-07, "loss": 0.8408015370368958, "step": 973 }, { "ce_loss": 4.00294920837041e-05, "cls_loss": 0.07177734375, "epoch": 0.5843843843843843, "mask_bce_loss": 0.5446999073028564, "mask_dice_loss": 0.12933282554149628, "mask_loss": 0.6740327477455139, "step": 973 }, { "epoch": 0.584984984984985, "grad_norm": 27.554250717163086, "learning_rate": 7.772081479372245e-07, "loss": 0.7759069204330444, "step": 974 }, { "ce_loss": 0.0009221665095537901, "cls_loss": 0.0400390625, "epoch": 0.584984984984985, "mask_bce_loss": 0.593469500541687, "mask_dice_loss": 0.0665060356259346, "mask_loss": 0.659975528717041, "step": 974 }, { "epoch": 0.5855855855855856, "grad_norm": 22.499366760253906, "learning_rate": 7.753122041070339e-07, "loss": 0.826788067817688, "step": 975 }, { "ce_loss": 3.1872208637651056e-05, "cls_loss": 0.057861328125, "epoch": 0.5855855855855856, "mask_bce_loss": 0.6914511919021606, "mask_dice_loss": 0.07028087228536606, "mask_loss": 0.7617320418357849, "step": 975 }, { "epoch": 0.5861861861861862, "grad_norm": 25.426830291748047, "learning_rate": 7.734171105021728e-07, "loss": 0.8505719900131226, "step": 976 }, { "ce_loss": 0.0008544019074179232, "cls_loss": 0.0810546875, "epoch": 0.5861861861861862, "mask_bce_loss": 0.4867558479309082, "mask_dice_loss": 0.06428436934947968, "mask_loss": 0.5510402321815491, "step": 976 }, { "epoch": 0.5867867867867868, "grad_norm": 37.00030517578125, "learning_rate": 7.715228742937319e-07, "loss": 0.7466528415679932, "step": 977 }, { "ce_loss": 8.833234460325912e-05, "cls_loss": 0.059326171875, "epoch": 0.5867867867867868, "mask_bce_loss": 0.4801637828350067, "mask_dice_loss": 0.10162989050149918, "mask_loss": 0.5817936658859253, "step": 977 }, { "epoch": 0.5873873873873874, "grad_norm": 35.13600540161133, "learning_rate": 7.696295026495576e-07, "loss": 0.8784116506576538, "step": 978 }, { "ce_loss": 0.0005526014138013124, "cls_loss": 0.043212890625, "epoch": 0.5873873873873874, "mask_bce_loss": 0.7361059784889221, "mask_dice_loss": 0.1082194447517395, "mask_loss": 0.8443254232406616, "step": 978 }, { "epoch": 0.587987987987988, "grad_norm": 34.93531799316406, "learning_rate": 7.677370027342249e-07, "loss": 0.9523354768753052, "step": 979 }, { "ce_loss": 0.00037781978608109057, "cls_loss": 0.037109375, "epoch": 0.587987987987988, "mask_bce_loss": 0.5504356622695923, "mask_dice_loss": 0.032759279012680054, "mask_loss": 0.5831949710845947, "step": 979 }, { "epoch": 0.5885885885885885, "grad_norm": 40.14404296875, "learning_rate": 7.658453817090094e-07, "loss": 0.8186146020889282, "step": 980 }, { "ce_loss": 5.01655122207012e-05, "cls_loss": 0.0322265625, "epoch": 0.5885885885885885, "mask_bce_loss": 0.3437090814113617, "mask_dice_loss": 0.06154434755444527, "mask_loss": 0.40525344014167786, "step": 980 }, { "epoch": 0.5891891891891892, "grad_norm": 23.767913818359375, "learning_rate": 7.639546467318628e-07, "loss": 0.8788177371025085, "step": 981 }, { "ce_loss": 0.00025100528728216887, "cls_loss": 0.05615234375, "epoch": 0.5891891891891892, "mask_bce_loss": 0.817354679107666, "mask_dice_loss": 0.07777359336614609, "mask_loss": 0.8951282501220703, "step": 981 }, { "epoch": 0.5897897897897898, "grad_norm": 27.474609375, "learning_rate": 7.620648049573814e-07, "loss": 0.8766063451766968, "step": 982 }, { "ce_loss": 3.087173536187038e-05, "cls_loss": 0.032958984375, "epoch": 0.5897897897897898, "mask_bce_loss": 0.4510825276374817, "mask_dice_loss": 0.02520136907696724, "mask_loss": 0.4762839078903198, "step": 982 }, { "epoch": 0.5903903903903904, "grad_norm": 32.12261962890625, "learning_rate": 7.601758635367834e-07, "loss": 0.899978518486023, "step": 983 }, { "ce_loss": 6.387945904862136e-05, "cls_loss": 0.064453125, "epoch": 0.5903903903903904, "mask_bce_loss": 1.222631812095642, "mask_dice_loss": 0.11229514330625534, "mask_loss": 1.334926962852478, "step": 983 }, { "epoch": 0.590990990990991, "grad_norm": 20.572877883911133, "learning_rate": 7.582878296178797e-07, "loss": 0.7606647610664368, "step": 984 }, { "ce_loss": 6.483087781816721e-05, "cls_loss": 0.051513671875, "epoch": 0.590990990990991, "mask_bce_loss": 0.4272204339504242, "mask_dice_loss": 0.09987234324216843, "mask_loss": 0.5270927548408508, "step": 984 }, { "epoch": 0.5915915915915916, "grad_norm": 26.603914260864258, "learning_rate": 7.564007103450471e-07, "loss": 0.8237791061401367, "step": 985 }, { "ce_loss": 5.3232255595503375e-05, "cls_loss": 0.0546875, "epoch": 0.5915915915915916, "mask_bce_loss": 0.4483307898044586, "mask_dice_loss": 0.07371736317873001, "mask_loss": 0.5220481753349304, "step": 985 }, { "epoch": 0.5921921921921922, "grad_norm": 30.798107147216797, "learning_rate": 7.545145128592008e-07, "loss": 0.7796255350112915, "step": 986 }, { "ce_loss": 3.5688419302459806e-05, "cls_loss": 0.041259765625, "epoch": 0.5921921921921922, "mask_bce_loss": 0.5363925099372864, "mask_dice_loss": 0.03920598328113556, "mask_loss": 0.5755984783172607, "step": 986 }, { "epoch": 0.5927927927927928, "grad_norm": 24.74870491027832, "learning_rate": 7.526292442977691e-07, "loss": 0.8602657318115234, "step": 987 }, { "ce_loss": 0.00014061563706491143, "cls_loss": 0.0654296875, "epoch": 0.5927927927927928, "mask_bce_loss": 0.7398970723152161, "mask_dice_loss": 0.09646203368902206, "mask_loss": 0.8363590836524963, "step": 987 }, { "epoch": 0.5933933933933934, "grad_norm": 82.79418182373047, "learning_rate": 7.50744911794664e-07, "loss": 0.8434298038482666, "step": 988 }, { "ce_loss": 0.0010166196152567863, "cls_loss": 0.045166015625, "epoch": 0.5933933933933934, "mask_bce_loss": 0.9613193869590759, "mask_dice_loss": 0.07794246077537537, "mask_loss": 1.039261817932129, "step": 988 }, { "epoch": 0.593993993993994, "grad_norm": 22.534461975097656, "learning_rate": 7.488615224802559e-07, "loss": 0.7350594997406006, "step": 989 }, { "ce_loss": 3.300439493614249e-05, "cls_loss": 0.037353515625, "epoch": 0.593993993993994, "mask_bce_loss": 0.38472601771354675, "mask_dice_loss": 0.03669318929314613, "mask_loss": 0.4214192032814026, "step": 989 }, { "epoch": 0.5945945945945946, "grad_norm": 47.09672546386719, "learning_rate": 7.469790834813467e-07, "loss": 0.8581554889678955, "step": 990 }, { "ce_loss": 6.311700417427346e-05, "cls_loss": 0.059814453125, "epoch": 0.5945945945945946, "mask_bce_loss": 0.6851657032966614, "mask_dice_loss": 0.09018193930387497, "mask_loss": 0.7753476500511169, "step": 990 }, { "epoch": 0.5951951951951951, "grad_norm": 29.20288848876953, "learning_rate": 7.450976019211415e-07, "loss": 0.8577831983566284, "step": 991 }, { "ce_loss": 0.0025078982580453157, "cls_loss": 0.0498046875, "epoch": 0.5951951951951951, "mask_bce_loss": 0.7267052531242371, "mask_dice_loss": 0.09609701484441757, "mask_loss": 0.8228022456169128, "step": 991 }, { "epoch": 0.5957957957957958, "grad_norm": 484.3070373535156, "learning_rate": 7.432170849192224e-07, "loss": 0.9776645302772522, "step": 992 }, { "ce_loss": 2.8236647267476656e-05, "cls_loss": 0.04541015625, "epoch": 0.5957957957957958, "mask_bce_loss": 0.29071342945098877, "mask_dice_loss": 0.04674756899476051, "mask_loss": 0.337460994720459, "step": 992 }, { "epoch": 0.5963963963963964, "grad_norm": 28.363784790039062, "learning_rate": 7.413375395915225e-07, "loss": 0.884605348110199, "step": 993 }, { "ce_loss": 0.0005925294826738536, "cls_loss": 0.068359375, "epoch": 0.5963963963963964, "mask_bce_loss": 0.16605623066425323, "mask_dice_loss": 0.09564101696014404, "mask_loss": 0.2616972327232361, "step": 993 }, { "epoch": 0.596996996996997, "grad_norm": 20.1596622467041, "learning_rate": 7.394589730502971e-07, "loss": 0.8597900867462158, "step": 994 }, { "ce_loss": 3.82114558306057e-05, "cls_loss": 0.04052734375, "epoch": 0.596996996996997, "mask_bce_loss": 0.575151801109314, "mask_dice_loss": 0.03864067792892456, "mask_loss": 0.6137924790382385, "step": 994 }, { "epoch": 0.5975975975975976, "grad_norm": 23.602676391601562, "learning_rate": 7.375813924040981e-07, "loss": 0.8546030521392822, "step": 995 }, { "ce_loss": 0.0012964168563485146, "cls_loss": 0.041015625, "epoch": 0.5975975975975976, "mask_bce_loss": 0.4343665540218353, "mask_dice_loss": 0.03697875887155533, "mask_loss": 0.47134530544281006, "step": 995 }, { "epoch": 0.5981981981981982, "grad_norm": 33.07323455810547, "learning_rate": 7.35704804757747e-07, "loss": 0.8303126692771912, "step": 996 }, { "ce_loss": 0.000832044577691704, "cls_loss": 0.03466796875, "epoch": 0.5981981981981982, "mask_bce_loss": 0.7413159608840942, "mask_dice_loss": 0.0627821609377861, "mask_loss": 0.8040981292724609, "step": 996 }, { "epoch": 0.5987987987987988, "grad_norm": 37.66202926635742, "learning_rate": 7.338292172123073e-07, "loss": 0.9367097616195679, "step": 997 }, { "ce_loss": 1.7320693586952984e-05, "cls_loss": 0.052001953125, "epoch": 0.5987987987987988, "mask_bce_loss": 0.5613507032394409, "mask_dice_loss": 0.10053427517414093, "mask_loss": 0.6618849635124207, "step": 997 }, { "epoch": 0.5993993993993993, "grad_norm": 19.79988670349121, "learning_rate": 7.319546368650579e-07, "loss": 0.8803048133850098, "step": 998 }, { "ce_loss": 5.537162724067457e-05, "cls_loss": 0.05078125, "epoch": 0.5993993993993993, "mask_bce_loss": 0.6814755797386169, "mask_dice_loss": 0.06439163535833359, "mask_loss": 0.7458671927452087, "step": 998 }, { "epoch": 0.6, "grad_norm": 38.546226501464844, "learning_rate": 7.300810708094676e-07, "loss": 0.9088278412818909, "step": 999 }, { "ce_loss": 2.8167038180981763e-05, "cls_loss": 0.0654296875, "epoch": 0.6, "mask_bce_loss": 0.42807555198669434, "mask_dice_loss": 0.03905550390481949, "mask_loss": 0.4671310484409332, "step": 999 }, { "epoch": 0.6006006006006006, "grad_norm": 19.6030330657959, "learning_rate": 7.282085261351658e-07, "loss": 0.6605385541915894, "step": 1000 }, { "ce_loss": 4.649246693588793e-05, "cls_loss": 0.035400390625, "epoch": 0.6006006006006006, "mask_bce_loss": 0.7974574565887451, "mask_dice_loss": 0.0613284707069397, "mask_loss": 0.8587859272956848, "step": 1000 }, { "epoch": 0.6012012012012012, "grad_norm": 48.44000244140625, "learning_rate": 7.263370099279171e-07, "loss": 0.82895827293396, "step": 1001 }, { "ce_loss": 9.588969260221347e-05, "cls_loss": 0.09033203125, "epoch": 0.6012012012012012, "mask_bce_loss": 0.5142056941986084, "mask_dice_loss": 0.1078479066491127, "mask_loss": 0.6220536231994629, "step": 1001 }, { "epoch": 0.6018018018018018, "grad_norm": 34.79450988769531, "learning_rate": 7.244665292695955e-07, "loss": 0.7954471111297607, "step": 1002 }, { "ce_loss": 8.739981421967968e-05, "cls_loss": 0.055419921875, "epoch": 0.6018018018018018, "mask_bce_loss": 0.6684242486953735, "mask_dice_loss": 0.06783495098352432, "mask_loss": 0.7362592220306396, "step": 1002 }, { "epoch": 0.6024024024024024, "grad_norm": 33.90718078613281, "learning_rate": 7.225970912381556e-07, "loss": 0.8728381395339966, "step": 1003 }, { "ce_loss": 3.292554902145639e-05, "cls_loss": 0.059814453125, "epoch": 0.6024024024024024, "mask_bce_loss": 0.8603386878967285, "mask_dice_loss": 0.09383025765419006, "mask_loss": 0.9541689157485962, "step": 1003 }, { "epoch": 0.603003003003003, "grad_norm": 64.38113403320312, "learning_rate": 7.207287029076062e-07, "loss": 0.747726321220398, "step": 1004 }, { "ce_loss": 0.000549064832739532, "cls_loss": 0.0306396484375, "epoch": 0.603003003003003, "mask_bce_loss": 0.1795639544725418, "mask_dice_loss": 0.023014718666672707, "mask_loss": 0.20257867872714996, "step": 1004 }, { "epoch": 0.6036036036036037, "grad_norm": 20.184526443481445, "learning_rate": 7.188613713479847e-07, "loss": 0.7592847943305969, "step": 1005 }, { "ce_loss": 6.781297270208597e-05, "cls_loss": 0.06103515625, "epoch": 0.6036036036036037, "mask_bce_loss": 1.4210492372512817, "mask_dice_loss": 0.1399022489786148, "mask_loss": 1.5609514713287354, "step": 1005 }, { "epoch": 0.6042042042042042, "grad_norm": 41.62115478515625, "learning_rate": 7.169951036253295e-07, "loss": 0.776650607585907, "step": 1006 }, { "ce_loss": 0.00011583978630369529, "cls_loss": 0.052001953125, "epoch": 0.6042042042042042, "mask_bce_loss": 1.0210611820220947, "mask_dice_loss": 0.12094715982675552, "mask_loss": 1.1420083045959473, "step": 1006 }, { "epoch": 0.6048048048048048, "grad_norm": 28.603376388549805, "learning_rate": 7.151299068016541e-07, "loss": 0.8512059450149536, "step": 1007 }, { "ce_loss": 4.1844643419608474e-05, "cls_loss": 0.060546875, "epoch": 0.6048048048048048, "mask_bce_loss": 0.7042644619941711, "mask_dice_loss": 0.08551639318466187, "mask_loss": 0.789780855178833, "step": 1007 }, { "epoch": 0.6054054054054054, "grad_norm": 204.0592498779297, "learning_rate": 7.13265787934918e-07, "loss": 0.8706986308097839, "step": 1008 }, { "ce_loss": 5.4825079132569954e-05, "cls_loss": 0.055419921875, "epoch": 0.6054054054054054, "mask_bce_loss": 0.5258875489234924, "mask_dice_loss": 0.10657300800085068, "mask_loss": 0.6324605345726013, "step": 1008 }, { "epoch": 0.606006006006006, "grad_norm": 126.43660736083984, "learning_rate": 7.114027540790029e-07, "loss": 0.8933515548706055, "step": 1009 }, { "ce_loss": 6.72116075293161e-05, "cls_loss": 0.061279296875, "epoch": 0.606006006006006, "mask_bce_loss": 2.0919175148010254, "mask_dice_loss": 0.17393051087856293, "mask_loss": 2.26584792137146, "step": 1009 }, { "epoch": 0.6066066066066066, "grad_norm": 41.13108444213867, "learning_rate": 7.095408122836847e-07, "loss": 0.8983743190765381, "step": 1010 }, { "ce_loss": 0.00010707805631682277, "cls_loss": 0.0289306640625, "epoch": 0.6066066066066066, "mask_bce_loss": 1.0176504850387573, "mask_dice_loss": 0.04810784012079239, "mask_loss": 1.0657583475112915, "step": 1010 }, { "epoch": 0.6072072072072072, "grad_norm": 27.977603912353516, "learning_rate": 7.076799695946071e-07, "loss": 0.8330540657043457, "step": 1011 }, { "ce_loss": 3.98051633965224e-05, "cls_loss": 0.038818359375, "epoch": 0.6072072072072072, "mask_bce_loss": 0.27469292283058167, "mask_dice_loss": 0.03406468406319618, "mask_loss": 0.30875760316848755, "step": 1011 }, { "epoch": 0.6078078078078079, "grad_norm": 34.4753303527832, "learning_rate": 7.058202330532537e-07, "loss": 0.7002869844436646, "step": 1012 }, { "ce_loss": 0.00013926465180702507, "cls_loss": 0.0306396484375, "epoch": 0.6078078078078079, "mask_bce_loss": 0.5206958651542664, "mask_dice_loss": 0.03892485424876213, "mask_loss": 0.55962073802948, "step": 1012 }, { "epoch": 0.6084084084084084, "grad_norm": 63.304664611816406, "learning_rate": 7.039616096969236e-07, "loss": 0.9886430501937866, "step": 1013 }, { "ce_loss": 0.00010046800161944702, "cls_loss": 0.036376953125, "epoch": 0.6084084084084084, "mask_bce_loss": 0.2654756009578705, "mask_dice_loss": 0.029823405668139458, "mask_loss": 0.2952989935874939, "step": 1013 }, { "epoch": 0.609009009009009, "grad_norm": 42.19860076904297, "learning_rate": 7.02104106558703e-07, "loss": 0.8315311670303345, "step": 1014 }, { "ce_loss": 6.687081622658297e-05, "cls_loss": 0.055419921875, "epoch": 0.609009009009009, "mask_bce_loss": 0.9164747595787048, "mask_dice_loss": 0.13784104585647583, "mask_loss": 1.0543158054351807, "step": 1014 }, { "epoch": 0.6096096096096096, "grad_norm": 26.314847946166992, "learning_rate": 7.002477306674388e-07, "loss": 0.8817092180252075, "step": 1015 }, { "ce_loss": 4.8515572416363284e-05, "cls_loss": 0.03369140625, "epoch": 0.6096096096096096, "mask_bce_loss": 0.25817933678627014, "mask_dice_loss": 0.026481812819838524, "mask_loss": 0.2846611440181732, "step": 1015 }, { "epoch": 0.6102102102102102, "grad_norm": 70.8653564453125, "learning_rate": 6.983924890477137e-07, "loss": 0.8032901287078857, "step": 1016 }, { "ce_loss": 3.1599236535839736e-05, "cls_loss": 0.06298828125, "epoch": 0.6102102102102102, "mask_bce_loss": 0.4783388674259186, "mask_dice_loss": 0.09867052733898163, "mask_loss": 0.577009379863739, "step": 1016 }, { "epoch": 0.6108108108108108, "grad_norm": 17.51882553100586, "learning_rate": 6.965383887198169e-07, "loss": 0.7244220972061157, "step": 1017 }, { "ce_loss": 4.609841562341899e-05, "cls_loss": 0.056640625, "epoch": 0.6108108108108108, "mask_bce_loss": 0.39499443769454956, "mask_dice_loss": 0.07331410050392151, "mask_loss": 0.46830853819847107, "step": 1017 }, { "epoch": 0.6114114114114114, "grad_norm": 22.776697158813477, "learning_rate": 6.946854366997187e-07, "loss": 0.8102864027023315, "step": 1018 }, { "ce_loss": 0.001049302751198411, "cls_loss": 0.03173828125, "epoch": 0.6114114114114114, "mask_bce_loss": 0.3167661130428314, "mask_dice_loss": 0.023283040151000023, "mask_loss": 0.340049147605896, "step": 1018 }, { "epoch": 0.612012012012012, "grad_norm": 18.60266876220703, "learning_rate": 6.92833639999046e-07, "loss": 0.7684731483459473, "step": 1019 }, { "ce_loss": 0.0006661581574007869, "cls_loss": 0.05029296875, "epoch": 0.612012012012012, "mask_bce_loss": 0.594716489315033, "mask_dice_loss": 0.06256385147571564, "mask_loss": 0.6572803258895874, "step": 1019 }, { "epoch": 0.6126126126126126, "grad_norm": 36.51486587524414, "learning_rate": 6.909830056250526e-07, "loss": 0.9027771353721619, "step": 1020 }, { "ce_loss": 8.690387767273933e-05, "cls_loss": 0.044921875, "epoch": 0.6126126126126126, "mask_bce_loss": 0.551805317401886, "mask_dice_loss": 0.06737890839576721, "mask_loss": 0.6191842555999756, "step": 1020 }, { "epoch": 0.6132132132132132, "grad_norm": 19.374292373657227, "learning_rate": 6.891335405805937e-07, "loss": 0.9827214479446411, "step": 1021 }, { "ce_loss": 7.571605237899348e-05, "cls_loss": 0.06787109375, "epoch": 0.6132132132132132, "mask_bce_loss": 1.3572287559509277, "mask_dice_loss": 0.11559569090604782, "mask_loss": 1.4728244543075562, "step": 1021 }, { "epoch": 0.6138138138138138, "grad_norm": 29.849361419677734, "learning_rate": 6.872852518641008e-07, "loss": 0.8806887865066528, "step": 1022 }, { "ce_loss": 2.2733889636583626e-05, "cls_loss": 0.060546875, "epoch": 0.6138138138138138, "mask_bce_loss": 0.20644818246364594, "mask_dice_loss": 0.07294868677854538, "mask_loss": 0.2793968617916107, "step": 1022 }, { "epoch": 0.6144144144144145, "grad_norm": 28.752275466918945, "learning_rate": 6.854381464695534e-07, "loss": 0.9401310682296753, "step": 1023 }, { "ce_loss": 3.862734592985362e-05, "cls_loss": 0.06591796875, "epoch": 0.6144144144144145, "mask_bce_loss": 0.5502747893333435, "mask_dice_loss": 0.129934161901474, "mask_loss": 0.6802089214324951, "step": 1023 }, { "epoch": 0.615015015015015, "grad_norm": 22.31058120727539, "learning_rate": 6.835922313864531e-07, "loss": 0.8755064010620117, "step": 1024 }, { "ce_loss": 0.0004153322661295533, "cls_loss": 0.055419921875, "epoch": 0.615015015015015, "mask_bce_loss": 1.244409203529358, "mask_dice_loss": 0.13811276853084564, "mask_loss": 1.3825219869613647, "step": 1024 }, { "epoch": 0.6156156156156156, "grad_norm": 70.35555267333984, "learning_rate": 6.817475135997986e-07, "loss": 0.8169569969177246, "step": 1025 }, { "ce_loss": 0.0009257858619093895, "cls_loss": 0.048828125, "epoch": 0.6156156156156156, "mask_bce_loss": 0.2882842421531677, "mask_dice_loss": 0.06006394699215889, "mask_loss": 0.3483482003211975, "step": 1025 }, { "epoch": 0.6162162162162163, "grad_norm": 29.984432220458984, "learning_rate": 6.799040000900562e-07, "loss": 0.9397627711296082, "step": 1026 }, { "ce_loss": 0.17032501101493835, "cls_loss": 0.0732421875, "epoch": 0.6162162162162163, "mask_bce_loss": 1.8973703384399414, "mask_dice_loss": 0.0725208967924118, "mask_loss": 1.9698911905288696, "step": 1026 }, { "epoch": 0.6168168168168168, "grad_norm": 47.42194747924805, "learning_rate": 6.780616978331362e-07, "loss": 0.8985476493835449, "step": 1027 }, { "ce_loss": 6.288850272540003e-05, "cls_loss": 0.03466796875, "epoch": 0.6168168168168168, "mask_bce_loss": 0.3989337682723999, "mask_dice_loss": 0.026464378461241722, "mask_loss": 0.4253981411457062, "step": 1027 }, { "epoch": 0.6174174174174174, "grad_norm": 19.294458389282227, "learning_rate": 6.762206138003659e-07, "loss": 0.7259501218795776, "step": 1028 }, { "ce_loss": 4.952650488121435e-05, "cls_loss": 0.056640625, "epoch": 0.6174174174174174, "mask_bce_loss": 0.6027783751487732, "mask_dice_loss": 0.07851745933294296, "mask_loss": 0.6812958121299744, "step": 1028 }, { "epoch": 0.618018018018018, "grad_norm": 13.7121000289917, "learning_rate": 6.74380754958462e-07, "loss": 0.7526868581771851, "step": 1029 }, { "ce_loss": 7.734022074146196e-05, "cls_loss": 0.1103515625, "epoch": 0.618018018018018, "mask_bce_loss": 1.3735946416854858, "mask_dice_loss": 0.09725994616746902, "mask_loss": 1.470854640007019, "step": 1029 }, { "epoch": 0.6186186186186187, "grad_norm": 21.251026153564453, "learning_rate": 6.725421282695048e-07, "loss": 0.7937528491020203, "step": 1030 }, { "ce_loss": 2.311524258402642e-05, "cls_loss": 0.0228271484375, "epoch": 0.6186186186186187, "mask_bce_loss": 0.27812954783439636, "mask_dice_loss": 0.01685536652803421, "mask_loss": 0.29498490691185, "step": 1030 }, { "epoch": 0.6192192192192192, "grad_norm": 17.78633689880371, "learning_rate": 6.707047406909135e-07, "loss": 0.5841040015220642, "step": 1031 }, { "ce_loss": 4.224963777232915e-05, "cls_loss": 0.053955078125, "epoch": 0.6192192192192192, "mask_bce_loss": 1.24183988571167, "mask_dice_loss": 0.07629027962684631, "mask_loss": 1.3181301355361938, "step": 1031 }, { "epoch": 0.6198198198198198, "grad_norm": 14.197965621948242, "learning_rate": 6.688685991754168e-07, "loss": 0.8163205981254578, "step": 1032 }, { "ce_loss": 2.8135886168456636e-05, "cls_loss": 0.059814453125, "epoch": 0.6198198198198198, "mask_bce_loss": 0.9791017770767212, "mask_dice_loss": 0.07942155003547668, "mask_loss": 1.0585232973098755, "step": 1032 }, { "epoch": 0.6204204204204204, "grad_norm": 21.357574462890625, "learning_rate": 6.670337106710293e-07, "loss": 0.8132756948471069, "step": 1033 }, { "ce_loss": 6.523379852296785e-05, "cls_loss": 0.04541015625, "epoch": 0.6204204204204204, "mask_bce_loss": 0.49123525619506836, "mask_dice_loss": 0.10918116569519043, "mask_loss": 0.6004164218902588, "step": 1033 }, { "epoch": 0.621021021021021, "grad_norm": 17.71416664123535, "learning_rate": 6.652000821210242e-07, "loss": 0.7226579189300537, "step": 1034 }, { "ce_loss": 2.0241684978827834e-05, "cls_loss": 0.041015625, "epoch": 0.621021021021021, "mask_bce_loss": 0.9586052894592285, "mask_dice_loss": 0.05265306308865547, "mask_loss": 1.0112583637237549, "step": 1034 }, { "epoch": 0.6216216216216216, "grad_norm": 30.923656463623047, "learning_rate": 6.633677204639064e-07, "loss": 0.8652178049087524, "step": 1035 }, { "ce_loss": 7.851242844481021e-05, "cls_loss": 0.055419921875, "epoch": 0.6216216216216216, "mask_bce_loss": 0.7742827534675598, "mask_dice_loss": 0.10355972498655319, "mask_loss": 0.8778424859046936, "step": 1035 }, { "epoch": 0.6222222222222222, "grad_norm": 36.6099853515625, "learning_rate": 6.615366326333866e-07, "loss": 0.8945047855377197, "step": 1036 }, { "ce_loss": 7.162817928474396e-05, "cls_loss": 0.03857421875, "epoch": 0.6222222222222222, "mask_bce_loss": 0.5627283453941345, "mask_dice_loss": 0.06702488660812378, "mask_loss": 0.6297532320022583, "step": 1036 }, { "epoch": 0.6228228228228229, "grad_norm": 28.471887588500977, "learning_rate": 6.597068255583569e-07, "loss": 0.9300252199172974, "step": 1037 }, { "ce_loss": 5.128289194544777e-05, "cls_loss": 0.0673828125, "epoch": 0.6228228228228229, "mask_bce_loss": 0.5713179707527161, "mask_dice_loss": 0.12710636854171753, "mask_loss": 0.6984243392944336, "step": 1037 }, { "epoch": 0.6234234234234234, "grad_norm": 22.185916900634766, "learning_rate": 6.578783061628614e-07, "loss": 0.8469698429107666, "step": 1038 }, { "ce_loss": 2.824248258548323e-05, "cls_loss": 0.030029296875, "epoch": 0.6234234234234234, "mask_bce_loss": 0.6599313616752625, "mask_dice_loss": 0.02941202186048031, "mask_loss": 0.6893433928489685, "step": 1038 }, { "epoch": 0.624024024024024, "grad_norm": 39.032203674316406, "learning_rate": 6.560510813660718e-07, "loss": 0.8938833475112915, "step": 1039 }, { "ce_loss": 0.00023714064445812255, "cls_loss": 0.05517578125, "epoch": 0.624024024024024, "mask_bce_loss": 1.1327834129333496, "mask_dice_loss": 0.08760466426610947, "mask_loss": 1.2203880548477173, "step": 1039 }, { "epoch": 0.6246246246246246, "grad_norm": 26.3149471282959, "learning_rate": 6.542251580822617e-07, "loss": 0.815117359161377, "step": 1040 }, { "ce_loss": 2.7377469450584613e-05, "cls_loss": 0.061767578125, "epoch": 0.6246246246246246, "mask_bce_loss": 0.06354989856481552, "mask_dice_loss": 0.057533182203769684, "mask_loss": 0.1210830807685852, "step": 1040 }, { "epoch": 0.6252252252252253, "grad_norm": 45.254390716552734, "learning_rate": 6.524005432207794e-07, "loss": 1.028184175491333, "step": 1041 }, { "ce_loss": 9.496764687355608e-05, "cls_loss": 0.05029296875, "epoch": 0.6252252252252253, "mask_bce_loss": 0.40992480516433716, "mask_dice_loss": 0.05482357740402222, "mask_loss": 0.4647483825683594, "step": 1041 }, { "epoch": 0.6258258258258258, "grad_norm": 21.027204513549805, "learning_rate": 6.505772436860215e-07, "loss": 0.7330482602119446, "step": 1042 }, { "ce_loss": 2.8419732188922353e-05, "cls_loss": 0.07958984375, "epoch": 0.6258258258258258, "mask_bce_loss": 0.6839763522148132, "mask_dice_loss": 0.08535724878311157, "mask_loss": 0.7693336009979248, "step": 1042 }, { "epoch": 0.6264264264264264, "grad_norm": 66.97217559814453, "learning_rate": 6.487552663774085e-07, "loss": 0.7421690225601196, "step": 1043 }, { "ce_loss": 0.0012128535890951753, "cls_loss": 0.053466796875, "epoch": 0.6264264264264264, "mask_bce_loss": 2.023660898208618, "mask_dice_loss": 0.09065796434879303, "mask_loss": 2.11431884765625, "step": 1043 }, { "epoch": 0.6270270270270271, "grad_norm": 27.431203842163086, "learning_rate": 6.469346181893567e-07, "loss": 0.9139910936355591, "step": 1044 }, { "ce_loss": 2.9337286832742393e-05, "cls_loss": 0.05810546875, "epoch": 0.6270270270270271, "mask_bce_loss": 1.1285533905029297, "mask_dice_loss": 0.10644326359033585, "mask_loss": 1.2349966764450073, "step": 1044 }, { "epoch": 0.6276276276276276, "grad_norm": 24.075180053710938, "learning_rate": 6.451153060112528e-07, "loss": 0.8011153936386108, "step": 1045 }, { "ce_loss": 4.1526775021338835e-05, "cls_loss": 0.05517578125, "epoch": 0.6276276276276276, "mask_bce_loss": 0.8181840777397156, "mask_dice_loss": 0.1255553662776947, "mask_loss": 0.9437394142150879, "step": 1045 }, { "epoch": 0.6282282282282282, "grad_norm": 32.30304718017578, "learning_rate": 6.432973367274292e-07, "loss": 0.9081242680549622, "step": 1046 }, { "ce_loss": 0.0066858502104878426, "cls_loss": 0.0235595703125, "epoch": 0.6282282282282282, "mask_bce_loss": 0.24773335456848145, "mask_dice_loss": 0.01801338605582714, "mask_loss": 0.26574674248695374, "step": 1046 }, { "epoch": 0.6288288288288288, "grad_norm": 24.44843864440918, "learning_rate": 6.414807172171355e-07, "loss": 0.8584402799606323, "step": 1047 }, { "ce_loss": 0.00010613544145599008, "cls_loss": 0.047119140625, "epoch": 0.6288288288288288, "mask_bce_loss": 0.8694931864738464, "mask_dice_loss": 0.0873073935508728, "mask_loss": 0.9568005800247192, "step": 1047 }, { "epoch": 0.6294294294294295, "grad_norm": 27.851354598999023, "learning_rate": 6.396654543545139e-07, "loss": 0.8253355026245117, "step": 1048 }, { "ce_loss": 0.00048803735990077257, "cls_loss": 0.05078125, "epoch": 0.6294294294294295, "mask_bce_loss": 0.45212918519973755, "mask_dice_loss": 0.056921977549791336, "mask_loss": 0.5090511441230774, "step": 1048 }, { "epoch": 0.63003003003003, "grad_norm": 18.17875862121582, "learning_rate": 6.37851555008574e-07, "loss": 0.7513272762298584, "step": 1049 }, { "ce_loss": 3.90061795769725e-05, "cls_loss": 0.055908203125, "epoch": 0.63003003003003, "mask_bce_loss": 0.3800128698348999, "mask_dice_loss": 0.08914899826049805, "mask_loss": 0.46916186809539795, "step": 1049 }, { "epoch": 0.6306306306306306, "grad_norm": 31.673154830932617, "learning_rate": 6.360390260431647e-07, "loss": 0.8597162365913391, "step": 1050 }, { "ce_loss": 5.926414087298326e-05, "cls_loss": 0.057861328125, "epoch": 0.6306306306306306, "mask_bce_loss": 0.886784017086029, "mask_dice_loss": 0.11656459420919418, "mask_loss": 1.0033485889434814, "step": 1050 }, { "epoch": 0.6312312312312313, "grad_norm": 23.782445907592773, "learning_rate": 6.342278743169495e-07, "loss": 0.933528482913971, "step": 1051 }, { "ce_loss": 0.00011996431567240506, "cls_loss": 0.03271484375, "epoch": 0.6312312312312313, "mask_bce_loss": 0.24118033051490784, "mask_dice_loss": 0.02383543737232685, "mask_loss": 0.26501578092575073, "step": 1051 }, { "epoch": 0.6318318318318318, "grad_norm": 18.276885986328125, "learning_rate": 6.32418106683381e-07, "loss": 0.7194221615791321, "step": 1052 }, { "ce_loss": 9.51084730331786e-05, "cls_loss": 0.04541015625, "epoch": 0.6318318318318318, "mask_bce_loss": 0.7507302165031433, "mask_dice_loss": 0.05697400122880936, "mask_loss": 0.8077042102813721, "step": 1052 }, { "epoch": 0.6324324324324324, "grad_norm": 24.13121795654297, "learning_rate": 6.306097299906738e-07, "loss": 0.8211721777915955, "step": 1053 }, { "ce_loss": 7.016334711806849e-05, "cls_loss": 0.048828125, "epoch": 0.6324324324324324, "mask_bce_loss": 0.9528465270996094, "mask_dice_loss": 0.09048324078321457, "mask_loss": 1.0433297157287598, "step": 1053 }, { "epoch": 0.633033033033033, "grad_norm": 43.962791442871094, "learning_rate": 6.288027510817791e-07, "loss": 0.9508538246154785, "step": 1054 }, { "ce_loss": 0.0002025692374445498, "cls_loss": 0.0673828125, "epoch": 0.633033033033033, "mask_bce_loss": 2.6378161907196045, "mask_dice_loss": 0.11203797161579132, "mask_loss": 2.74985408782959, "step": 1054 }, { "epoch": 0.6336336336336337, "grad_norm": 25.13448143005371, "learning_rate": 6.269971767943596e-07, "loss": 0.84116530418396, "step": 1055 }, { "ce_loss": 0.0034291213378310204, "cls_loss": 0.05419921875, "epoch": 0.6336336336336337, "mask_bce_loss": 1.3389701843261719, "mask_dice_loss": 0.12159699201583862, "mask_loss": 1.4605672359466553, "step": 1055 }, { "epoch": 0.6342342342342342, "grad_norm": 29.762014389038086, "learning_rate": 6.251930139607621e-07, "loss": 0.871665358543396, "step": 1056 }, { "ce_loss": 5.4805164836579934e-05, "cls_loss": 0.051513671875, "epoch": 0.6342342342342342, "mask_bce_loss": 0.7171808481216431, "mask_dice_loss": 0.06912501901388168, "mask_loss": 0.786305844783783, "step": 1056 }, { "epoch": 0.6348348348348348, "grad_norm": 22.8749942779541, "learning_rate": 6.233902694079923e-07, "loss": 0.84441077709198, "step": 1057 }, { "ce_loss": 8.529572369297966e-05, "cls_loss": 0.05517578125, "epoch": 0.6348348348348348, "mask_bce_loss": 0.5614801645278931, "mask_dice_loss": 0.07456635683774948, "mask_loss": 0.6360465288162231, "step": 1057 }, { "epoch": 0.6354354354354355, "grad_norm": 22.804845809936523, "learning_rate": 6.215889499576897e-07, "loss": 0.8942234516143799, "step": 1058 }, { "ce_loss": 3.1267612939700484e-05, "cls_loss": 0.060546875, "epoch": 0.6354354354354355, "mask_bce_loss": 0.8859120607376099, "mask_dice_loss": 0.0889880359172821, "mask_loss": 0.9749001264572144, "step": 1058 }, { "epoch": 0.6360360360360361, "grad_norm": 50.82362365722656, "learning_rate": 6.197890624261011e-07, "loss": 0.7708170413970947, "step": 1059 }, { "ce_loss": 5.0421644118614495e-05, "cls_loss": 0.099609375, "epoch": 0.6360360360360361, "mask_bce_loss": 2.924372434616089, "mask_dice_loss": 0.14044393599033356, "mask_loss": 3.064816474914551, "step": 1059 }, { "epoch": 0.6366366366366366, "grad_norm": 21.24911880493164, "learning_rate": 6.179906136240541e-07, "loss": 0.7702158689498901, "step": 1060 }, { "ce_loss": 3.598104740376584e-05, "cls_loss": 0.04443359375, "epoch": 0.6366366366366366, "mask_bce_loss": 0.916417121887207, "mask_dice_loss": 0.03679628297686577, "mask_loss": 0.9532133936882019, "step": 1060 }, { "epoch": 0.6372372372372372, "grad_norm": 34.4188117980957, "learning_rate": 6.161936103569332e-07, "loss": 0.8898429870605469, "step": 1061 }, { "ce_loss": 0.01915779709815979, "cls_loss": 0.0390625, "epoch": 0.6372372372372372, "mask_bce_loss": 0.4499473571777344, "mask_dice_loss": 0.03691883757710457, "mask_loss": 0.48686620593070984, "step": 1061 }, { "epoch": 0.6378378378378379, "grad_norm": 25.75209617614746, "learning_rate": 6.143980594246523e-07, "loss": 0.6560245752334595, "step": 1062 }, { "ce_loss": 0.000350288231857121, "cls_loss": 0.04541015625, "epoch": 0.6378378378378379, "mask_bce_loss": 0.6562290191650391, "mask_dice_loss": 0.05382717773318291, "mask_loss": 0.7100561857223511, "step": 1062 }, { "epoch": 0.6384384384384384, "grad_norm": 22.41826057434082, "learning_rate": 6.126039676216293e-07, "loss": 0.7122003436088562, "step": 1063 }, { "ce_loss": 7.133725011954084e-05, "cls_loss": 0.07421875, "epoch": 0.6384384384384384, "mask_bce_loss": 0.2541399896144867, "mask_dice_loss": 0.12945567071437836, "mask_loss": 0.38359564542770386, "step": 1063 }, { "epoch": 0.639039039039039, "grad_norm": 21.683610916137695, "learning_rate": 6.10811341736762e-07, "loss": 0.7769410610198975, "step": 1064 }, { "ce_loss": 0.00034672184847295284, "cls_loss": 0.04931640625, "epoch": 0.639039039039039, "mask_bce_loss": 0.7120162844657898, "mask_dice_loss": 0.11372649669647217, "mask_loss": 0.825742781162262, "step": 1064 }, { "epoch": 0.6396396396396397, "grad_norm": 28.215412139892578, "learning_rate": 6.090201885533996e-07, "loss": 0.7387319803237915, "step": 1065 }, { "ce_loss": 2.6041714590974152e-05, "cls_loss": 0.029541015625, "epoch": 0.6396396396396397, "mask_bce_loss": 0.30042192339897156, "mask_dice_loss": 0.021751349791884422, "mask_loss": 0.32217326760292053, "step": 1065 }, { "epoch": 0.6402402402402403, "grad_norm": 32.39112091064453, "learning_rate": 6.072305148493194e-07, "loss": 0.773805558681488, "step": 1066 }, { "ce_loss": 3.1346757168648764e-05, "cls_loss": 0.051513671875, "epoch": 0.6402402402402403, "mask_bce_loss": 0.932569146156311, "mask_dice_loss": 0.1063869521021843, "mask_loss": 1.0389560461044312, "step": 1066 }, { "epoch": 0.6408408408408408, "grad_norm": 33.95850372314453, "learning_rate": 6.054423273967004e-07, "loss": 1.074865698814392, "step": 1067 }, { "ce_loss": 0.00010453880531713367, "cls_loss": 0.05859375, "epoch": 0.6408408408408408, "mask_bce_loss": 0.836202085018158, "mask_dice_loss": 0.09988020360469818, "mask_loss": 0.9360823035240173, "step": 1067 }, { "epoch": 0.6414414414414414, "grad_norm": 23.662797927856445, "learning_rate": 6.036556329620972e-07, "loss": 0.7060127258300781, "step": 1068 }, { "ce_loss": 5.568929191213101e-05, "cls_loss": 0.06640625, "epoch": 0.6414414414414414, "mask_bce_loss": 0.8377974629402161, "mask_dice_loss": 0.09045432507991791, "mask_loss": 0.9282518029212952, "step": 1068 }, { "epoch": 0.6420420420420421, "grad_norm": 25.97455406188965, "learning_rate": 6.018704383064144e-07, "loss": 0.8046417236328125, "step": 1069 }, { "ce_loss": 0.010762551799416542, "cls_loss": 0.06298828125, "epoch": 0.6420420420420421, "mask_bce_loss": 2.233177900314331, "mask_dice_loss": 0.1223655715584755, "mask_loss": 2.355543375015259, "step": 1069 }, { "epoch": 0.6426426426426426, "grad_norm": 53.35335159301758, "learning_rate": 6.000867501848827e-07, "loss": 0.9747555255889893, "step": 1070 }, { "ce_loss": 5.1236780564067885e-05, "cls_loss": 0.038818359375, "epoch": 0.6426426426426426, "mask_bce_loss": 0.22199292480945587, "mask_dice_loss": 0.0353488028049469, "mask_loss": 0.25734174251556396, "step": 1070 }, { "epoch": 0.6432432432432432, "grad_norm": 18.784290313720703, "learning_rate": 5.983045753470307e-07, "loss": 0.7500845193862915, "step": 1071 }, { "ce_loss": 2.8779431886505336e-05, "cls_loss": 0.0673828125, "epoch": 0.6432432432432432, "mask_bce_loss": 0.741344153881073, "mask_dice_loss": 0.12138461321592331, "mask_loss": 0.8627287745475769, "step": 1071 }, { "epoch": 0.6438438438438439, "grad_norm": 24.376197814941406, "learning_rate": 5.96523920536661e-07, "loss": 0.9016768336296082, "step": 1072 }, { "ce_loss": 6.493805994978175e-05, "cls_loss": 0.046142578125, "epoch": 0.6438438438438439, "mask_bce_loss": 1.1891376972198486, "mask_dice_loss": 0.0504097118973732, "mask_loss": 1.2395473718643188, "step": 1072 }, { "epoch": 0.6444444444444445, "grad_norm": 76.04017639160156, "learning_rate": 5.947447924918253e-07, "loss": 0.8303893804550171, "step": 1073 }, { "ce_loss": 0.00020008232968393713, "cls_loss": 0.05224609375, "epoch": 0.6444444444444445, "mask_bce_loss": 0.6596128940582275, "mask_dice_loss": 0.08191845566034317, "mask_loss": 0.7415313720703125, "step": 1073 }, { "epoch": 0.645045045045045, "grad_norm": 18.670862197875977, "learning_rate": 5.929671979447967e-07, "loss": 0.8141762018203735, "step": 1074 }, { "ce_loss": 0.0023608608171343803, "cls_loss": 0.053955078125, "epoch": 0.645045045045045, "mask_bce_loss": 1.848665475845337, "mask_dice_loss": 0.08096040785312653, "mask_loss": 1.9296258687973022, "step": 1074 }, { "epoch": 0.6456456456456456, "grad_norm": 18.53375816345215, "learning_rate": 5.911911436220463e-07, "loss": 0.8085805177688599, "step": 1075 }, { "ce_loss": 0.00038176256930455565, "cls_loss": 0.0712890625, "epoch": 0.6456456456456456, "mask_bce_loss": 0.6277614831924438, "mask_dice_loss": 0.08041471987962723, "mask_loss": 0.7081761956214905, "step": 1075 }, { "epoch": 0.6462462462462463, "grad_norm": 16.944677352905273, "learning_rate": 5.894166362442167e-07, "loss": 0.8309458494186401, "step": 1076 }, { "ce_loss": 5.501518535311334e-05, "cls_loss": 0.02734375, "epoch": 0.6462462462462463, "mask_bce_loss": 0.22279350459575653, "mask_dice_loss": 0.01894446648657322, "mask_loss": 0.2417379766702652, "step": 1076 }, { "epoch": 0.6468468468468469, "grad_norm": 15.466407775878906, "learning_rate": 5.876436825260967e-07, "loss": 0.722930908203125, "step": 1077 }, { "ce_loss": 5.984995368635282e-05, "cls_loss": 0.039794921875, "epoch": 0.6468468468468469, "mask_bce_loss": 0.3138387203216553, "mask_dice_loss": 0.035920556634664536, "mask_loss": 0.3497592806816101, "step": 1077 }, { "epoch": 0.6474474474474474, "grad_norm": 22.647367477416992, "learning_rate": 5.858722891765958e-07, "loss": 0.8065639138221741, "step": 1078 }, { "ce_loss": 0.00012404637527652085, "cls_loss": 0.042236328125, "epoch": 0.6474474474474474, "mask_bce_loss": 0.5756774544715881, "mask_dice_loss": 0.05425538495182991, "mask_loss": 0.6299328207969666, "step": 1078 }, { "epoch": 0.648048048048048, "grad_norm": 28.44256019592285, "learning_rate": 5.841024628987199e-07, "loss": 0.8660650253295898, "step": 1079 }, { "ce_loss": 5.5445751058869064e-05, "cls_loss": 0.05810546875, "epoch": 0.648048048048048, "mask_bce_loss": 0.8602668642997742, "mask_dice_loss": 0.11494338512420654, "mask_loss": 0.9752102494239807, "step": 1079 }, { "epoch": 0.6486486486486487, "grad_norm": 16.976228713989258, "learning_rate": 5.823342103895443e-07, "loss": 0.7457882761955261, "step": 1080 }, { "ce_loss": 0.00023458281066268682, "cls_loss": 0.04541015625, "epoch": 0.6486486486486487, "mask_bce_loss": 0.746669590473175, "mask_dice_loss": 0.06860969960689545, "mask_loss": 0.8152793049812317, "step": 1080 }, { "epoch": 0.6492492492492492, "grad_norm": 16.24930763244629, "learning_rate": 5.80567538340189e-07, "loss": 0.7268856763839722, "step": 1081 }, { "ce_loss": 3.44936124747619e-05, "cls_loss": 0.055908203125, "epoch": 0.6492492492492492, "mask_bce_loss": 1.2674236297607422, "mask_dice_loss": 0.10155313462018967, "mask_loss": 1.3689767122268677, "step": 1081 }, { "epoch": 0.6498498498498498, "grad_norm": 41.07123565673828, "learning_rate": 5.788024534357943e-07, "loss": 0.8392034769058228, "step": 1082 }, { "ce_loss": 3.257519347243942e-05, "cls_loss": 0.043701171875, "epoch": 0.6498498498498498, "mask_bce_loss": 0.5508887767791748, "mask_dice_loss": 0.06649675965309143, "mask_loss": 0.6173855066299438, "step": 1082 }, { "epoch": 0.6504504504504505, "grad_norm": 48.98479461669922, "learning_rate": 5.770389623554938e-07, "loss": 0.8866141438484192, "step": 1083 }, { "ce_loss": 4.009090116596781e-05, "cls_loss": 0.03515625, "epoch": 0.6504504504504505, "mask_bce_loss": 0.2750660181045532, "mask_dice_loss": 0.02865542843937874, "mask_loss": 0.30372145771980286, "step": 1083 }, { "epoch": 0.6510510510510511, "grad_norm": 21.609342575073242, "learning_rate": 5.752770717723902e-07, "loss": 0.7783610820770264, "step": 1084 }, { "ce_loss": 0.0009512259857729077, "cls_loss": 0.047607421875, "epoch": 0.6510510510510511, "mask_bce_loss": 0.6573162078857422, "mask_dice_loss": 0.06432564556598663, "mask_loss": 0.7216418385505676, "step": 1084 }, { "epoch": 0.6516516516516516, "grad_norm": 32.065914154052734, "learning_rate": 5.735167883535313e-07, "loss": 0.8592770099639893, "step": 1085 }, { "ce_loss": 3.922179530491121e-05, "cls_loss": 0.04833984375, "epoch": 0.6516516516516516, "mask_bce_loss": 0.6993817687034607, "mask_dice_loss": 0.06188012287020683, "mask_loss": 0.7612618803977966, "step": 1085 }, { "epoch": 0.6522522522522523, "grad_norm": 29.137187957763672, "learning_rate": 5.717581187598809e-07, "loss": 0.8882476091384888, "step": 1086 }, { "ce_loss": 0.0011824293760582805, "cls_loss": 0.04833984375, "epoch": 0.6522522522522523, "mask_bce_loss": 0.9332088828086853, "mask_dice_loss": 0.07973796874284744, "mask_loss": 1.0129468441009521, "step": 1086 }, { "epoch": 0.6528528528528529, "grad_norm": 30.562856674194336, "learning_rate": 5.700010696462972e-07, "loss": 0.8950607776641846, "step": 1087 }, { "ce_loss": 0.00010707159526646137, "cls_loss": 0.055908203125, "epoch": 0.6528528528528529, "mask_bce_loss": 0.7535130977630615, "mask_dice_loss": 0.07470845431089401, "mask_loss": 0.8282215595245361, "step": 1087 }, { "epoch": 0.6534534534534534, "grad_norm": 20.924434661865234, "learning_rate": 5.682456476615072e-07, "loss": 0.6360899806022644, "step": 1088 }, { "ce_loss": 3.2426996767753735e-05, "cls_loss": 0.05859375, "epoch": 0.6534534534534534, "mask_bce_loss": 0.2531389594078064, "mask_dice_loss": 0.0738477110862732, "mask_loss": 0.3269866704940796, "step": 1088 }, { "epoch": 0.654054054054054, "grad_norm": 17.985883712768555, "learning_rate": 5.664918594480799e-07, "loss": 0.8119602799415588, "step": 1089 }, { "ce_loss": 0.00012855300155933946, "cls_loss": 0.057861328125, "epoch": 0.654054054054054, "mask_bce_loss": 0.6167773008346558, "mask_dice_loss": 0.06285638362169266, "mask_loss": 0.6796336770057678, "step": 1089 }, { "epoch": 0.6546546546546547, "grad_norm": 15.681777954101562, "learning_rate": 5.647397116424018e-07, "loss": 0.7850281596183777, "step": 1090 }, { "ce_loss": 5.214670454734005e-05, "cls_loss": 0.04052734375, "epoch": 0.6546546546546547, "mask_bce_loss": 0.46543702483177185, "mask_dice_loss": 0.05135071277618408, "mask_loss": 0.5167877674102783, "step": 1090 }, { "epoch": 0.6552552552552553, "grad_norm": 33.73982620239258, "learning_rate": 5.629892108746532e-07, "loss": 0.8033097982406616, "step": 1091 }, { "ce_loss": 8.503627759637311e-05, "cls_loss": 0.047607421875, "epoch": 0.6552552552552553, "mask_bce_loss": 0.52936851978302, "mask_dice_loss": 0.046268463134765625, "mask_loss": 0.5756369829177856, "step": 1091 }, { "epoch": 0.6558558558558558, "grad_norm": 32.759124755859375, "learning_rate": 5.612403637687813e-07, "loss": 0.7654881477355957, "step": 1092 }, { "ce_loss": 0.0003638124908320606, "cls_loss": 0.04296875, "epoch": 0.6558558558558558, "mask_bce_loss": 1.3081446886062622, "mask_dice_loss": 0.047911386936903, "mask_loss": 1.3560560941696167, "step": 1092 }, { "epoch": 0.6564564564564564, "grad_norm": 23.733694076538086, "learning_rate": 5.594931769424747e-07, "loss": 0.7833601832389832, "step": 1093 }, { "ce_loss": 0.0008658190490677953, "cls_loss": 0.049072265625, "epoch": 0.6564564564564564, "mask_bce_loss": 0.7234537601470947, "mask_dice_loss": 0.08654172718524933, "mask_loss": 0.8099954724311829, "step": 1093 }, { "epoch": 0.6570570570570571, "grad_norm": 26.158597946166992, "learning_rate": 5.577476570071424e-07, "loss": 0.9461795091629028, "step": 1094 }, { "ce_loss": 0.00016400164167862386, "cls_loss": 0.05322265625, "epoch": 0.6570570570570571, "mask_bce_loss": 0.7582090497016907, "mask_dice_loss": 0.1465895026922226, "mask_loss": 0.9047985672950745, "step": 1094 }, { "epoch": 0.6576576576576577, "grad_norm": 54.99756622314453, "learning_rate": 5.560038105678823e-07, "loss": 0.854467511177063, "step": 1095 }, { "ce_loss": 2.8530255804071203e-05, "cls_loss": 0.048095703125, "epoch": 0.6576576576576577, "mask_bce_loss": 0.1834159940481186, "mask_dice_loss": 0.04903111606836319, "mask_loss": 0.23244711756706238, "step": 1095 }, { "epoch": 0.6582582582582582, "grad_norm": 40.621002197265625, "learning_rate": 5.542616442234618e-07, "loss": 0.816595733165741, "step": 1096 }, { "ce_loss": 0.0001444839290343225, "cls_loss": 0.06787109375, "epoch": 0.6582582582582582, "mask_bce_loss": 1.008893609046936, "mask_dice_loss": 0.12886402010917664, "mask_loss": 1.137757658958435, "step": 1096 }, { "epoch": 0.6588588588588589, "grad_norm": 17.926815032958984, "learning_rate": 5.525211645662909e-07, "loss": 0.7720363736152649, "step": 1097 }, { "ce_loss": 0.0002955140662379563, "cls_loss": 0.034912109375, "epoch": 0.6588588588588589, "mask_bce_loss": 0.2859567105770111, "mask_dice_loss": 0.0297720767557621, "mask_loss": 0.3157287836074829, "step": 1097 }, { "epoch": 0.6594594594594595, "grad_norm": 26.250267028808594, "learning_rate": 5.50782378182396e-07, "loss": 0.783390998840332, "step": 1098 }, { "ce_loss": 4.5881810365244746e-05, "cls_loss": 0.06396484375, "epoch": 0.6594594594594595, "mask_bce_loss": 0.3999800384044647, "mask_dice_loss": 0.15471425652503967, "mask_loss": 0.5546942949295044, "step": 1098 }, { "epoch": 0.66006006006006, "grad_norm": 36.17412567138672, "learning_rate": 5.490452916513966e-07, "loss": 0.8107259273529053, "step": 1099 }, { "ce_loss": 3.425883187446743e-05, "cls_loss": 0.05712890625, "epoch": 0.66006006006006, "mask_bce_loss": 0.5644844174385071, "mask_dice_loss": 0.14687104523181915, "mask_loss": 0.711355447769165, "step": 1099 }, { "epoch": 0.6606606606606606, "grad_norm": 36.227325439453125, "learning_rate": 5.47309911546481e-07, "loss": 0.7128921151161194, "step": 1100 }, { "ce_loss": 0.00015998944581951946, "cls_loss": 0.05908203125, "epoch": 0.6606606606606606, "mask_bce_loss": 1.822829246520996, "mask_dice_loss": 0.10740398615598679, "mask_loss": 1.9302332401275635, "step": 1100 }, { "epoch": 0.6612612612612613, "grad_norm": 23.733182907104492, "learning_rate": 5.455762444343784e-07, "loss": 0.8800241947174072, "step": 1101 }, { "ce_loss": 0.00011350620479788631, "cls_loss": 0.0302734375, "epoch": 0.6612612612612613, "mask_bce_loss": 0.8265017867088318, "mask_dice_loss": 0.06284927576780319, "mask_loss": 0.8893510699272156, "step": 1101 }, { "epoch": 0.6618618618618619, "grad_norm": 21.35089111328125, "learning_rate": 5.438442968753373e-07, "loss": 0.8663875460624695, "step": 1102 }, { "ce_loss": 3.904668119503185e-05, "cls_loss": 0.037109375, "epoch": 0.6618618618618619, "mask_bce_loss": 0.8347412943840027, "mask_dice_loss": 0.08790981769561768, "mask_loss": 0.9226511120796204, "step": 1102 }, { "epoch": 0.6624624624624624, "grad_norm": 18.978376388549805, "learning_rate": 5.421140754230991e-07, "loss": 0.8169749975204468, "step": 1103 }, { "ce_loss": 0.0029503123369067907, "cls_loss": 0.0257568359375, "epoch": 0.6624624624624624, "mask_bce_loss": 0.23658128082752228, "mask_dice_loss": 0.019597196951508522, "mask_loss": 0.25617846846580505, "step": 1103 }, { "epoch": 0.6630630630630631, "grad_norm": 29.006881713867188, "learning_rate": 5.40385586624873e-07, "loss": 0.8153474926948547, "step": 1104 }, { "ce_loss": 0.0016990413423627615, "cls_loss": 0.042724609375, "epoch": 0.6630630630630631, "mask_bce_loss": 1.0164037942886353, "mask_dice_loss": 0.07055419683456421, "mask_loss": 1.0869579315185547, "step": 1104 }, { "epoch": 0.6636636636636637, "grad_norm": 17.53903579711914, "learning_rate": 5.386588370213123e-07, "loss": 0.8211942315101624, "step": 1105 }, { "ce_loss": 2.9715840355493128e-05, "cls_loss": 0.032958984375, "epoch": 0.6636636636636637, "mask_bce_loss": 0.3999798893928528, "mask_dice_loss": 0.02829734794795513, "mask_loss": 0.42827722430229187, "step": 1105 }, { "epoch": 0.6642642642642642, "grad_norm": 26.546295166015625, "learning_rate": 5.369338331464894e-07, "loss": 0.8685742616653442, "step": 1106 }, { "ce_loss": 0.00011065557191614062, "cls_loss": 0.05078125, "epoch": 0.6642642642642642, "mask_bce_loss": 0.5606155395507812, "mask_dice_loss": 0.07781963050365448, "mask_loss": 0.6384351849555969, "step": 1106 }, { "epoch": 0.6648648648648648, "grad_norm": 24.65392303466797, "learning_rate": 5.352105815278704e-07, "loss": 0.6883245706558228, "step": 1107 }, { "ce_loss": 0.00012129079550504684, "cls_loss": 0.059326171875, "epoch": 0.6648648648648648, "mask_bce_loss": 0.6263196468353271, "mask_dice_loss": 0.10146673023700714, "mask_loss": 0.7277863621711731, "step": 1107 }, { "epoch": 0.6654654654654655, "grad_norm": 24.465288162231445, "learning_rate": 5.334890886862905e-07, "loss": 0.8534889221191406, "step": 1108 }, { "ce_loss": 0.0034408080391585827, "cls_loss": 0.044677734375, "epoch": 0.6654654654654655, "mask_bce_loss": 0.9637832045555115, "mask_dice_loss": 0.06911099702119827, "mask_loss": 1.032894253730774, "step": 1108 }, { "epoch": 0.6660660660660661, "grad_norm": 20.13339614868164, "learning_rate": 5.317693611359306e-07, "loss": 0.7846341729164124, "step": 1109 }, { "ce_loss": 6.230138387763873e-05, "cls_loss": 0.04150390625, "epoch": 0.6660660660660661, "mask_bce_loss": 0.7334949374198914, "mask_dice_loss": 0.05589587241411209, "mask_loss": 0.7893908023834229, "step": 1109 }, { "epoch": 0.6666666666666666, "grad_norm": 21.627548217773438, "learning_rate": 5.300514053842909e-07, "loss": 0.8274050951004028, "step": 1110 }, { "ce_loss": 0.0006451454828493297, "cls_loss": 0.042236328125, "epoch": 0.6666666666666666, "mask_bce_loss": 0.6647471785545349, "mask_dice_loss": 0.04983432590961456, "mask_loss": 0.7145814895629883, "step": 1110 }, { "epoch": 0.6672672672672673, "grad_norm": 28.751842498779297, "learning_rate": 5.283352279321673e-07, "loss": 0.8793879747390747, "step": 1111 }, { "ce_loss": 2.6679397706175223e-05, "cls_loss": 0.03857421875, "epoch": 0.6672672672672673, "mask_bce_loss": 0.3583337664604187, "mask_dice_loss": 0.029949504882097244, "mask_loss": 0.38828328251838684, "step": 1111 }, { "epoch": 0.6678678678678679, "grad_norm": 23.256629943847656, "learning_rate": 5.266208352736262e-07, "loss": 0.6909419298171997, "step": 1112 }, { "ce_loss": 8.451208123005927e-05, "cls_loss": 0.0211181640625, "epoch": 0.6678678678678679, "mask_bce_loss": 0.16836261749267578, "mask_dice_loss": 0.03560016676783562, "mask_loss": 0.2039627879858017, "step": 1112 }, { "epoch": 0.6684684684684684, "grad_norm": 21.66558837890625, "learning_rate": 5.249082338959809e-07, "loss": 0.8519477844238281, "step": 1113 }, { "ce_loss": 7.35605790396221e-05, "cls_loss": 0.0299072265625, "epoch": 0.6684684684684684, "mask_bce_loss": 0.19223855435848236, "mask_dice_loss": 0.04382941871881485, "mask_loss": 0.2360679805278778, "step": 1113 }, { "epoch": 0.669069069069069, "grad_norm": 32.08285140991211, "learning_rate": 5.231974302797659e-07, "loss": 0.8257429599761963, "step": 1114 }, { "ce_loss": 2.6328641979489475e-05, "cls_loss": 0.050048828125, "epoch": 0.669069069069069, "mask_bce_loss": 0.9687663912773132, "mask_dice_loss": 0.06293385475873947, "mask_loss": 1.0317002534866333, "step": 1114 }, { "epoch": 0.6696696696696697, "grad_norm": 62.394832611083984, "learning_rate": 5.214884308987136e-07, "loss": 0.8957319855690002, "step": 1115 }, { "ce_loss": 5.230031456449069e-05, "cls_loss": 0.056640625, "epoch": 0.6696696696696697, "mask_bce_loss": 0.5684045553207397, "mask_dice_loss": 0.1057010293006897, "mask_loss": 0.6741055846214294, "step": 1115 }, { "epoch": 0.6702702702702703, "grad_norm": 49.811405181884766, "learning_rate": 5.197812422197285e-07, "loss": 0.8515551090240479, "step": 1116 }, { "ce_loss": 0.0002443798875901848, "cls_loss": 0.045654296875, "epoch": 0.6702702702702703, "mask_bce_loss": 0.8126985430717468, "mask_dice_loss": 0.06526744365692139, "mask_loss": 0.8779659867286682, "step": 1116 }, { "epoch": 0.6708708708708708, "grad_norm": 38.52351760864258, "learning_rate": 5.180758707028631e-07, "loss": 0.9928258061408997, "step": 1117 }, { "ce_loss": 3.52290844602976e-05, "cls_loss": 0.030029296875, "epoch": 0.6708708708708708, "mask_bce_loss": 0.24183709919452667, "mask_dice_loss": 0.02196481265127659, "mask_loss": 0.2638019025325775, "step": 1117 }, { "epoch": 0.6714714714714715, "grad_norm": 23.370315551757812, "learning_rate": 5.163723228012949e-07, "loss": 0.8138827085494995, "step": 1118 }, { "ce_loss": 6.055615085642785e-05, "cls_loss": 0.027099609375, "epoch": 0.6714714714714715, "mask_bce_loss": 0.23161418735980988, "mask_dice_loss": 0.02239459566771984, "mask_loss": 0.25400876998901367, "step": 1118 }, { "epoch": 0.6720720720720721, "grad_norm": 27.954191207885742, "learning_rate": 5.146706049612996e-07, "loss": 0.8735445141792297, "step": 1119 }, { "ce_loss": 5.291398338158615e-05, "cls_loss": 0.057861328125, "epoch": 0.6720720720720721, "mask_bce_loss": 0.711558997631073, "mask_dice_loss": 0.06747634708881378, "mask_loss": 0.7790353298187256, "step": 1119 }, { "epoch": 0.6726726726726727, "grad_norm": 29.41691780090332, "learning_rate": 5.129707236222287e-07, "loss": 0.7439494729042053, "step": 1120 }, { "ce_loss": 2.4127211872837506e-05, "cls_loss": 0.033935546875, "epoch": 0.6726726726726727, "mask_bce_loss": 0.3608381450176239, "mask_dice_loss": 0.04716401547193527, "mask_loss": 0.40800216794013977, "step": 1120 }, { "epoch": 0.6732732732732732, "grad_norm": 97.22419738769531, "learning_rate": 5.112726852164835e-07, "loss": 0.8604656457901001, "step": 1121 }, { "ce_loss": 3.460602601990104e-05, "cls_loss": 0.0634765625, "epoch": 0.6732732732732732, "mask_bce_loss": 0.6604005694389343, "mask_dice_loss": 0.08413369953632355, "mask_loss": 0.7445342540740967, "step": 1121 }, { "epoch": 0.6738738738738739, "grad_norm": 38.56829833984375, "learning_rate": 5.095764961694922e-07, "loss": 0.7550793886184692, "step": 1122 }, { "ce_loss": 0.00025781310978345573, "cls_loss": 0.04931640625, "epoch": 0.6738738738738739, "mask_bce_loss": 0.7572016716003418, "mask_dice_loss": 0.1274232119321823, "mask_loss": 0.8846248984336853, "step": 1122 }, { "epoch": 0.6744744744744745, "grad_norm": 23.75174903869629, "learning_rate": 5.078821628996844e-07, "loss": 0.7727333307266235, "step": 1123 }, { "ce_loss": 3.435094185988419e-05, "cls_loss": 0.0478515625, "epoch": 0.6744744744744745, "mask_bce_loss": 0.7674962878227234, "mask_dice_loss": 0.05598428472876549, "mask_loss": 0.8234805464744568, "step": 1123 }, { "epoch": 0.675075075075075, "grad_norm": 23.743560791015625, "learning_rate": 5.061896918184686e-07, "loss": 0.7422523498535156, "step": 1124 }, { "ce_loss": 8.119444100884721e-05, "cls_loss": 0.0390625, "epoch": 0.675075075075075, "mask_bce_loss": 0.5808861255645752, "mask_dice_loss": 0.04455335810780525, "mask_loss": 0.625439465045929, "step": 1124 }, { "epoch": 0.6756756756756757, "grad_norm": 36.6866455078125, "learning_rate": 5.044990893302055e-07, "loss": 0.8197721242904663, "step": 1125 }, { "ce_loss": 0.00868211593478918, "cls_loss": 0.0791015625, "epoch": 0.6756756756756757, "mask_bce_loss": 2.3677642345428467, "mask_dice_loss": 0.14709047973155975, "mask_loss": 2.514854669570923, "step": 1125 }, { "epoch": 0.6762762762762763, "grad_norm": 24.01559066772461, "learning_rate": 5.028103618321851e-07, "loss": 0.8133821487426758, "step": 1126 }, { "ce_loss": 4.536878986982629e-05, "cls_loss": 0.03369140625, "epoch": 0.6762762762762763, "mask_bce_loss": 0.5493848919868469, "mask_dice_loss": 0.07035116851329803, "mask_loss": 0.6197360754013062, "step": 1126 }, { "epoch": 0.6768768768768769, "grad_norm": 21.25687599182129, "learning_rate": 5.011235157146032e-07, "loss": 0.8278915882110596, "step": 1127 }, { "ce_loss": 0.000792351143900305, "cls_loss": 0.0693359375, "epoch": 0.6768768768768769, "mask_bce_loss": 0.31875425577163696, "mask_dice_loss": 0.09259939193725586, "mask_loss": 0.4113536477088928, "step": 1127 }, { "epoch": 0.6774774774774774, "grad_norm": 16.479660034179688, "learning_rate": 4.994385573605359e-07, "loss": 0.8234478235244751, "step": 1128 }, { "ce_loss": 0.0033542446326464415, "cls_loss": 0.041748046875, "epoch": 0.6774774774774774, "mask_bce_loss": 0.8891078233718872, "mask_dice_loss": 0.046367909759283066, "mask_loss": 0.9354757070541382, "step": 1128 }, { "epoch": 0.6780780780780781, "grad_norm": 22.923126220703125, "learning_rate": 4.977554931459155e-07, "loss": 0.8578071594238281, "step": 1129 }, { "ce_loss": 0.00037192352465353906, "cls_loss": 0.028564453125, "epoch": 0.6780780780780781, "mask_bce_loss": 0.19552980363368988, "mask_dice_loss": 0.04475477337837219, "mask_loss": 0.24028457701206207, "step": 1129 }, { "epoch": 0.6786786786786787, "grad_norm": 64.68428802490234, "learning_rate": 4.960743294395076e-07, "loss": 0.8570294380187988, "step": 1130 }, { "ce_loss": 0.000338762387400493, "cls_loss": 0.0546875, "epoch": 0.6786786786786787, "mask_bce_loss": 0.9216718673706055, "mask_dice_loss": 0.12200441211462021, "mask_loss": 1.0436762571334839, "step": 1130 }, { "epoch": 0.6792792792792792, "grad_norm": 16.331857681274414, "learning_rate": 4.943950726028857e-07, "loss": 0.7387335300445557, "step": 1131 }, { "ce_loss": 0.0001853672438301146, "cls_loss": 0.06298828125, "epoch": 0.6792792792792792, "mask_bce_loss": 0.5758478045463562, "mask_dice_loss": 0.11473844200372696, "mask_loss": 0.690586268901825, "step": 1131 }, { "epoch": 0.6798798798798799, "grad_norm": 17.239118576049805, "learning_rate": 4.927177289904074e-07, "loss": 0.827846884727478, "step": 1132 }, { "ce_loss": 3.6487621400738135e-05, "cls_loss": 0.0712890625, "epoch": 0.6798798798798799, "mask_bce_loss": 0.8369947671890259, "mask_dice_loss": 0.13908621668815613, "mask_loss": 0.9760810136795044, "step": 1132 }, { "epoch": 0.6804804804804805, "grad_norm": 23.006267547607422, "learning_rate": 4.910423049491919e-07, "loss": 0.7430271506309509, "step": 1133 }, { "ce_loss": 9.260119259124622e-05, "cls_loss": 0.05419921875, "epoch": 0.6804804804804805, "mask_bce_loss": 0.6643691658973694, "mask_dice_loss": 0.100799061357975, "mask_loss": 0.7651682496070862, "step": 1133 }, { "epoch": 0.6810810810810811, "grad_norm": 45.717830657958984, "learning_rate": 4.893688068190932e-07, "loss": 0.7438549995422363, "step": 1134 }, { "ce_loss": 3.936944267479703e-05, "cls_loss": 0.06201171875, "epoch": 0.6810810810810811, "mask_bce_loss": 0.8565943837165833, "mask_dice_loss": 0.08431460708379745, "mask_loss": 0.9409089684486389, "step": 1134 }, { "epoch": 0.6816816816816816, "grad_norm": 28.078338623046875, "learning_rate": 4.876972409326782e-07, "loss": 0.8959178924560547, "step": 1135 }, { "ce_loss": 6.056439451640472e-05, "cls_loss": 0.06298828125, "epoch": 0.6816816816816816, "mask_bce_loss": 0.6126371622085571, "mask_dice_loss": 0.1281200498342514, "mask_loss": 0.7407572269439697, "step": 1135 }, { "epoch": 0.6822822822822823, "grad_norm": 18.72952651977539, "learning_rate": 4.860276136152025e-07, "loss": 0.8393098711967468, "step": 1136 }, { "ce_loss": 0.0001364795316476375, "cls_loss": 0.0546875, "epoch": 0.6822822822822823, "mask_bce_loss": 0.5450107455253601, "mask_dice_loss": 0.12434177845716476, "mask_loss": 0.6693525314331055, "step": 1136 }, { "epoch": 0.6828828828828829, "grad_norm": 30.794740676879883, "learning_rate": 4.843599311845856e-07, "loss": 0.8259680867195129, "step": 1137 }, { "ce_loss": 2.9491671739378944e-05, "cls_loss": 0.06103515625, "epoch": 0.6828828828828829, "mask_bce_loss": 1.083146572113037, "mask_dice_loss": 0.09033676236867905, "mask_loss": 1.1734833717346191, "step": 1137 }, { "epoch": 0.6834834834834835, "grad_norm": 22.415578842163086, "learning_rate": 4.826941999513876e-07, "loss": 0.7972350120544434, "step": 1138 }, { "ce_loss": 2.5807170459302142e-05, "cls_loss": 0.06103515625, "epoch": 0.6834834834834835, "mask_bce_loss": 0.45243191719055176, "mask_dice_loss": 0.12954625487327576, "mask_loss": 0.5819782018661499, "step": 1138 }, { "epoch": 0.684084084084084, "grad_norm": 19.70414161682129, "learning_rate": 4.810304262187851e-07, "loss": 0.7538164854049683, "step": 1139 }, { "ce_loss": 4.277740663383156e-05, "cls_loss": 0.06005859375, "epoch": 0.684084084084084, "mask_bce_loss": 0.6049984097480774, "mask_dice_loss": 0.12683290243148804, "mask_loss": 0.7318313121795654, "step": 1139 }, { "epoch": 0.6846846846846847, "grad_norm": 57.446529388427734, "learning_rate": 4.793686162825476e-07, "loss": 0.7780901789665222, "step": 1140 }, { "ce_loss": 4.495868051890284e-05, "cls_loss": 0.053955078125, "epoch": 0.6846846846846847, "mask_bce_loss": 0.3986762464046478, "mask_dice_loss": 0.11334868520498276, "mask_loss": 0.5120249390602112, "step": 1140 }, { "epoch": 0.6852852852852853, "grad_norm": 32.966495513916016, "learning_rate": 4.777087764310134e-07, "loss": 0.8516100645065308, "step": 1141 }, { "ce_loss": 5.777739352197386e-05, "cls_loss": 0.047607421875, "epoch": 0.6852852852852853, "mask_bce_loss": 0.4648350775241852, "mask_dice_loss": 0.0661962553858757, "mask_loss": 0.5310313105583191, "step": 1141 }, { "epoch": 0.6858858858858858, "grad_norm": 44.50619125366211, "learning_rate": 4.760509129450665e-07, "loss": 0.8967255353927612, "step": 1142 }, { "ce_loss": 0.0001470099959988147, "cls_loss": 0.015625, "epoch": 0.6858858858858858, "mask_bce_loss": 0.1883268505334854, "mask_dice_loss": 0.009472179226577282, "mask_loss": 0.19779902696609497, "step": 1142 }, { "epoch": 0.6864864864864865, "grad_norm": 41.839908599853516, "learning_rate": 4.743950320981116e-07, "loss": 1.0308434963226318, "step": 1143 }, { "ce_loss": 9.690222213976085e-05, "cls_loss": 0.068359375, "epoch": 0.6864864864864865, "mask_bce_loss": 0.2708527147769928, "mask_dice_loss": 0.09597416967153549, "mask_loss": 0.3668268918991089, "step": 1143 }, { "epoch": 0.6870870870870871, "grad_norm": 37.000675201416016, "learning_rate": 4.7274114015605093e-07, "loss": 0.8493385314941406, "step": 1144 }, { "ce_loss": 8.162917220033705e-05, "cls_loss": 0.0299072265625, "epoch": 0.6870870870870871, "mask_bce_loss": 0.27371251583099365, "mask_dice_loss": 0.04510653764009476, "mask_loss": 0.3188190460205078, "step": 1144 }, { "epoch": 0.6876876876876877, "grad_norm": 24.880050659179688, "learning_rate": 4.710892433772616e-07, "loss": 0.7485247254371643, "step": 1145 }, { "ce_loss": 2.6959271053783596e-05, "cls_loss": 0.06787109375, "epoch": 0.6876876876876877, "mask_bce_loss": 0.7021887898445129, "mask_dice_loss": 0.09158699959516525, "mask_loss": 0.7937757968902588, "step": 1145 }, { "epoch": 0.6882882882882883, "grad_norm": 36.40140914916992, "learning_rate": 4.6943934801257025e-07, "loss": 0.7791830897331238, "step": 1146 }, { "ce_loss": 0.00016889294784050435, "cls_loss": 0.04638671875, "epoch": 0.6882882882882883, "mask_bce_loss": 0.9601742029190063, "mask_dice_loss": 0.10608130693435669, "mask_loss": 1.0662555694580078, "step": 1146 }, { "epoch": 0.6888888888888889, "grad_norm": 28.559329986572266, "learning_rate": 4.6779146030523e-07, "loss": 0.8213381767272949, "step": 1147 }, { "ce_loss": 3.776098310481757e-05, "cls_loss": 0.038818359375, "epoch": 0.6888888888888889, "mask_bce_loss": 0.4229496121406555, "mask_dice_loss": 0.052668094635009766, "mask_loss": 0.4756177067756653, "step": 1147 }, { "epoch": 0.6894894894894895, "grad_norm": 19.578048706054688, "learning_rate": 4.661455864908971e-07, "loss": 0.8252648115158081, "step": 1148 }, { "ce_loss": 5.0101123633794487e-05, "cls_loss": 0.052001953125, "epoch": 0.6894894894894895, "mask_bce_loss": 0.3443085849285126, "mask_dice_loss": 0.07498065382242203, "mask_loss": 0.419289231300354, "step": 1148 }, { "epoch": 0.69009009009009, "grad_norm": 31.91105842590332, "learning_rate": 4.645017327976073e-07, "loss": 0.695225179195404, "step": 1149 }, { "ce_loss": 4.0857757994672284e-05, "cls_loss": 0.042236328125, "epoch": 0.69009009009009, "mask_bce_loss": 0.48493385314941406, "mask_dice_loss": 0.052426647394895554, "mask_loss": 0.5373604893684387, "step": 1149 }, { "epoch": 0.6906906906906907, "grad_norm": 24.569129943847656, "learning_rate": 4.628599054457517e-07, "loss": 0.7607369422912598, "step": 1150 }, { "ce_loss": 0.0005857331561855972, "cls_loss": 0.03369140625, "epoch": 0.6906906906906907, "mask_bce_loss": 0.38987597823143005, "mask_dice_loss": 0.04226161912083626, "mask_loss": 0.4321376085281372, "step": 1150 }, { "epoch": 0.6912912912912913, "grad_norm": 16.092670440673828, "learning_rate": 4.6122011064805477e-07, "loss": 0.8142052888870239, "step": 1151 }, { "ce_loss": 0.00015938421711325645, "cls_loss": 0.0546875, "epoch": 0.6912912912912913, "mask_bce_loss": 0.9016212821006775, "mask_dice_loss": 0.08187609165906906, "mask_loss": 0.9834973812103271, "step": 1151 }, { "epoch": 0.6918918918918919, "grad_norm": 35.87517547607422, "learning_rate": 4.595823546095485e-07, "loss": 0.8619617223739624, "step": 1152 }, { "ce_loss": 0.0008390736184082925, "cls_loss": 0.05615234375, "epoch": 0.6918918918918919, "mask_bce_loss": 0.7432164549827576, "mask_dice_loss": 0.15483631193637848, "mask_loss": 0.8980527520179749, "step": 1152 }, { "epoch": 0.6924924924924925, "grad_norm": 11.064460754394531, "learning_rate": 4.5794664352755054e-07, "loss": 0.8196966648101807, "step": 1153 }, { "ce_loss": 6.213867163751274e-05, "cls_loss": 0.024169921875, "epoch": 0.6924924924924925, "mask_bce_loss": 0.541601300239563, "mask_dice_loss": 0.016846762970089912, "mask_loss": 0.558448076248169, "step": 1153 }, { "epoch": 0.6930930930930931, "grad_norm": 57.405311584472656, "learning_rate": 4.5631298359164103e-07, "loss": 0.832563042640686, "step": 1154 }, { "ce_loss": 3.15528450300917e-05, "cls_loss": 0.072265625, "epoch": 0.6930930930930931, "mask_bce_loss": 0.6586489677429199, "mask_dice_loss": 0.12968158721923828, "mask_loss": 0.7883305549621582, "step": 1154 }, { "epoch": 0.6936936936936937, "grad_norm": 55.985939025878906, "learning_rate": 4.546813809836376e-07, "loss": 0.974033534526825, "step": 1155 }, { "ce_loss": 0.00014681681932415813, "cls_loss": 0.03466796875, "epoch": 0.6936936936936937, "mask_bce_loss": 0.4357403814792633, "mask_dice_loss": 0.05280221626162529, "mask_loss": 0.4885425865650177, "step": 1155 }, { "epoch": 0.6942942942942943, "grad_norm": 20.715320587158203, "learning_rate": 4.530518418775733e-07, "loss": 0.8541744947433472, "step": 1156 }, { "ce_loss": 2.178672184527386e-05, "cls_loss": 0.06103515625, "epoch": 0.6942942942942943, "mask_bce_loss": 1.1563634872436523, "mask_dice_loss": 0.08182156085968018, "mask_loss": 1.2381850481033325, "step": 1156 }, { "epoch": 0.6948948948948949, "grad_norm": 45.27634811401367, "learning_rate": 4.514243724396727e-07, "loss": 0.9321890473365784, "step": 1157 }, { "ce_loss": 6.909380317665637e-05, "cls_loss": 0.059326171875, "epoch": 0.6948948948948949, "mask_bce_loss": 0.5008495450019836, "mask_dice_loss": 0.07360950857400894, "mask_loss": 0.5744590759277344, "step": 1157 }, { "epoch": 0.6954954954954955, "grad_norm": 23.47991943359375, "learning_rate": 4.497989788283284e-07, "loss": 0.7685770988464355, "step": 1158 }, { "ce_loss": 5.084984877612442e-05, "cls_loss": 0.0361328125, "epoch": 0.6954954954954955, "mask_bce_loss": 0.9392598271369934, "mask_dice_loss": 0.056692369282245636, "mask_loss": 0.9959521889686584, "step": 1158 }, { "epoch": 0.6960960960960961, "grad_norm": 19.915437698364258, "learning_rate": 4.4817566719407906e-07, "loss": 0.7402971982955933, "step": 1159 }, { "ce_loss": 7.460681808879599e-05, "cls_loss": 0.05615234375, "epoch": 0.6960960960960961, "mask_bce_loss": 1.1788082122802734, "mask_dice_loss": 0.1552373766899109, "mask_loss": 1.334045648574829, "step": 1159 }, { "epoch": 0.6966966966966966, "grad_norm": 21.76595115661621, "learning_rate": 4.465544436795842e-07, "loss": 0.7486002445220947, "step": 1160 }, { "ce_loss": 3.5219822166254744e-05, "cls_loss": 0.04638671875, "epoch": 0.6966966966966966, "mask_bce_loss": 0.5338829755783081, "mask_dice_loss": 0.05753563717007637, "mask_loss": 0.5914186239242554, "step": 1160 }, { "epoch": 0.6972972972972973, "grad_norm": 61.352699279785156, "learning_rate": 4.449353144196014e-07, "loss": 0.9775881767272949, "step": 1161 }, { "ce_loss": 8.918061212170869e-05, "cls_loss": 0.05859375, "epoch": 0.6972972972972973, "mask_bce_loss": 0.7294099926948547, "mask_dice_loss": 0.15274445712566376, "mask_loss": 0.8821544647216797, "step": 1161 }, { "epoch": 0.6978978978978979, "grad_norm": 21.645788192749023, "learning_rate": 4.4331828554096506e-07, "loss": 0.8946480751037598, "step": 1162 }, { "ce_loss": 4.45028163085226e-05, "cls_loss": 0.061279296875, "epoch": 0.6978978978978979, "mask_bce_loss": 0.7368199825286865, "mask_dice_loss": 0.12571783363819122, "mask_loss": 0.8625378012657166, "step": 1162 }, { "epoch": 0.6984984984984985, "grad_norm": 20.15625, "learning_rate": 4.417033631625602e-07, "loss": 0.8862178325653076, "step": 1163 }, { "ce_loss": 3.6160872696200386e-05, "cls_loss": 0.0419921875, "epoch": 0.6984984984984985, "mask_bce_loss": 0.17076025903224945, "mask_dice_loss": 0.03684141859412193, "mask_loss": 0.20760168135166168, "step": 1163 }, { "epoch": 0.6990990990990991, "grad_norm": 21.29073143005371, "learning_rate": 4.4009055339530145e-07, "loss": 0.8241890072822571, "step": 1164 }, { "ce_loss": 0.0037724950816482306, "cls_loss": 0.0419921875, "epoch": 0.6990990990990991, "mask_bce_loss": 0.5279273986816406, "mask_dice_loss": 0.062865249812603, "mask_loss": 0.5907926559448242, "step": 1164 }, { "epoch": 0.6996996996996997, "grad_norm": 40.66654968261719, "learning_rate": 4.384798623421092e-07, "loss": 0.8708781599998474, "step": 1165 }, { "ce_loss": 2.8109692721045576e-05, "cls_loss": 0.051513671875, "epoch": 0.6996996996996997, "mask_bce_loss": 0.8849459886550903, "mask_dice_loss": 0.054152775555849075, "mask_loss": 0.9390987753868103, "step": 1165 }, { "epoch": 0.7003003003003003, "grad_norm": 27.730464935302734, "learning_rate": 4.368712960978863e-07, "loss": 0.7735524773597717, "step": 1166 }, { "ce_loss": 2.1506077246158384e-05, "cls_loss": 0.027099609375, "epoch": 0.7003003003003003, "mask_bce_loss": 0.1499936729669571, "mask_dice_loss": 0.017697423696517944, "mask_loss": 0.16769109666347504, "step": 1166 }, { "epoch": 0.7009009009009008, "grad_norm": 27.456087112426758, "learning_rate": 4.352648607494952e-07, "loss": 0.959111213684082, "step": 1167 }, { "ce_loss": 4.5460634282790124e-05, "cls_loss": 0.06103515625, "epoch": 0.7009009009009008, "mask_bce_loss": 0.639943540096283, "mask_dice_loss": 0.1365656703710556, "mask_loss": 0.7765092253684998, "step": 1167 }, { "epoch": 0.7015015015015015, "grad_norm": 17.599369049072266, "learning_rate": 4.33660562375736e-07, "loss": 0.7133200168609619, "step": 1168 }, { "ce_loss": 3.3962776797125116e-05, "cls_loss": 0.05859375, "epoch": 0.7015015015015015, "mask_bce_loss": 1.5133458375930786, "mask_dice_loss": 0.16682405769824982, "mask_loss": 1.680169939994812, "step": 1168 }, { "epoch": 0.7021021021021021, "grad_norm": 52.777462005615234, "learning_rate": 4.320584070473212e-07, "loss": 0.8995513319969177, "step": 1169 }, { "ce_loss": 0.00016072012658696622, "cls_loss": 0.05615234375, "epoch": 0.7021021021021021, "mask_bce_loss": 1.0189639329910278, "mask_dice_loss": 0.0975046381354332, "mask_loss": 1.1164685487747192, "step": 1169 }, { "epoch": 0.7027027027027027, "grad_norm": 31.270971298217773, "learning_rate": 4.3045840082685405e-07, "loss": 0.8423636555671692, "step": 1170 }, { "ce_loss": 0.0001435066806152463, "cls_loss": 0.055419921875, "epoch": 0.7027027027027027, "mask_bce_loss": 0.5658193826675415, "mask_dice_loss": 0.08936191350221634, "mask_loss": 0.6551812887191772, "step": 1170 }, { "epoch": 0.7033033033033033, "grad_norm": 54.46002197265625, "learning_rate": 4.288605497688067e-07, "loss": 0.8376576900482178, "step": 1171 }, { "ce_loss": 4.391165930428542e-05, "cls_loss": 0.05419921875, "epoch": 0.7033033033033033, "mask_bce_loss": 0.6437821388244629, "mask_dice_loss": 0.08221843838691711, "mask_loss": 0.7260005474090576, "step": 1171 }, { "epoch": 0.7039039039039039, "grad_norm": 27.53957176208496, "learning_rate": 4.272648599194948e-07, "loss": 0.9227584004402161, "step": 1172 }, { "ce_loss": 0.00023461955424863845, "cls_loss": 0.0634765625, "epoch": 0.7039039039039039, "mask_bce_loss": 0.39974457025527954, "mask_dice_loss": 0.09901909530162811, "mask_loss": 0.49876368045806885, "step": 1172 }, { "epoch": 0.7045045045045045, "grad_norm": 26.628955841064453, "learning_rate": 4.256713373170564e-07, "loss": 0.7740477919578552, "step": 1173 }, { "ce_loss": 0.0001314259134232998, "cls_loss": 0.06103515625, "epoch": 0.7045045045045045, "mask_bce_loss": 0.4372677803039551, "mask_dice_loss": 0.08140963315963745, "mask_loss": 0.5186774134635925, "step": 1173 }, { "epoch": 0.7051051051051052, "grad_norm": 38.01972198486328, "learning_rate": 4.2407998799142875e-07, "loss": 0.7150953412055969, "step": 1174 }, { "ce_loss": 4.4517353671835735e-05, "cls_loss": 0.05322265625, "epoch": 0.7051051051051052, "mask_bce_loss": 1.0761898756027222, "mask_dice_loss": 0.0749962329864502, "mask_loss": 1.1511861085891724, "step": 1174 }, { "epoch": 0.7057057057057057, "grad_norm": 70.0971908569336, "learning_rate": 4.2249081796432495e-07, "loss": 1.1772772073745728, "step": 1175 }, { "ce_loss": 3.3960965083679184e-05, "cls_loss": 0.06201171875, "epoch": 0.7057057057057057, "mask_bce_loss": 0.6502525210380554, "mask_dice_loss": 0.1275099366903305, "mask_loss": 0.7777624726295471, "step": 1175 }, { "epoch": 0.7063063063063063, "grad_norm": 17.474166870117188, "learning_rate": 4.2090383324921176e-07, "loss": 0.8726473450660706, "step": 1176 }, { "ce_loss": 7.414520950987935e-05, "cls_loss": 0.049072265625, "epoch": 0.7063063063063063, "mask_bce_loss": 0.3399368226528168, "mask_dice_loss": 0.04867235943675041, "mask_loss": 0.3886091709136963, "step": 1176 }, { "epoch": 0.7069069069069069, "grad_norm": 19.87905502319336, "learning_rate": 4.1931903985128736e-07, "loss": 0.7552181482315063, "step": 1177 }, { "ce_loss": 8.840102964313701e-05, "cls_loss": 0.0732421875, "epoch": 0.7069069069069069, "mask_bce_loss": 0.8973572850227356, "mask_dice_loss": 0.13456279039382935, "mask_loss": 1.031920075416565, "step": 1177 }, { "epoch": 0.7075075075075075, "grad_norm": 34.70404815673828, "learning_rate": 4.1773644376745684e-07, "loss": 0.8703258037567139, "step": 1178 }, { "ce_loss": 0.00022014562273398042, "cls_loss": 0.053466796875, "epoch": 0.7075075075075075, "mask_bce_loss": 0.8939834833145142, "mask_dice_loss": 0.09821917116641998, "mask_loss": 0.992202639579773, "step": 1178 }, { "epoch": 0.7081081081081081, "grad_norm": 17.79653549194336, "learning_rate": 4.161560509863108e-07, "loss": 0.7329350709915161, "step": 1179 }, { "ce_loss": 0.0008188760839402676, "cls_loss": 0.0751953125, "epoch": 0.7081081081081081, "mask_bce_loss": 0.1828712671995163, "mask_dice_loss": 0.08717665821313858, "mask_loss": 0.2700479328632355, "step": 1179 }, { "epoch": 0.7087087087087087, "grad_norm": 63.283077239990234, "learning_rate": 4.145778674881033e-07, "loss": 1.0338435173034668, "step": 1180 }, { "ce_loss": 2.809460784192197e-05, "cls_loss": 0.043212890625, "epoch": 0.7087087087087087, "mask_bce_loss": 1.3322235345840454, "mask_dice_loss": 0.05414996296167374, "mask_loss": 1.386373519897461, "step": 1180 }, { "epoch": 0.7093093093093094, "grad_norm": 19.627893447875977, "learning_rate": 4.1300189924472763e-07, "loss": 0.7997511625289917, "step": 1181 }, { "ce_loss": 2.922951716755051e-05, "cls_loss": 0.05859375, "epoch": 0.7093093093093094, "mask_bce_loss": 1.3755580186843872, "mask_dice_loss": 0.17593571543693542, "mask_loss": 1.551493763923645, "step": 1181 }, { "epoch": 0.7099099099099099, "grad_norm": 16.814638137817383, "learning_rate": 4.114281522196946e-07, "loss": 0.7746900320053101, "step": 1182 }, { "ce_loss": 5.359272108762525e-05, "cls_loss": 0.05126953125, "epoch": 0.7099099099099099, "mask_bce_loss": 0.6689712405204773, "mask_dice_loss": 0.028001364320516586, "mask_loss": 0.6969726085662842, "step": 1182 }, { "epoch": 0.7105105105105105, "grad_norm": 46.41055679321289, "learning_rate": 4.0985663236811e-07, "loss": 0.8839671611785889, "step": 1183 }, { "ce_loss": 3.641826333478093e-05, "cls_loss": 0.04296875, "epoch": 0.7105105105105105, "mask_bce_loss": 0.5094202160835266, "mask_dice_loss": 0.04194264858961105, "mask_loss": 0.5513628721237183, "step": 1183 }, { "epoch": 0.7111111111111111, "grad_norm": 28.75340461730957, "learning_rate": 4.082873456366518e-07, "loss": 0.7711114287376404, "step": 1184 }, { "ce_loss": 6.014635437168181e-05, "cls_loss": 0.047119140625, "epoch": 0.7111111111111111, "mask_bce_loss": 0.4130900800228119, "mask_dice_loss": 0.14602875709533691, "mask_loss": 0.5591188669204712, "step": 1184 }, { "epoch": 0.7117117117117117, "grad_norm": 39.93330764770508, "learning_rate": 4.0672029796354757e-07, "loss": 0.9524592757225037, "step": 1185 }, { "ce_loss": 0.00045033247442916036, "cls_loss": 0.0703125, "epoch": 0.7117117117117117, "mask_bce_loss": 0.8395175337791443, "mask_dice_loss": 0.09689140319824219, "mask_loss": 0.9364089369773865, "step": 1185 }, { "epoch": 0.7123123123123123, "grad_norm": 30.482961654663086, "learning_rate": 4.0515549527855306e-07, "loss": 0.6781511902809143, "step": 1186 }, { "ce_loss": 6.338535604299977e-05, "cls_loss": 0.0299072265625, "epoch": 0.7123123123123123, "mask_bce_loss": 0.21983149647712708, "mask_dice_loss": 0.022087665274739265, "mask_loss": 0.2419191598892212, "step": 1186 }, { "epoch": 0.7129129129129129, "grad_norm": 21.059524536132812, "learning_rate": 4.0359294350292826e-07, "loss": 0.8140021562576294, "step": 1187 }, { "ce_loss": 3.107693191850558e-05, "cls_loss": 0.04248046875, "epoch": 0.7129129129129129, "mask_bce_loss": 0.69495689868927, "mask_dice_loss": 0.044418394565582275, "mask_loss": 0.7393752932548523, "step": 1187 }, { "epoch": 0.7135135135135136, "grad_norm": 23.50175666809082, "learning_rate": 4.020326485494152e-07, "loss": 0.7148928046226501, "step": 1188 }, { "ce_loss": 2.501717426639516e-05, "cls_loss": 0.05078125, "epoch": 0.7135135135135136, "mask_bce_loss": 0.5074854493141174, "mask_dice_loss": 0.10530436038970947, "mask_loss": 0.6127898097038269, "step": 1188 }, { "epoch": 0.7141141141141141, "grad_norm": 22.30803871154785, "learning_rate": 4.004746163222175e-07, "loss": 0.7723706960678101, "step": 1189 }, { "ce_loss": 6.923300679773092e-05, "cls_loss": 0.05224609375, "epoch": 0.7141141141141141, "mask_bce_loss": 0.6119275689125061, "mask_dice_loss": 0.0702594667673111, "mask_loss": 0.682187020778656, "step": 1189 }, { "epoch": 0.7147147147147147, "grad_norm": 43.87041091918945, "learning_rate": 3.989188527169749e-07, "loss": 0.832994818687439, "step": 1190 }, { "ce_loss": 0.003366809571161866, "cls_loss": 0.0439453125, "epoch": 0.7147147147147147, "mask_bce_loss": 0.3004375100135803, "mask_dice_loss": 0.03862423822283745, "mask_loss": 0.3390617370605469, "step": 1190 }, { "epoch": 0.7153153153153153, "grad_norm": 19.66072654724121, "learning_rate": 3.973653636207437e-07, "loss": 0.8184452056884766, "step": 1191 }, { "ce_loss": 8.068182069109753e-05, "cls_loss": 0.0283203125, "epoch": 0.7153153153153153, "mask_bce_loss": 0.206385999917984, "mask_dice_loss": 0.06608037650585175, "mask_loss": 0.27246636152267456, "step": 1191 }, { "epoch": 0.715915915915916, "grad_norm": 30.988861083984375, "learning_rate": 3.958141549119729e-07, "loss": 0.937623143196106, "step": 1192 }, { "ce_loss": 0.00037033046828582883, "cls_loss": 0.0341796875, "epoch": 0.715915915915916, "mask_bce_loss": 0.2829675078392029, "mask_dice_loss": 0.07209137082099915, "mask_loss": 0.355058878660202, "step": 1192 }, { "epoch": 0.7165165165165165, "grad_norm": 35.859588623046875, "learning_rate": 3.942652324604826e-07, "loss": 0.8220299482345581, "step": 1193 }, { "ce_loss": 3.461242522462271e-05, "cls_loss": 0.0673828125, "epoch": 0.7165165165165165, "mask_bce_loss": 1.0128759145736694, "mask_dice_loss": 0.12313675880432129, "mask_loss": 1.1360126733779907, "step": 1193 }, { "epoch": 0.7171171171171171, "grad_norm": 45.72538375854492, "learning_rate": 3.9271860212744113e-07, "loss": 0.9208501577377319, "step": 1194 }, { "ce_loss": 0.0032940523233264685, "cls_loss": 0.03271484375, "epoch": 0.7171171171171171, "mask_bce_loss": 0.19670800864696503, "mask_dice_loss": 0.052880026400089264, "mask_loss": 0.2495880424976349, "step": 1194 }, { "epoch": 0.7177177177177178, "grad_norm": 12.82708740234375, "learning_rate": 3.911742697653447e-07, "loss": 0.7879343628883362, "step": 1195 }, { "ce_loss": 4.300466025597416e-05, "cls_loss": 0.060546875, "epoch": 0.7177177177177178, "mask_bce_loss": 1.5340827703475952, "mask_dice_loss": 0.09922438114881516, "mask_loss": 1.6333070993423462, "step": 1195 }, { "epoch": 0.7183183183183183, "grad_norm": 24.13795280456543, "learning_rate": 3.896322412179929e-07, "loss": 0.7087438106536865, "step": 1196 }, { "ce_loss": 5.735406375606544e-05, "cls_loss": 0.07275390625, "epoch": 0.7183183183183183, "mask_bce_loss": 0.8402977585792542, "mask_dice_loss": 0.09685030579566956, "mask_loss": 0.9371480941772461, "step": 1196 }, { "epoch": 0.7189189189189189, "grad_norm": 37.742088317871094, "learning_rate": 3.880925223204671e-07, "loss": 0.988191545009613, "step": 1197 }, { "ce_loss": 7.070637366268784e-05, "cls_loss": 0.033203125, "epoch": 0.7189189189189189, "mask_bce_loss": 0.5427871942520142, "mask_dice_loss": 0.04058389738202095, "mask_loss": 0.583371102809906, "step": 1197 }, { "epoch": 0.7195195195195195, "grad_norm": 21.152301788330078, "learning_rate": 3.865551188991104e-07, "loss": 0.6676958203315735, "step": 1198 }, { "ce_loss": 5.731411147280596e-05, "cls_loss": 0.06201171875, "epoch": 0.7195195195195195, "mask_bce_loss": 1.3582098484039307, "mask_dice_loss": 0.13123540580272675, "mask_loss": 1.4894452095031738, "step": 1198 }, { "epoch": 0.7201201201201202, "grad_norm": 29.11947250366211, "learning_rate": 3.850200367715031e-07, "loss": 0.8814848065376282, "step": 1199 }, { "ce_loss": 3.0275576136773452e-05, "cls_loss": 0.04443359375, "epoch": 0.7201201201201202, "mask_bce_loss": 0.5461921691894531, "mask_dice_loss": 0.057918768376111984, "mask_loss": 0.6041109561920166, "step": 1199 }, { "epoch": 0.7207207207207207, "grad_norm": 26.038951873779297, "learning_rate": 3.834872817464415e-07, "loss": 0.743344783782959, "step": 1200 }, { "ce_loss": 7.501478103222325e-05, "cls_loss": 0.05224609375, "epoch": 0.7207207207207207, "mask_bce_loss": 1.419071078300476, "mask_dice_loss": 0.13755492866039276, "mask_loss": 1.5566259622573853, "step": 1200 }, { "epoch": 0.7213213213213213, "grad_norm": 29.75655174255371, "learning_rate": 3.8195685962391664e-07, "loss": 0.8933892250061035, "step": 1201 }, { "ce_loss": 0.0001460396160837263, "cls_loss": 0.053955078125, "epoch": 0.7213213213213213, "mask_bce_loss": 0.605767548084259, "mask_dice_loss": 0.056031256914138794, "mask_loss": 0.6617988348007202, "step": 1201 }, { "epoch": 0.721921921921922, "grad_norm": 37.14545440673828, "learning_rate": 3.804287761950915e-07, "loss": 0.7793497443199158, "step": 1202 }, { "ce_loss": 7.085644756443799e-05, "cls_loss": 0.06689453125, "epoch": 0.721921921921922, "mask_bce_loss": 1.5428094863891602, "mask_dice_loss": 0.14386849105358124, "mask_loss": 1.6866779327392578, "step": 1202 }, { "epoch": 0.7225225225225225, "grad_norm": 25.401992797851562, "learning_rate": 3.789030372422789e-07, "loss": 0.7730610370635986, "step": 1203 }, { "ce_loss": 0.005864724051207304, "cls_loss": 0.055419921875, "epoch": 0.7225225225225225, "mask_bce_loss": 1.2555402517318726, "mask_dice_loss": 0.09126932919025421, "mask_loss": 1.3468096256256104, "step": 1203 }, { "epoch": 0.7231231231231231, "grad_norm": 22.22156524658203, "learning_rate": 3.7737964853892143e-07, "loss": 0.8026453256607056, "step": 1204 }, { "ce_loss": 0.00026995636289939284, "cls_loss": 0.03857421875, "epoch": 0.7231231231231231, "mask_bce_loss": 0.32268598675727844, "mask_dice_loss": 0.032363396137952805, "mask_loss": 0.35504937171936035, "step": 1204 }, { "epoch": 0.7237237237237237, "grad_norm": 21.902101516723633, "learning_rate": 3.7585861584956715e-07, "loss": 0.7100709676742554, "step": 1205 }, { "ce_loss": 5.0549984734971076e-05, "cls_loss": 0.0625, "epoch": 0.7237237237237237, "mask_bce_loss": 0.4866878390312195, "mask_dice_loss": 0.0553341805934906, "mask_loss": 0.5420219898223877, "step": 1205 }, { "epoch": 0.7243243243243244, "grad_norm": 25.04231071472168, "learning_rate": 3.7433994492984875e-07, "loss": 0.7841419577598572, "step": 1206 }, { "ce_loss": 0.00020779033366125077, "cls_loss": 0.05419921875, "epoch": 0.7243243243243244, "mask_bce_loss": 0.6937829256057739, "mask_dice_loss": 0.05626748129725456, "mask_loss": 0.75005042552948, "step": 1206 }, { "epoch": 0.7249249249249249, "grad_norm": 32.10662078857422, "learning_rate": 3.7282364152646295e-07, "loss": 0.7144466638565063, "step": 1207 }, { "ce_loss": 0.00011378181079635397, "cls_loss": 0.041015625, "epoch": 0.7249249249249249, "mask_bce_loss": 0.4104050099849701, "mask_dice_loss": 0.05283054709434509, "mask_loss": 0.4632355570793152, "step": 1207 }, { "epoch": 0.7255255255255255, "grad_norm": 46.591243743896484, "learning_rate": 3.7130971137714684e-07, "loss": 0.8433560729026794, "step": 1208 }, { "ce_loss": 3.6596247809939086e-05, "cls_loss": 0.05517578125, "epoch": 0.7255255255255255, "mask_bce_loss": 0.5245408415794373, "mask_dice_loss": 0.09313847124576569, "mask_loss": 0.6176792979240417, "step": 1208 }, { "epoch": 0.7261261261261261, "grad_norm": 24.801101684570312, "learning_rate": 3.6979816021065734e-07, "loss": 0.7106590270996094, "step": 1209 }, { "ce_loss": 6.789878534618765e-05, "cls_loss": 0.0791015625, "epoch": 0.7261261261261261, "mask_bce_loss": 0.4715322256088257, "mask_dice_loss": 0.04502585157752037, "mask_loss": 0.516558051109314, "step": 1209 }, { "epoch": 0.7267267267267268, "grad_norm": 33.85209274291992, "learning_rate": 3.6828899374674925e-07, "loss": 0.7775120139122009, "step": 1210 }, { "ce_loss": 3.566772284102626e-05, "cls_loss": 0.039794921875, "epoch": 0.7267267267267268, "mask_bce_loss": 0.3176184594631195, "mask_dice_loss": 0.035035744309425354, "mask_loss": 0.35265421867370605, "step": 1210 }, { "epoch": 0.7273273273273273, "grad_norm": 26.183372497558594, "learning_rate": 3.6678221769615335e-07, "loss": 0.9213501811027527, "step": 1211 }, { "ce_loss": 2.986877370858565e-05, "cls_loss": 0.0673828125, "epoch": 0.7273273273273273, "mask_bce_loss": 0.9405078887939453, "mask_dice_loss": 0.10431434959173203, "mask_loss": 1.0448222160339355, "step": 1211 }, { "epoch": 0.7279279279279279, "grad_norm": 23.019115447998047, "learning_rate": 3.652778377605551e-07, "loss": 0.8140228390693665, "step": 1212 }, { "ce_loss": 2.5238761736545712e-05, "cls_loss": 0.055419921875, "epoch": 0.7279279279279279, "mask_bce_loss": 0.64070063829422, "mask_dice_loss": 0.10659211128950119, "mask_loss": 0.7472927570343018, "step": 1212 }, { "epoch": 0.7285285285285286, "grad_norm": 13.784943580627441, "learning_rate": 3.6377585963257363e-07, "loss": 0.616732656955719, "step": 1213 }, { "ce_loss": 2.419922202534508e-05, "cls_loss": 0.05712890625, "epoch": 0.7285285285285286, "mask_bce_loss": 1.060556173324585, "mask_dice_loss": 0.1071828231215477, "mask_loss": 1.1677390336990356, "step": 1213 }, { "epoch": 0.7291291291291291, "grad_norm": 46.168968200683594, "learning_rate": 3.622762889957387e-07, "loss": 1.0687370300292969, "step": 1214 }, { "ce_loss": 0.00040123332291841507, "cls_loss": 0.0654296875, "epoch": 0.7291291291291291, "mask_bce_loss": 0.8769707679748535, "mask_dice_loss": 0.10595095157623291, "mask_loss": 0.9829217195510864, "step": 1214 }, { "epoch": 0.7297297297297297, "grad_norm": 31.217004776000977, "learning_rate": 3.6077913152447005e-07, "loss": 1.0010161399841309, "step": 1215 }, { "ce_loss": 0.00018159854516852647, "cls_loss": 0.05810546875, "epoch": 0.7297297297297297, "mask_bce_loss": 0.4430945813655853, "mask_dice_loss": 0.0555981807410717, "mask_loss": 0.49869275093078613, "step": 1215 }, { "epoch": 0.7303303303303303, "grad_norm": 24.60272789001465, "learning_rate": 3.59284392884057e-07, "loss": 0.8091835975646973, "step": 1216 }, { "ce_loss": 4.5912645873613656e-05, "cls_loss": 0.02392578125, "epoch": 0.7303303303303303, "mask_bce_loss": 0.23023734986782074, "mask_dice_loss": 0.015122498385608196, "mask_loss": 0.2453598529100418, "step": 1216 }, { "epoch": 0.730930930930931, "grad_norm": 35.0788688659668, "learning_rate": 3.577920787306349e-07, "loss": 0.7518908977508545, "step": 1217 }, { "ce_loss": 2.8200491215102375e-05, "cls_loss": 0.05859375, "epoch": 0.730930930930931, "mask_bce_loss": 0.7930306792259216, "mask_dice_loss": 0.09721168130636215, "mask_loss": 0.890242338180542, "step": 1217 }, { "epoch": 0.7315315315315315, "grad_norm": 18.837427139282227, "learning_rate": 3.5630219471116484e-07, "loss": 0.810312032699585, "step": 1218 }, { "ce_loss": 0.0017538649262860417, "cls_loss": 0.03369140625, "epoch": 0.7315315315315315, "mask_bce_loss": 0.4976232647895813, "mask_dice_loss": 0.13393275439739227, "mask_loss": 0.6315560340881348, "step": 1218 }, { "epoch": 0.7321321321321321, "grad_norm": 53.68178176879883, "learning_rate": 3.548147464634127e-07, "loss": 0.6382511854171753, "step": 1219 }, { "ce_loss": 5.6560100347269326e-05, "cls_loss": 0.059814453125, "epoch": 0.7321321321321321, "mask_bce_loss": 0.8367375731468201, "mask_dice_loss": 0.1271825134754181, "mask_loss": 0.9639201164245605, "step": 1219 }, { "epoch": 0.7327327327327328, "grad_norm": 25.196969985961914, "learning_rate": 3.5332973961592695e-07, "loss": 0.7890166640281677, "step": 1220 }, { "ce_loss": 4.181093390798196e-05, "cls_loss": 0.05859375, "epoch": 0.7327327327327328, "mask_bce_loss": 0.5557634234428406, "mask_dice_loss": 0.09448104351758957, "mask_loss": 0.6502444744110107, "step": 1220 }, { "epoch": 0.7333333333333333, "grad_norm": 22.25613784790039, "learning_rate": 3.5184717978801757e-07, "loss": 0.7663915753364563, "step": 1221 }, { "ce_loss": 5.1088936743326485e-05, "cls_loss": 0.0458984375, "epoch": 0.7333333333333333, "mask_bce_loss": 0.7527710795402527, "mask_dice_loss": 0.04455771669745445, "mask_loss": 0.797328770160675, "step": 1221 }, { "epoch": 0.7339339339339339, "grad_norm": 19.083038330078125, "learning_rate": 3.5036707258973584e-07, "loss": 0.6829128861427307, "step": 1222 }, { "ce_loss": 3.679075598483905e-05, "cls_loss": 0.033447265625, "epoch": 0.7339339339339339, "mask_bce_loss": 0.564560055732727, "mask_dice_loss": 0.027943922206759453, "mask_loss": 0.5925039649009705, "step": 1222 }, { "epoch": 0.7345345345345345, "grad_norm": 32.5217399597168, "learning_rate": 3.4888942362185127e-07, "loss": 0.8890805840492249, "step": 1223 }, { "ce_loss": 3.840792487608269e-05, "cls_loss": 0.038818359375, "epoch": 0.7345345345345345, "mask_bce_loss": 0.4949617385864258, "mask_dice_loss": 0.044481128454208374, "mask_loss": 0.5394428968429565, "step": 1223 }, { "epoch": 0.7351351351351352, "grad_norm": 24.400623321533203, "learning_rate": 3.4741423847583127e-07, "loss": 0.7448335886001587, "step": 1224 }, { "ce_loss": 5.049199171480723e-05, "cls_loss": 0.051513671875, "epoch": 0.7351351351351352, "mask_bce_loss": 0.3052065968513489, "mask_dice_loss": 0.10635235160589218, "mask_loss": 0.41155895590782166, "step": 1224 }, { "epoch": 0.7357357357357357, "grad_norm": 42.79981994628906, "learning_rate": 3.459415227338214e-07, "loss": 0.8976421356201172, "step": 1225 }, { "ce_loss": 4.737489143735729e-05, "cls_loss": 0.044677734375, "epoch": 0.7357357357357357, "mask_bce_loss": 0.6374999284744263, "mask_dice_loss": 0.046183373779058456, "mask_loss": 0.6836832761764526, "step": 1225 }, { "epoch": 0.7363363363363363, "grad_norm": 33.62487030029297, "learning_rate": 3.444712819686213e-07, "loss": 0.761591374874115, "step": 1226 }, { "ce_loss": 0.13130977749824524, "cls_loss": 0.06689453125, "epoch": 0.7363363363363363, "mask_bce_loss": 0.771327018737793, "mask_dice_loss": 0.14030110836029053, "mask_loss": 0.9116281270980835, "step": 1226 }, { "epoch": 0.736936936936937, "grad_norm": 44.94691848754883, "learning_rate": 3.430035217436663e-07, "loss": 0.7848438024520874, "step": 1227 }, { "ce_loss": 2.3189079001895152e-05, "cls_loss": 0.06982421875, "epoch": 0.736936936936937, "mask_bce_loss": 0.548376739025116, "mask_dice_loss": 0.13171470165252686, "mask_loss": 0.6800914406776428, "step": 1227 }, { "epoch": 0.7375375375375376, "grad_norm": 25.030855178833008, "learning_rate": 3.4153824761300465e-07, "loss": 0.8690197467803955, "step": 1228 }, { "ce_loss": 5.588664498645812e-05, "cls_loss": 0.03173828125, "epoch": 0.7375375375375376, "mask_bce_loss": 0.578101634979248, "mask_dice_loss": 0.04428219795227051, "mask_loss": 0.6223838329315186, "step": 1228 }, { "epoch": 0.7381381381381381, "grad_norm": 19.85849952697754, "learning_rate": 3.400754651212776e-07, "loss": 0.8221956491470337, "step": 1229 }, { "ce_loss": 0.0002075853117275983, "cls_loss": 0.04541015625, "epoch": 0.7381381381381381, "mask_bce_loss": 0.51225346326828, "mask_dice_loss": 0.08341207355260849, "mask_loss": 0.5956655144691467, "step": 1229 }, { "epoch": 0.7387387387387387, "grad_norm": 39.557559967041016, "learning_rate": 3.3861517980369745e-07, "loss": 0.8061417937278748, "step": 1230 }, { "ce_loss": 3.4689339372562245e-05, "cls_loss": 0.03466796875, "epoch": 0.7387387387387387, "mask_bce_loss": 0.3790600597858429, "mask_dice_loss": 0.03354129195213318, "mask_loss": 0.4126013517379761, "step": 1230 }, { "epoch": 0.7393393393393394, "grad_norm": 47.875545501708984, "learning_rate": 3.37157397186028e-07, "loss": 0.9461987614631653, "step": 1231 }, { "ce_loss": 4.2148600186919793e-05, "cls_loss": 0.06103515625, "epoch": 0.7393393393393394, "mask_bce_loss": 1.1648401021957397, "mask_dice_loss": 0.09981578588485718, "mask_loss": 1.2646558284759521, "step": 1231 }, { "epoch": 0.7399399399399399, "grad_norm": 20.790035247802734, "learning_rate": 3.3570212278456177e-07, "loss": 0.786768913269043, "step": 1232 }, { "ce_loss": 4.874280057265423e-05, "cls_loss": 0.038818359375, "epoch": 0.7399399399399399, "mask_bce_loss": 0.35374608635902405, "mask_dice_loss": 0.03518366068601608, "mask_loss": 0.3889297544956207, "step": 1232 }, { "epoch": 0.7405405405405405, "grad_norm": 34.574729919433594, "learning_rate": 3.3424936210610055e-07, "loss": 0.8155653476715088, "step": 1233 }, { "ce_loss": 0.000263711204752326, "cls_loss": 0.05419921875, "epoch": 0.7405405405405405, "mask_bce_loss": 0.6090500950813293, "mask_dice_loss": 0.1292978823184967, "mask_loss": 0.7383480072021484, "step": 1233 }, { "epoch": 0.7411411411411412, "grad_norm": 32.5678596496582, "learning_rate": 3.3279912064793457e-07, "loss": 0.953860342502594, "step": 1234 }, { "ce_loss": 2.8359934731270187e-05, "cls_loss": 0.03759765625, "epoch": 0.7411411411411412, "mask_bce_loss": 0.3129439055919647, "mask_dice_loss": 0.04007622227072716, "mask_loss": 0.3530201315879822, "step": 1234 }, { "epoch": 0.7417417417417418, "grad_norm": 24.49575424194336, "learning_rate": 3.313514038978205e-07, "loss": 0.8923415541648865, "step": 1235 }, { "ce_loss": 4.060275387018919e-05, "cls_loss": 0.031982421875, "epoch": 0.7417417417417418, "mask_bce_loss": 0.3814306855201721, "mask_dice_loss": 0.09321186691522598, "mask_loss": 0.4746425449848175, "step": 1235 }, { "epoch": 0.7423423423423423, "grad_norm": 19.369674682617188, "learning_rate": 3.2990621733396186e-07, "loss": 0.7167350053787231, "step": 1236 }, { "ce_loss": 2.97336773655843e-05, "cls_loss": 0.04833984375, "epoch": 0.7423423423423423, "mask_bce_loss": 0.5073367953300476, "mask_dice_loss": 0.04230458661913872, "mask_loss": 0.5496413707733154, "step": 1236 }, { "epoch": 0.7429429429429429, "grad_norm": 15.084417343139648, "learning_rate": 3.2846356642498765e-07, "loss": 0.7481588125228882, "step": 1237 }, { "ce_loss": 0.0008240020833909512, "cls_loss": 0.057373046875, "epoch": 0.7429429429429429, "mask_bce_loss": 1.268843650817871, "mask_dice_loss": 0.1232563778758049, "mask_loss": 1.3920999765396118, "step": 1237 }, { "epoch": 0.7435435435435436, "grad_norm": 31.613590240478516, "learning_rate": 3.270234566299319e-07, "loss": 0.8758573532104492, "step": 1238 }, { "ce_loss": 5.503529609995894e-05, "cls_loss": 0.023681640625, "epoch": 0.7435435435435436, "mask_bce_loss": 0.19348876178264618, "mask_dice_loss": 0.016780054196715355, "mask_loss": 0.2102688103914261, "step": 1238 }, { "epoch": 0.7441441441441441, "grad_norm": 27.285018920898438, "learning_rate": 3.255858933982131e-07, "loss": 0.9053739309310913, "step": 1239 }, { "ce_loss": 6.382548599503934e-05, "cls_loss": 0.06689453125, "epoch": 0.7441441441441441, "mask_bce_loss": 0.9331147074699402, "mask_dice_loss": 0.0890795961022377, "mask_loss": 1.022194266319275, "step": 1239 }, { "epoch": 0.7447447447447447, "grad_norm": 23.720455169677734, "learning_rate": 3.241508821696137e-07, "loss": 0.8078469038009644, "step": 1240 }, { "ce_loss": 0.00010542282689129934, "cls_loss": 0.040283203125, "epoch": 0.7447447447447447, "mask_bce_loss": 0.8816419839859009, "mask_dice_loss": 0.04869980737566948, "mask_loss": 0.9303417801856995, "step": 1240 }, { "epoch": 0.7453453453453454, "grad_norm": 32.92934036254883, "learning_rate": 3.227184283742591e-07, "loss": 0.8352999687194824, "step": 1241 }, { "ce_loss": 7.042475772323087e-05, "cls_loss": 0.04736328125, "epoch": 0.7453453453453454, "mask_bce_loss": 1.043950080871582, "mask_dice_loss": 0.06588853150606155, "mask_loss": 1.109838604927063, "step": 1241 }, { "epoch": 0.745945945945946, "grad_norm": 35.56839370727539, "learning_rate": 3.2128853743259654e-07, "loss": 0.8802164793014526, "step": 1242 }, { "ce_loss": 2.473158747307025e-05, "cls_loss": 0.06689453125, "epoch": 0.745945945945946, "mask_bce_loss": 2.158252000808716, "mask_dice_loss": 0.14652122557163239, "mask_loss": 2.3047733306884766, "step": 1242 }, { "epoch": 0.7465465465465465, "grad_norm": 27.350297927856445, "learning_rate": 3.1986121475537696e-07, "loss": 0.932133138179779, "step": 1243 }, { "ce_loss": 2.5408979126950726e-05, "cls_loss": 0.0245361328125, "epoch": 0.7465465465465465, "mask_bce_loss": 0.2530951201915741, "mask_dice_loss": 0.01869446225464344, "mask_loss": 0.2717895805835724, "step": 1243 }, { "epoch": 0.7471471471471471, "grad_norm": 17.668840408325195, "learning_rate": 3.184364657436316e-07, "loss": 0.7344602346420288, "step": 1244 }, { "ce_loss": 3.894828478223644e-05, "cls_loss": 0.039794921875, "epoch": 0.7471471471471471, "mask_bce_loss": 0.3338838517665863, "mask_dice_loss": 0.031701184809207916, "mask_loss": 0.3655850291252136, "step": 1244 }, { "epoch": 0.7477477477477478, "grad_norm": 17.20301055908203, "learning_rate": 3.1701429578865323e-07, "loss": 0.7026867866516113, "step": 1245 }, { "ce_loss": 8.281347254524007e-05, "cls_loss": 0.045654296875, "epoch": 0.7477477477477478, "mask_bce_loss": 0.2055940479040146, "mask_dice_loss": 0.03820216283202171, "mask_loss": 0.2437962144613266, "step": 1245 }, { "epoch": 0.7483483483483484, "grad_norm": 54.47358703613281, "learning_rate": 3.1559471027197537e-07, "loss": 0.903831958770752, "step": 1246 }, { "ce_loss": 0.00023928053269628435, "cls_loss": 0.048583984375, "epoch": 0.7483483483483484, "mask_bce_loss": 0.7244144082069397, "mask_dice_loss": 0.07518471777439117, "mask_loss": 0.7995991110801697, "step": 1246 }, { "epoch": 0.7489489489489489, "grad_norm": 35.2391357421875, "learning_rate": 3.1417771456535213e-07, "loss": 0.8595766425132751, "step": 1247 }, { "ce_loss": 7.968938734848052e-05, "cls_loss": 0.03271484375, "epoch": 0.7489489489489489, "mask_bce_loss": 0.41865596175193787, "mask_dice_loss": 0.027640745043754578, "mask_loss": 0.44629669189453125, "step": 1247 }, { "epoch": 0.7495495495495496, "grad_norm": 22.752267837524414, "learning_rate": 3.127633140307373e-07, "loss": 0.7497740387916565, "step": 1248 }, { "ce_loss": 0.0015301245730370283, "cls_loss": 0.042236328125, "epoch": 0.7495495495495496, "mask_bce_loss": 0.5124723315238953, "mask_dice_loss": 0.08372002094984055, "mask_loss": 0.5961923599243164, "step": 1248 }, { "epoch": 0.7501501501501502, "grad_norm": 38.4029426574707, "learning_rate": 3.1135151402026514e-07, "loss": 0.8635261654853821, "step": 1249 }, { "ce_loss": 3.0900646379450336e-05, "cls_loss": 0.04931640625, "epoch": 0.7501501501501502, "mask_bce_loss": 1.1786761283874512, "mask_dice_loss": 0.05777498707175255, "mask_loss": 1.2364511489868164, "step": 1249 }, { "epoch": 0.7507507507507507, "grad_norm": 51.021854400634766, "learning_rate": 3.09942319876229e-07, "loss": 0.7539944648742676, "step": 1250 }, { "ce_loss": 2.304595363966655e-05, "cls_loss": 0.03466796875, "epoch": 0.7507507507507507, "mask_bce_loss": 0.4328935742378235, "mask_dice_loss": 0.026852643117308617, "mask_loss": 0.45974621176719666, "step": 1250 }, { "epoch": 0.7513513513513513, "grad_norm": 23.008630752563477, "learning_rate": 3.085357369310612e-07, "loss": 0.78126060962677, "step": 1251 }, { "ce_loss": 7.669332990190014e-05, "cls_loss": 0.06005859375, "epoch": 0.7513513513513513, "mask_bce_loss": 0.6567699313163757, "mask_dice_loss": 0.09821423143148422, "mask_loss": 0.7549841403961182, "step": 1251 }, { "epoch": 0.751951951951952, "grad_norm": 23.14560317993164, "learning_rate": 3.071317705073141e-07, "loss": 0.7702865600585938, "step": 1252 }, { "ce_loss": 9.776234946912155e-05, "cls_loss": 0.064453125, "epoch": 0.751951951951952, "mask_bce_loss": 0.7263028025627136, "mask_dice_loss": 0.09889689832925797, "mask_loss": 0.8251997232437134, "step": 1252 }, { "epoch": 0.7525525525525526, "grad_norm": 39.274757385253906, "learning_rate": 3.057304259176384e-07, "loss": 0.8085547089576721, "step": 1253 }, { "ce_loss": 3.927783473045565e-05, "cls_loss": 0.041015625, "epoch": 0.7525525525525526, "mask_bce_loss": 0.32977214455604553, "mask_dice_loss": 0.04383091256022453, "mask_loss": 0.37360304594039917, "step": 1253 }, { "epoch": 0.7531531531531531, "grad_norm": 23.241451263427734, "learning_rate": 3.04331708464764e-07, "loss": 0.7539820671081543, "step": 1254 }, { "ce_loss": 3.712894977070391e-05, "cls_loss": 0.040283203125, "epoch": 0.7531531531531531, "mask_bce_loss": 0.5381708145141602, "mask_dice_loss": 0.053194381296634674, "mask_loss": 0.5913652181625366, "step": 1254 }, { "epoch": 0.7537537537537538, "grad_norm": 44.92750549316406, "learning_rate": 3.0293562344147917e-07, "loss": 0.9233477115631104, "step": 1255 }, { "ce_loss": 5.7509711041348055e-05, "cls_loss": 0.029541015625, "epoch": 0.7537537537537538, "mask_bce_loss": 0.22452044486999512, "mask_dice_loss": 0.04464135318994522, "mask_loss": 0.26916179060935974, "step": 1255 }, { "epoch": 0.7543543543543544, "grad_norm": 26.33859634399414, "learning_rate": 3.015421761306116e-07, "loss": 0.7928010821342468, "step": 1256 }, { "ce_loss": 4.0443581383442506e-05, "cls_loss": 0.0380859375, "epoch": 0.7543543543543544, "mask_bce_loss": 0.23166470229625702, "mask_dice_loss": 0.032495222985744476, "mask_loss": 0.2641599178314209, "step": 1256 }, { "epoch": 0.7549549549549549, "grad_norm": 45.80812072753906, "learning_rate": 3.0015137180500703e-07, "loss": 0.9905149936676025, "step": 1257 }, { "ce_loss": 0.00025587441632524133, "cls_loss": 0.05908203125, "epoch": 0.7549549549549549, "mask_bce_loss": 0.835699737071991, "mask_dice_loss": 0.18682809174060822, "mask_loss": 1.022527813911438, "step": 1257 }, { "epoch": 0.7555555555555555, "grad_norm": 45.41179275512695, "learning_rate": 2.987632157275114e-07, "loss": 0.8454571962356567, "step": 1258 }, { "ce_loss": 3.274534537922591e-05, "cls_loss": 0.057373046875, "epoch": 0.7555555555555555, "mask_bce_loss": 0.5176783800125122, "mask_dice_loss": 0.08532960712909698, "mask_loss": 0.603007972240448, "step": 1258 }, { "epoch": 0.7561561561561562, "grad_norm": 17.379304885864258, "learning_rate": 2.973777131509483e-07, "loss": 0.7752524614334106, "step": 1259 }, { "ce_loss": 0.00013905524974688888, "cls_loss": 0.046142578125, "epoch": 0.7561561561561562, "mask_bce_loss": 0.7078516483306885, "mask_dice_loss": 0.07599174231290817, "mask_loss": 0.7838433980941772, "step": 1259 }, { "epoch": 0.7567567567567568, "grad_norm": 27.380563735961914, "learning_rate": 2.959948693181006e-07, "loss": 0.840067982673645, "step": 1260 }, { "ce_loss": 0.00010090240539284423, "cls_loss": 0.059326171875, "epoch": 0.7567567567567568, "mask_bce_loss": 0.7778657674789429, "mask_dice_loss": 0.15811380743980408, "mask_loss": 0.9359796047210693, "step": 1260 }, { "epoch": 0.7573573573573573, "grad_norm": 29.402658462524414, "learning_rate": 2.9461468946169133e-07, "loss": 0.8944909572601318, "step": 1261 }, { "ce_loss": 4.069066926604137e-05, "cls_loss": 0.072265625, "epoch": 0.7573573573573573, "mask_bce_loss": 0.9623657464981079, "mask_dice_loss": 0.17811976373195648, "mask_loss": 1.1404855251312256, "step": 1261 }, { "epoch": 0.757957957957958, "grad_norm": 37.08781433105469, "learning_rate": 2.9323717880436173e-07, "loss": 0.7367217540740967, "step": 1262 }, { "ce_loss": 3.9440874388674274e-05, "cls_loss": 0.05322265625, "epoch": 0.757957957957958, "mask_bce_loss": 1.6900207996368408, "mask_dice_loss": 0.0898602232336998, "mask_loss": 1.7798810005187988, "step": 1262 }, { "epoch": 0.7585585585585586, "grad_norm": 30.568992614746094, "learning_rate": 2.918623425586536e-07, "loss": 0.7615388631820679, "step": 1263 }, { "ce_loss": 7.58034220780246e-05, "cls_loss": 0.046875, "epoch": 0.7585585585585586, "mask_bce_loss": 0.71546471118927, "mask_dice_loss": 0.05336308106780052, "mask_loss": 0.7688277959823608, "step": 1263 }, { "epoch": 0.7591591591591592, "grad_norm": 37.94512176513672, "learning_rate": 2.90490185926988e-07, "loss": 1.043683648109436, "step": 1264 }, { "ce_loss": 8.295559382531792e-05, "cls_loss": 0.0654296875, "epoch": 0.7591591591591592, "mask_bce_loss": 0.7066196799278259, "mask_dice_loss": 0.11086177825927734, "mask_loss": 0.8174814581871033, "step": 1264 }, { "epoch": 0.7597597597597597, "grad_norm": 28.827924728393555, "learning_rate": 2.8912071410164695e-07, "loss": 0.8900686502456665, "step": 1265 }, { "ce_loss": 4.512951272772625e-05, "cls_loss": 0.05224609375, "epoch": 0.7597597597597597, "mask_bce_loss": 0.3902904689311981, "mask_dice_loss": 0.1267489194869995, "mask_loss": 0.51703941822052, "step": 1265 }, { "epoch": 0.7603603603603604, "grad_norm": 28.911415100097656, "learning_rate": 2.877539322647522e-07, "loss": 0.8998241424560547, "step": 1266 }, { "ce_loss": 0.0006408373592421412, "cls_loss": 0.0439453125, "epoch": 0.7603603603603604, "mask_bce_loss": 0.8411955833435059, "mask_dice_loss": 0.07390635460615158, "mask_loss": 0.915101945400238, "step": 1266 }, { "epoch": 0.760960960960961, "grad_norm": 53.12643051147461, "learning_rate": 2.8638984558824774e-07, "loss": 0.8638435006141663, "step": 1267 }, { "ce_loss": 2.7383161068428308e-05, "cls_loss": 0.05517578125, "epoch": 0.760960960960961, "mask_bce_loss": 0.5966305136680603, "mask_dice_loss": 0.06618736684322357, "mask_loss": 0.6628178954124451, "step": 1267 }, { "epoch": 0.7615615615615615, "grad_norm": 28.958559036254883, "learning_rate": 2.8502845923387795e-07, "loss": 0.8588871955871582, "step": 1268 }, { "ce_loss": 3.2275802368531004e-05, "cls_loss": 0.0439453125, "epoch": 0.7615615615615615, "mask_bce_loss": 0.5496335029602051, "mask_dice_loss": 0.06287391483783722, "mask_loss": 0.6125074028968811, "step": 1268 }, { "epoch": 0.7621621621621621, "grad_norm": 38.77426528930664, "learning_rate": 2.836697783531692e-07, "loss": 0.7793681025505066, "step": 1269 }, { "ce_loss": 0.0001922009396366775, "cls_loss": 0.05615234375, "epoch": 0.7621621621621621, "mask_bce_loss": 0.1667192280292511, "mask_dice_loss": 0.2074209749698639, "mask_loss": 0.374140202999115, "step": 1269 }, { "epoch": 0.7627627627627628, "grad_norm": 42.290931701660156, "learning_rate": 2.8231380808741104e-07, "loss": 0.88679438829422, "step": 1270 }, { "ce_loss": 6.651192961726338e-05, "cls_loss": 0.05615234375, "epoch": 0.7627627627627628, "mask_bce_loss": 1.2348397970199585, "mask_dice_loss": 0.12539887428283691, "mask_loss": 1.3602386713027954, "step": 1270 }, { "epoch": 0.7633633633633634, "grad_norm": 43.283111572265625, "learning_rate": 2.8096055356763515e-07, "loss": 0.828153133392334, "step": 1271 }, { "ce_loss": 0.010794716887176037, "cls_loss": 0.0390625, "epoch": 0.7633633633633634, "mask_bce_loss": 0.8910835385322571, "mask_dice_loss": 0.04917234182357788, "mask_loss": 0.940255880355835, "step": 1271 }, { "epoch": 0.7639639639639639, "grad_norm": 34.78557586669922, "learning_rate": 2.796100199145971e-07, "loss": 0.8148632645606995, "step": 1272 }, { "ce_loss": 0.00015105665079317987, "cls_loss": 0.043701171875, "epoch": 0.7639639639639639, "mask_bce_loss": 0.3630998432636261, "mask_dice_loss": 0.04541891813278198, "mask_loss": 0.4085187613964081, "step": 1272 }, { "epoch": 0.7645645645645646, "grad_norm": 31.294788360595703, "learning_rate": 2.7826221223875655e-07, "loss": 0.8701642155647278, "step": 1273 }, { "ce_loss": 0.0003428676282055676, "cls_loss": 0.05517578125, "epoch": 0.7645645645645646, "mask_bce_loss": 0.7431487441062927, "mask_dice_loss": 0.09968941658735275, "mask_loss": 0.8428381681442261, "step": 1273 }, { "epoch": 0.7651651651651652, "grad_norm": 29.887025833129883, "learning_rate": 2.7691713564025797e-07, "loss": 0.9584216475486755, "step": 1274 }, { "ce_loss": 3.1804313039174303e-05, "cls_loss": 0.046875, "epoch": 0.7651651651651652, "mask_bce_loss": 0.7570773959159851, "mask_dice_loss": 0.07880252599716187, "mask_loss": 0.835879921913147, "step": 1274 }, { "epoch": 0.7657657657657657, "grad_norm": 45.88132858276367, "learning_rate": 2.7557479520891104e-07, "loss": 0.8488572835922241, "step": 1275 }, { "ce_loss": 2.4955270419013686e-05, "cls_loss": 0.0289306640625, "epoch": 0.7657657657657657, "mask_bce_loss": 0.3057369291782379, "mask_dice_loss": 0.020535670220851898, "mask_loss": 0.3262726068496704, "step": 1275 }, { "epoch": 0.7663663663663663, "grad_norm": 53.98517608642578, "learning_rate": 2.742351960241728e-07, "loss": 1.0326381921768188, "step": 1276 }, { "ce_loss": 3.716661740327254e-05, "cls_loss": 0.05908203125, "epoch": 0.7663663663663663, "mask_bce_loss": 1.127140760421753, "mask_dice_loss": 0.1155974492430687, "mask_loss": 1.2427382469177246, "step": 1276 }, { "epoch": 0.766966966966967, "grad_norm": 30.009479522705078, "learning_rate": 2.728983431551261e-07, "loss": 0.8304004669189453, "step": 1277 }, { "ce_loss": 2.465713259880431e-05, "cls_loss": 0.047119140625, "epoch": 0.766966966966967, "mask_bce_loss": 1.3862276077270508, "mask_dice_loss": 0.18608197569847107, "mask_loss": 1.5723096132278442, "step": 1277 }, { "epoch": 0.7675675675675676, "grad_norm": 15.473066329956055, "learning_rate": 2.715642416604619e-07, "loss": 0.7533632516860962, "step": 1278 }, { "ce_loss": 0.0009081661701202393, "cls_loss": 0.05322265625, "epoch": 0.7675675675675676, "mask_bce_loss": 1.306541085243225, "mask_dice_loss": 0.13503201305866241, "mask_loss": 1.441573143005371, "step": 1278 }, { "epoch": 0.7681681681681681, "grad_norm": 27.096715927124023, "learning_rate": 2.702328965884606e-07, "loss": 0.7330906391143799, "step": 1279 }, { "ce_loss": 3.69374974980019e-05, "cls_loss": 0.055908203125, "epoch": 0.7681681681681681, "mask_bce_loss": 0.3450259268283844, "mask_dice_loss": 0.10568827390670776, "mask_loss": 0.45071420073509216, "step": 1279 }, { "epoch": 0.7687687687687688, "grad_norm": 31.73426628112793, "learning_rate": 2.6890431297697156e-07, "loss": 0.7173330783843994, "step": 1280 }, { "ce_loss": 6.530023529194295e-05, "cls_loss": 0.06103515625, "epoch": 0.7687687687687688, "mask_bce_loss": 1.633423089981079, "mask_dice_loss": 0.09246431291103363, "mask_loss": 1.725887417793274, "step": 1280 }, { "epoch": 0.7693693693693694, "grad_norm": 24.077178955078125, "learning_rate": 2.6757849585339476e-07, "loss": 0.8039816617965698, "step": 1281 }, { "ce_loss": 0.0015294532058760524, "cls_loss": 0.064453125, "epoch": 0.7693693693693694, "mask_bce_loss": 0.8267555236816406, "mask_dice_loss": 0.09186365455389023, "mask_loss": 0.9186191558837891, "step": 1281 }, { "epoch": 0.76996996996997, "grad_norm": 26.380996704101562, "learning_rate": 2.6625545023466176e-07, "loss": 0.6862778663635254, "step": 1282 }, { "ce_loss": 2.5090592316701077e-05, "cls_loss": 0.06396484375, "epoch": 0.76996996996997, "mask_bce_loss": 1.3703619241714478, "mask_dice_loss": 0.08834884315729141, "mask_loss": 1.458710789680481, "step": 1282 }, { "epoch": 0.7705705705705705, "grad_norm": 18.672706604003906, "learning_rate": 2.649351811272167e-07, "loss": 0.778668999671936, "step": 1283 }, { "ce_loss": 4.402458216645755e-05, "cls_loss": 0.059814453125, "epoch": 0.7705705705705705, "mask_bce_loss": 0.5124253630638123, "mask_dice_loss": 0.09174885600805283, "mask_loss": 0.6041741967201233, "step": 1283 }, { "epoch": 0.7711711711711712, "grad_norm": 28.176498413085938, "learning_rate": 2.6361769352699703e-07, "loss": 0.8047678470611572, "step": 1284 }, { "ce_loss": 0.00014207033382263035, "cls_loss": 0.049560546875, "epoch": 0.7711711711711712, "mask_bce_loss": 0.9507009387016296, "mask_dice_loss": 0.10861524194478989, "mask_loss": 1.0593161582946777, "step": 1284 }, { "epoch": 0.7717717717717718, "grad_norm": 26.47753143310547, "learning_rate": 2.623029924194159e-07, "loss": 0.9532680511474609, "step": 1285 }, { "ce_loss": 3.2230880606221035e-05, "cls_loss": 0.05859375, "epoch": 0.7717717717717718, "mask_bce_loss": 1.1554449796676636, "mask_dice_loss": 0.12426748126745224, "mask_loss": 1.279712438583374, "step": 1285 }, { "epoch": 0.7723723723723723, "grad_norm": 19.040449142456055, "learning_rate": 2.60991082779341e-07, "loss": 0.8193613290786743, "step": 1286 }, { "ce_loss": 0.0006062669563107193, "cls_loss": 0.06103515625, "epoch": 0.7723723723723723, "mask_bce_loss": 1.2436299324035645, "mask_dice_loss": 0.1149054542183876, "mask_loss": 1.3585354089736938, "step": 1286 }, { "epoch": 0.772972972972973, "grad_norm": 77.45411682128906, "learning_rate": 2.5968196957107745e-07, "loss": 0.8331446051597595, "step": 1287 }, { "ce_loss": 6.605400994885713e-05, "cls_loss": 0.05908203125, "epoch": 0.772972972972973, "mask_bce_loss": 0.7294057011604309, "mask_dice_loss": 0.1313198059797287, "mask_loss": 0.8607255220413208, "step": 1287 }, { "epoch": 0.7735735735735736, "grad_norm": 32.34040451049805, "learning_rate": 2.5837565774834936e-07, "loss": 0.812321662902832, "step": 1288 }, { "ce_loss": 0.0004200583789497614, "cls_loss": 0.0279541015625, "epoch": 0.7735735735735736, "mask_bce_loss": 0.18327023088932037, "mask_dice_loss": 0.019344842061400414, "mask_loss": 0.20261506736278534, "step": 1288 }, { "epoch": 0.7741741741741742, "grad_norm": 46.198394775390625, "learning_rate": 2.570721522542793e-07, "loss": 0.9347101449966431, "step": 1289 }, { "ce_loss": 0.00021897899569012225, "cls_loss": 0.09375, "epoch": 0.7741741741741742, "mask_bce_loss": 1.9771595001220703, "mask_dice_loss": 0.13525210320949554, "mask_loss": 2.1124114990234375, "step": 1289 }, { "epoch": 0.7747747747747747, "grad_norm": 20.645334243774414, "learning_rate": 2.5577145802137103e-07, "loss": 0.8594854474067688, "step": 1290 }, { "ce_loss": 0.0007320301956497133, "cls_loss": 0.038330078125, "epoch": 0.7747747747747747, "mask_bce_loss": 0.4235458970069885, "mask_dice_loss": 0.05041028559207916, "mask_loss": 0.4739561676979065, "step": 1290 }, { "epoch": 0.7753753753753754, "grad_norm": 21.695642471313477, "learning_rate": 2.5447357997149024e-07, "loss": 0.819342315196991, "step": 1291 }, { "ce_loss": 0.00021233363077044487, "cls_loss": 0.0478515625, "epoch": 0.7753753753753754, "mask_bce_loss": 0.5605099201202393, "mask_dice_loss": 0.049175556749105453, "mask_loss": 0.609685480594635, "step": 1291 }, { "epoch": 0.775975975975976, "grad_norm": 21.22710609436035, "learning_rate": 2.5317852301584643e-07, "loss": 0.8726019859313965, "step": 1292 }, { "ce_loss": 7.575331983389333e-05, "cls_loss": 0.0732421875, "epoch": 0.775975975975976, "mask_bce_loss": 0.4421882629394531, "mask_dice_loss": 0.06639339029788971, "mask_loss": 0.5085816383361816, "step": 1292 }, { "epoch": 0.7765765765765765, "grad_norm": 45.57679748535156, "learning_rate": 2.5188629205497335e-07, "loss": 0.8935103416442871, "step": 1293 }, { "ce_loss": 0.0001429747644579038, "cls_loss": 0.059326171875, "epoch": 0.7765765765765765, "mask_bce_loss": 1.2569983005523682, "mask_dice_loss": 0.12110903114080429, "mask_loss": 1.3781073093414307, "step": 1293 }, { "epoch": 0.7771771771771772, "grad_norm": 25.561016082763672, "learning_rate": 2.5059689197871214e-07, "loss": 0.8080664873123169, "step": 1294 }, { "ce_loss": 6.449145439546555e-05, "cls_loss": 0.05419921875, "epoch": 0.7771771771771772, "mask_bce_loss": 0.6445444822311401, "mask_dice_loss": 0.06428717821836472, "mask_loss": 0.7088316679000854, "step": 1294 }, { "epoch": 0.7777777777777778, "grad_norm": 14.032647132873535, "learning_rate": 2.4931032766619096e-07, "loss": 0.6833379864692688, "step": 1295 }, { "ce_loss": 8.483052079100162e-05, "cls_loss": 0.05859375, "epoch": 0.7777777777777778, "mask_bce_loss": 0.7412015199661255, "mask_dice_loss": 0.16092266142368317, "mask_loss": 0.9021241664886475, "step": 1295 }, { "epoch": 0.7783783783783784, "grad_norm": 29.72516441345215, "learning_rate": 2.480266039858073e-07, "loss": 0.7063052654266357, "step": 1296 }, { "ce_loss": 5.453278936329298e-05, "cls_loss": 0.0546875, "epoch": 0.7783783783783784, "mask_bce_loss": 0.3822116553783417, "mask_dice_loss": 0.10219340771436691, "mask_loss": 0.4844050705432892, "step": 1296 }, { "epoch": 0.7789789789789789, "grad_norm": 28.92209815979004, "learning_rate": 2.467457257952105e-07, "loss": 0.9003547430038452, "step": 1297 }, { "ce_loss": 0.0007517910562455654, "cls_loss": 0.052001953125, "epoch": 0.7789789789789789, "mask_bce_loss": 0.34419041872024536, "mask_dice_loss": 0.05047475919127464, "mask_loss": 0.3946651816368103, "step": 1297 }, { "epoch": 0.7795795795795796, "grad_norm": 22.690027236938477, "learning_rate": 2.454676979412815e-07, "loss": 0.6597261428833008, "step": 1298 }, { "ce_loss": 9.564942592987791e-05, "cls_loss": 0.06396484375, "epoch": 0.7795795795795796, "mask_bce_loss": 0.9428636431694031, "mask_dice_loss": 0.13460172712802887, "mask_loss": 1.0774654150009155, "step": 1298 }, { "epoch": 0.7801801801801802, "grad_norm": 30.683395385742188, "learning_rate": 2.44192525260116e-07, "loss": 0.885648250579834, "step": 1299 }, { "ce_loss": 7.018266478553414e-05, "cls_loss": 0.04052734375, "epoch": 0.7801801801801802, "mask_bce_loss": 0.46138763427734375, "mask_dice_loss": 0.059280045330524445, "mask_loss": 0.5206676721572876, "step": 1299 }, { "epoch": 0.7807807807807807, "grad_norm": 15.308919906616211, "learning_rate": 2.4292021257700534e-07, "loss": 0.8083574771881104, "step": 1300 }, { "ce_loss": 0.00010656884842319414, "cls_loss": 0.036865234375, "epoch": 0.7807807807807807, "mask_bce_loss": 0.4438149631023407, "mask_dice_loss": 0.04580472782254219, "mask_loss": 0.4896197021007538, "step": 1300 }, { "epoch": 0.7813813813813814, "grad_norm": 25.35063362121582, "learning_rate": 2.4165076470641856e-07, "loss": 0.9106533527374268, "step": 1301 }, { "ce_loss": 6.32364462944679e-05, "cls_loss": 0.06005859375, "epoch": 0.7813813813813814, "mask_bce_loss": 0.41119304299354553, "mask_dice_loss": 0.10255800932645798, "mask_loss": 0.5137510299682617, "step": 1301 }, { "epoch": 0.781981981981982, "grad_norm": 26.327552795410156, "learning_rate": 2.4038418645198475e-07, "loss": 0.8339748382568359, "step": 1302 }, { "ce_loss": 3.85142229788471e-05, "cls_loss": 0.0634765625, "epoch": 0.781981981981982, "mask_bce_loss": 1.5921369791030884, "mask_dice_loss": 0.1562267690896988, "mask_loss": 1.748363733291626, "step": 1302 }, { "epoch": 0.7825825825825826, "grad_norm": 20.293476104736328, "learning_rate": 2.391204826064738e-07, "loss": 0.7688820958137512, "step": 1303 }, { "ce_loss": 0.007448554039001465, "cls_loss": 0.03662109375, "epoch": 0.7825825825825826, "mask_bce_loss": 0.4081604480743408, "mask_dice_loss": 0.10808243602514267, "mask_loss": 0.5162428617477417, "step": 1303 }, { "epoch": 0.7831831831831831, "grad_norm": 26.352689743041992, "learning_rate": 2.378596579517782e-07, "loss": 0.8629849553108215, "step": 1304 }, { "ce_loss": 2.9751701731584035e-05, "cls_loss": 0.06005859375, "epoch": 0.7831831831831831, "mask_bce_loss": 0.25000080466270447, "mask_dice_loss": 0.1349247694015503, "mask_loss": 0.38492557406425476, "step": 1304 }, { "epoch": 0.7837837837837838, "grad_norm": 15.887370109558105, "learning_rate": 2.36601717258897e-07, "loss": 0.7278624773025513, "step": 1305 }, { "ce_loss": 0.00015935616102069616, "cls_loss": 0.04248046875, "epoch": 0.7837837837837838, "mask_bce_loss": 0.3784044086933136, "mask_dice_loss": 0.04088642820715904, "mask_loss": 0.41929084062576294, "step": 1305 }, { "epoch": 0.7843843843843844, "grad_norm": 16.25271224975586, "learning_rate": 2.353466652879147e-07, "loss": 0.7065367698669434, "step": 1306 }, { "ce_loss": 8.719165634829551e-05, "cls_loss": 0.057861328125, "epoch": 0.7843843843843844, "mask_bce_loss": 0.7464329600334167, "mask_dice_loss": 0.10072594881057739, "mask_loss": 0.8471589088439941, "step": 1306 }, { "epoch": 0.784984984984985, "grad_norm": 55.46712112426758, "learning_rate": 2.340945067879857e-07, "loss": 0.8425866365432739, "step": 1307 }, { "ce_loss": 4.2483603465370834e-05, "cls_loss": 0.043212890625, "epoch": 0.784984984984985, "mask_bce_loss": 0.2482621967792511, "mask_dice_loss": 0.03915272280573845, "mask_loss": 0.28741490840911865, "step": 1307 }, { "epoch": 0.7855855855855856, "grad_norm": 30.375944137573242, "learning_rate": 2.3284524649731518e-07, "loss": 0.7338647842407227, "step": 1308 }, { "ce_loss": 0.00012347706069704145, "cls_loss": 0.05322265625, "epoch": 0.7855855855855856, "mask_bce_loss": 1.3831520080566406, "mask_dice_loss": 0.09961006790399551, "mask_loss": 1.482762098312378, "step": 1308 }, { "epoch": 0.7861861861861862, "grad_norm": 20.853883743286133, "learning_rate": 2.3159888914314119e-07, "loss": 0.8484901189804077, "step": 1309 }, { "ce_loss": 4.423140126164071e-05, "cls_loss": 0.048828125, "epoch": 0.7861861861861862, "mask_bce_loss": 0.6460970044136047, "mask_dice_loss": 0.04861369729042053, "mask_loss": 0.6947107315063477, "step": 1309 }, { "epoch": 0.7867867867867868, "grad_norm": 61.55032730102539, "learning_rate": 2.30355439441717e-07, "loss": 0.9131024479866028, "step": 1310 }, { "ce_loss": 2.0155574020463973e-05, "cls_loss": 0.048583984375, "epoch": 0.7867867867867868, "mask_bce_loss": 1.3032009601593018, "mask_dice_loss": 0.08358240127563477, "mask_loss": 1.3867833614349365, "step": 1310 }, { "epoch": 0.7873873873873873, "grad_norm": 28.80544662475586, "learning_rate": 2.2911490209829398e-07, "loss": 0.7953700423240662, "step": 1311 }, { "ce_loss": 0.00023369796690531075, "cls_loss": 0.053955078125, "epoch": 0.7873873873873873, "mask_bce_loss": 1.3334643840789795, "mask_dice_loss": 0.08851806074380875, "mask_loss": 1.4219824075698853, "step": 1311 }, { "epoch": 0.787987987987988, "grad_norm": 31.67040252685547, "learning_rate": 2.2787728180710218e-07, "loss": 0.7356173992156982, "step": 1312 }, { "ce_loss": 4.7382593038491905e-05, "cls_loss": 0.055419921875, "epoch": 0.787987987987988, "mask_bce_loss": 0.9625974893569946, "mask_dice_loss": 0.09160535782575607, "mask_loss": 1.0542027950286865, "step": 1312 }, { "epoch": 0.7885885885885886, "grad_norm": 21.688594818115234, "learning_rate": 2.2664258325133347e-07, "loss": 0.7506338357925415, "step": 1313 }, { "ce_loss": 3.4745604352792725e-05, "cls_loss": 0.05908203125, "epoch": 0.7885885885885886, "mask_bce_loss": 0.6663147807121277, "mask_dice_loss": 0.08816394209861755, "mask_loss": 0.7544786930084229, "step": 1313 }, { "epoch": 0.7891891891891892, "grad_norm": 41.31745910644531, "learning_rate": 2.2541081110312466e-07, "loss": 0.7944984436035156, "step": 1314 }, { "ce_loss": 0.0014218693831935525, "cls_loss": 0.023193359375, "epoch": 0.7891891891891892, "mask_bce_loss": 0.37307435274124146, "mask_dice_loss": 0.036862146109342575, "mask_loss": 0.40993648767471313, "step": 1314 }, { "epoch": 0.7897897897897898, "grad_norm": 41.12871170043945, "learning_rate": 2.241819700235379e-07, "loss": 0.8527882099151611, "step": 1315 }, { "ce_loss": 4.153543704887852e-05, "cls_loss": 0.04296875, "epoch": 0.7897897897897898, "mask_bce_loss": 0.3269607722759247, "mask_dice_loss": 0.04512527957558632, "mask_loss": 0.3720860481262207, "step": 1315 }, { "epoch": 0.7903903903903904, "grad_norm": 21.175222396850586, "learning_rate": 2.229560646625448e-07, "loss": 0.6496859192848206, "step": 1316 }, { "ce_loss": 0.009951145388185978, "cls_loss": 0.05810546875, "epoch": 0.7903903903903904, "mask_bce_loss": 0.5393615961074829, "mask_dice_loss": 0.11883914470672607, "mask_loss": 0.658200740814209, "step": 1316 }, { "epoch": 0.790990990990991, "grad_norm": 32.83853530883789, "learning_rate": 2.2173309965900765e-07, "loss": 0.7603904008865356, "step": 1317 }, { "ce_loss": 0.0005956952227279544, "cls_loss": 0.0400390625, "epoch": 0.790990990990991, "mask_bce_loss": 0.3726719915866852, "mask_dice_loss": 0.04969480633735657, "mask_loss": 0.42236679792404175, "step": 1317 }, { "epoch": 0.7915915915915915, "grad_norm": 36.21041488647461, "learning_rate": 2.2051307964066256e-07, "loss": 0.9419215321540833, "step": 1318 }, { "ce_loss": 2.7452797439764254e-05, "cls_loss": 0.04150390625, "epoch": 0.7915915915915915, "mask_bce_loss": 0.40404948592185974, "mask_dice_loss": 0.03875763341784477, "mask_loss": 0.4428071081638336, "step": 1318 }, { "epoch": 0.7921921921921922, "grad_norm": 46.03791809082031, "learning_rate": 2.1929600922410153e-07, "loss": 0.7949113845825195, "step": 1319 }, { "ce_loss": 3.9031790947774425e-05, "cls_loss": 0.02734375, "epoch": 0.7921921921921922, "mask_bce_loss": 0.22845745086669922, "mask_dice_loss": 0.018872562795877457, "mask_loss": 0.24733000993728638, "step": 1319 }, { "epoch": 0.7927927927927928, "grad_norm": 37.353763580322266, "learning_rate": 2.1808189301475589e-07, "loss": 0.8218072056770325, "step": 1320 }, { "ce_loss": 0.00012750033056363463, "cls_loss": 0.05712890625, "epoch": 0.7927927927927928, "mask_bce_loss": 0.635560929775238, "mask_dice_loss": 0.10662000626325607, "mask_loss": 0.7421809434890747, "step": 1320 }, { "epoch": 0.7933933933933934, "grad_norm": 27.748735427856445, "learning_rate": 2.168707356068773e-07, "loss": 0.7965947389602661, "step": 1321 }, { "ce_loss": 3.9957831177162006e-05, "cls_loss": 0.05859375, "epoch": 0.7933933933933934, "mask_bce_loss": 1.0920135974884033, "mask_dice_loss": 0.08585577458143234, "mask_loss": 1.1778693199157715, "step": 1321 }, { "epoch": 0.793993993993994, "grad_norm": 35.3870964050293, "learning_rate": 2.1566254158352125e-07, "loss": 0.952879786491394, "step": 1322 }, { "ce_loss": 0.0006159862969070673, "cls_loss": 0.0986328125, "epoch": 0.793993993993994, "mask_bce_loss": 0.7992674708366394, "mask_dice_loss": 0.08540980517864227, "mask_loss": 0.8846772909164429, "step": 1322 }, { "epoch": 0.7945945945945946, "grad_norm": 49.58879470825195, "learning_rate": 2.1445731551653067e-07, "loss": 0.874373197555542, "step": 1323 }, { "ce_loss": 0.00010125203698407859, "cls_loss": 0.06103515625, "epoch": 0.7945945945945946, "mask_bce_loss": 0.8777078986167908, "mask_dice_loss": 0.14084628224372864, "mask_loss": 1.0185542106628418, "step": 1323 }, { "epoch": 0.7951951951951952, "grad_norm": 32.000308990478516, "learning_rate": 2.1325506196651676e-07, "loss": 0.7982219457626343, "step": 1324 }, { "ce_loss": 0.00021433670190162957, "cls_loss": 0.055908203125, "epoch": 0.7951951951951952, "mask_bce_loss": 1.2352921962738037, "mask_dice_loss": 0.086388498544693, "mask_loss": 1.3216806650161743, "step": 1324 }, { "epoch": 0.7957957957957958, "grad_norm": 21.638879776000977, "learning_rate": 2.1205578548284265e-07, "loss": 0.8339399695396423, "step": 1325 }, { "ce_loss": 6.585996743524447e-05, "cls_loss": 0.04443359375, "epoch": 0.7957957957957958, "mask_bce_loss": 0.25672653317451477, "mask_dice_loss": 0.06799288094043732, "mask_loss": 0.3247194290161133, "step": 1325 }, { "epoch": 0.7963963963963964, "grad_norm": 15.2249755859375, "learning_rate": 2.1085949060360653e-07, "loss": 0.7095605134963989, "step": 1326 }, { "ce_loss": 3.097701846854761e-05, "cls_loss": 0.05126953125, "epoch": 0.7963963963963964, "mask_bce_loss": 0.6114835143089294, "mask_dice_loss": 0.05357605218887329, "mask_loss": 0.6650595664978027, "step": 1326 }, { "epoch": 0.796996996996997, "grad_norm": 14.739911079406738, "learning_rate": 2.0966618185562378e-07, "loss": 0.7703850865364075, "step": 1327 }, { "ce_loss": 0.03211794048547745, "cls_loss": 0.055419921875, "epoch": 0.796996996996997, "mask_bce_loss": 0.6026660203933716, "mask_dice_loss": 0.09369838982820511, "mask_loss": 0.6963644027709961, "step": 1327 }, { "epoch": 0.7975975975975976, "grad_norm": 52.35066604614258, "learning_rate": 2.084758637544103e-07, "loss": 0.8149200677871704, "step": 1328 }, { "ce_loss": 3.775313598453067e-05, "cls_loss": 0.035400390625, "epoch": 0.7975975975975976, "mask_bce_loss": 0.3854382038116455, "mask_dice_loss": 0.05038977414369583, "mask_loss": 0.43582797050476074, "step": 1328 }, { "epoch": 0.7981981981981981, "grad_norm": 39.527427673339844, "learning_rate": 2.072885408041658e-07, "loss": 0.8516405820846558, "step": 1329 }, { "ce_loss": 3.811206261161715e-05, "cls_loss": 0.068359375, "epoch": 0.7981981981981981, "mask_bce_loss": 0.4422364830970764, "mask_dice_loss": 0.11425594240427017, "mask_loss": 0.5564924478530884, "step": 1329 }, { "epoch": 0.7987987987987988, "grad_norm": 41.32123947143555, "learning_rate": 2.061042174977553e-07, "loss": 0.9061073064804077, "step": 1330 }, { "ce_loss": 2.9800545235048048e-05, "cls_loss": 0.04541015625, "epoch": 0.7987987987987988, "mask_bce_loss": 0.6604669094085693, "mask_dice_loss": 0.05728263780474663, "mask_loss": 0.7177495360374451, "step": 1330 }, { "epoch": 0.7993993993993994, "grad_norm": 31.408447265625, "learning_rate": 2.0492289831669362e-07, "loss": 0.9456274509429932, "step": 1331 }, { "ce_loss": 0.00015685406106058508, "cls_loss": 0.059814453125, "epoch": 0.7993993993993994, "mask_bce_loss": 0.9501320123672485, "mask_dice_loss": 0.08040403574705124, "mask_loss": 1.0305360555648804, "step": 1331 }, { "epoch": 0.8, "grad_norm": 24.038854598999023, "learning_rate": 2.0374458773112812e-07, "loss": 0.8758135437965393, "step": 1332 }, { "ce_loss": 6.500327435787767e-05, "cls_loss": 0.10400390625, "epoch": 0.8, "mask_bce_loss": 0.8296612501144409, "mask_dice_loss": 0.122315913438797, "mask_loss": 0.9519771337509155, "step": 1332 }, { "epoch": 0.8006006006006006, "grad_norm": 24.314210891723633, "learning_rate": 2.0256929019982104e-07, "loss": 0.736485481262207, "step": 1333 }, { "ce_loss": 0.0006247287965379655, "cls_loss": 0.03955078125, "epoch": 0.8006006006006006, "mask_bce_loss": 0.7127141356468201, "mask_dice_loss": 0.05219031125307083, "mask_loss": 0.7649044394493103, "step": 1333 }, { "epoch": 0.8012012012012012, "grad_norm": 42.273529052734375, "learning_rate": 2.0139701017013334e-07, "loss": 0.9008345007896423, "step": 1334 }, { "ce_loss": 3.9687674870947376e-05, "cls_loss": 0.046875, "epoch": 0.8012012012012012, "mask_bce_loss": 0.6152420043945312, "mask_dice_loss": 0.06536999344825745, "mask_loss": 0.6806119680404663, "step": 1334 }, { "epoch": 0.8018018018018018, "grad_norm": 24.620574951171875, "learning_rate": 2.002277520780077e-07, "loss": 0.7561342120170593, "step": 1335 }, { "ce_loss": 4.693863229476847e-05, "cls_loss": 0.049072265625, "epoch": 0.8018018018018018, "mask_bce_loss": 0.7508983016014099, "mask_dice_loss": 0.05716588720679283, "mask_loss": 0.8080641627311707, "step": 1335 }, { "epoch": 0.8024024024024023, "grad_norm": 16.284021377563477, "learning_rate": 1.9906152034795154e-07, "loss": 0.6921475529670715, "step": 1336 }, { "ce_loss": 2.1070241928100586e-05, "cls_loss": 0.027587890625, "epoch": 0.8024024024024023, "mask_bce_loss": 0.2663929760456085, "mask_dice_loss": 0.02027474157512188, "mask_loss": 0.28666770458221436, "step": 1336 }, { "epoch": 0.803003003003003, "grad_norm": 45.16517639160156, "learning_rate": 1.9789831939302038e-07, "loss": 1.139931559562683, "step": 1337 }, { "ce_loss": 5.5666369007667527e-05, "cls_loss": 0.0625, "epoch": 0.803003003003003, "mask_bce_loss": 0.7905518412590027, "mask_dice_loss": 0.1592988520860672, "mask_loss": 0.9498506784439087, "step": 1337 }, { "epoch": 0.8036036036036036, "grad_norm": 46.885379791259766, "learning_rate": 1.9673815361480195e-07, "loss": 0.8907042741775513, "step": 1338 }, { "ce_loss": 3.223575185984373e-05, "cls_loss": 0.028564453125, "epoch": 0.8036036036036036, "mask_bce_loss": 0.23004746437072754, "mask_dice_loss": 0.021204423159360886, "mask_loss": 0.25125187635421753, "step": 1338 }, { "epoch": 0.8042042042042042, "grad_norm": 56.17268371582031, "learning_rate": 1.9558102740339778e-07, "loss": 0.8105674386024475, "step": 1339 }, { "ce_loss": 9.885782492347062e-05, "cls_loss": 0.055908203125, "epoch": 0.8042042042042042, "mask_bce_loss": 0.5502247214317322, "mask_dice_loss": 0.10038655251264572, "mask_loss": 0.6506112813949585, "step": 1339 }, { "epoch": 0.8048048048048048, "grad_norm": 19.51468276977539, "learning_rate": 1.9442694513740797e-07, "loss": 0.764937162399292, "step": 1340 }, { "ce_loss": 8.523873111698776e-05, "cls_loss": 0.056640625, "epoch": 0.8048048048048048, "mask_bce_loss": 0.9740256667137146, "mask_dice_loss": 0.09566324204206467, "mask_loss": 1.0696889162063599, "step": 1340 }, { "epoch": 0.8054054054054054, "grad_norm": 21.183347702026367, "learning_rate": 1.932759111839146e-07, "loss": 0.848369836807251, "step": 1341 }, { "ce_loss": 6.988299719523638e-05, "cls_loss": 0.035888671875, "epoch": 0.8054054054054054, "mask_bce_loss": 0.774939239025116, "mask_dice_loss": 0.032304465770721436, "mask_loss": 0.8072437047958374, "step": 1341 }, { "epoch": 0.806006006006006, "grad_norm": 41.539100646972656, "learning_rate": 1.9212792989846428e-07, "loss": 0.9815176725387573, "step": 1342 }, { "ce_loss": 0.000299265404464677, "cls_loss": 0.025146484375, "epoch": 0.806006006006006, "mask_bce_loss": 0.27686381340026855, "mask_dice_loss": 0.017909279093146324, "mask_loss": 0.2947731018066406, "step": 1342 }, { "epoch": 0.8066066066066067, "grad_norm": 28.53610610961914, "learning_rate": 1.9098300562505264e-07, "loss": 1.0315971374511719, "step": 1343 }, { "ce_loss": 2.8318654585746117e-05, "cls_loss": 0.02490234375, "epoch": 0.8066066066066067, "mask_bce_loss": 0.44806256890296936, "mask_dice_loss": 0.05636715888977051, "mask_loss": 0.5044296979904175, "step": 1343 }, { "epoch": 0.8072072072072072, "grad_norm": 19.979873657226562, "learning_rate": 1.8984114269610719e-07, "loss": 0.6874842643737793, "step": 1344 }, { "ce_loss": 5.226022767601535e-05, "cls_loss": 0.04638671875, "epoch": 0.8072072072072072, "mask_bce_loss": 0.34311649203300476, "mask_dice_loss": 0.044056668877601624, "mask_loss": 0.3871731758117676, "step": 1344 }, { "epoch": 0.8078078078078078, "grad_norm": 18.371593475341797, "learning_rate": 1.8870234543247133e-07, "loss": 0.8285462856292725, "step": 1345 }, { "ce_loss": 0.017137285321950912, "cls_loss": 0.0654296875, "epoch": 0.8078078078078078, "mask_bce_loss": 2.2232935428619385, "mask_dice_loss": 0.11218108981847763, "mask_loss": 2.335474729537964, "step": 1345 }, { "epoch": 0.8084084084084084, "grad_norm": 26.02431869506836, "learning_rate": 1.875666181433877e-07, "loss": 0.8301670551300049, "step": 1346 }, { "ce_loss": 0.0001577291841385886, "cls_loss": 0.04052734375, "epoch": 0.8084084084084084, "mask_bce_loss": 0.33257099986076355, "mask_dice_loss": 0.03308189660310745, "mask_loss": 0.3656528890132904, "step": 1346 }, { "epoch": 0.809009009009009, "grad_norm": 37.97527313232422, "learning_rate": 1.864339651264828e-07, "loss": 0.7939654588699341, "step": 1347 }, { "ce_loss": 4.041820648126304e-05, "cls_loss": 0.03173828125, "epoch": 0.809009009009009, "mask_bce_loss": 0.21879224479198456, "mask_dice_loss": 0.024563096463680267, "mask_loss": 0.24335533380508423, "step": 1347 }, { "epoch": 0.8096096096096096, "grad_norm": 33.985511779785156, "learning_rate": 1.8530439066774917e-07, "loss": 0.7954474687576294, "step": 1348 }, { "ce_loss": 6.21064828010276e-05, "cls_loss": 0.057373046875, "epoch": 0.8096096096096096, "mask_bce_loss": 0.8485890626907349, "mask_dice_loss": 0.12866424024105072, "mask_loss": 0.9772533178329468, "step": 1348 }, { "epoch": 0.8102102102102102, "grad_norm": 26.425521850585938, "learning_rate": 1.8417789904153013e-07, "loss": 0.8021780252456665, "step": 1349 }, { "ce_loss": 6.449478678405285e-05, "cls_loss": 0.031494140625, "epoch": 0.8102102102102102, "mask_bce_loss": 0.3082263171672821, "mask_dice_loss": 0.046617720276117325, "mask_loss": 0.35484403371810913, "step": 1349 }, { "epoch": 0.8108108108108109, "grad_norm": 17.262155532836914, "learning_rate": 1.8305449451050404e-07, "loss": 0.7402417659759521, "step": 1350 }, { "ce_loss": 3.770422699744813e-05, "cls_loss": 0.05908203125, "epoch": 0.8108108108108109, "mask_bce_loss": 0.4255710244178772, "mask_dice_loss": 0.08242442458868027, "mask_loss": 0.5079954266548157, "step": 1350 }, { "epoch": 0.8114114114114114, "grad_norm": 83.9090347290039, "learning_rate": 1.8193418132566706e-07, "loss": 0.9993823170661926, "step": 1351 }, { "ce_loss": 5.317969043971971e-05, "cls_loss": 0.04638671875, "epoch": 0.8114114114114114, "mask_bce_loss": 0.3835972547531128, "mask_dice_loss": 0.048883743584156036, "mask_loss": 0.43248099088668823, "step": 1351 }, { "epoch": 0.812012012012012, "grad_norm": 28.560020446777344, "learning_rate": 1.8081696372631762e-07, "loss": 0.8754957914352417, "step": 1352 }, { "ce_loss": 5.268244058242999e-05, "cls_loss": 0.0361328125, "epoch": 0.812012012012012, "mask_bce_loss": 0.5388504862785339, "mask_dice_loss": 0.05367743968963623, "mask_loss": 0.5925279259681702, "step": 1352 }, { "epoch": 0.8126126126126126, "grad_norm": 27.397754669189453, "learning_rate": 1.7970284594004047e-07, "loss": 0.8770353198051453, "step": 1353 }, { "ce_loss": 8.180273289326578e-05, "cls_loss": 0.037353515625, "epoch": 0.8126126126126126, "mask_bce_loss": 0.764471173286438, "mask_dice_loss": 0.06800954788923264, "mask_loss": 0.8324807286262512, "step": 1353 }, { "epoch": 0.8132132132132132, "grad_norm": 29.589889526367188, "learning_rate": 1.785918321826907e-07, "loss": 0.8863227367401123, "step": 1354 }, { "ce_loss": 0.00015742433606646955, "cls_loss": 0.0546875, "epoch": 0.8132132132132132, "mask_bce_loss": 1.1867767572402954, "mask_dice_loss": 0.08376311510801315, "mask_loss": 1.2705398797988892, "step": 1354 }, { "epoch": 0.8138138138138138, "grad_norm": 16.25564956665039, "learning_rate": 1.774839266583772e-07, "loss": 0.7157902717590332, "step": 1355 }, { "ce_loss": 7.452830323018134e-05, "cls_loss": 0.051513671875, "epoch": 0.8138138138138138, "mask_bce_loss": 1.0317597389221191, "mask_dice_loss": 0.09618791192770004, "mask_loss": 1.1279476881027222, "step": 1355 }, { "epoch": 0.8144144144144144, "grad_norm": 33.7740478515625, "learning_rate": 1.763791335594481e-07, "loss": 0.6816951632499695, "step": 1356 }, { "ce_loss": 8.275932486867532e-05, "cls_loss": 0.05615234375, "epoch": 0.8144144144144144, "mask_bce_loss": 1.2073158025741577, "mask_dice_loss": 0.11328647285699844, "mask_loss": 1.320602297782898, "step": 1356 }, { "epoch": 0.815015015015015, "grad_norm": 15.481776237487793, "learning_rate": 1.7527745706647325e-07, "loss": 0.7380398511886597, "step": 1357 }, { "ce_loss": 3.0553208489436656e-05, "cls_loss": 0.0625, "epoch": 0.815015015015015, "mask_bce_loss": 1.0045548677444458, "mask_dice_loss": 0.09998597204685211, "mask_loss": 1.1045408248901367, "step": 1357 }, { "epoch": 0.8156156156156156, "grad_norm": 28.133028030395508, "learning_rate": 1.7417890134822911e-07, "loss": 0.7301615476608276, "step": 1358 }, { "ce_loss": 0.00012478639837354422, "cls_loss": 0.03759765625, "epoch": 0.8156156156156156, "mask_bce_loss": 0.5077187418937683, "mask_dice_loss": 0.04557123780250549, "mask_loss": 0.5532900094985962, "step": 1358 }, { "epoch": 0.8162162162162162, "grad_norm": 20.809314727783203, "learning_rate": 1.7308347056168383e-07, "loss": 0.851690411567688, "step": 1359 }, { "ce_loss": 8.725056977709755e-05, "cls_loss": 0.049072265625, "epoch": 0.8162162162162162, "mask_bce_loss": 0.7749024033546448, "mask_dice_loss": 0.053497642278671265, "mask_loss": 0.8284000158309937, "step": 1359 }, { "epoch": 0.8168168168168168, "grad_norm": 24.19793128967285, "learning_rate": 1.7199116885197996e-07, "loss": 0.7377110719680786, "step": 1360 }, { "ce_loss": 0.0024704127572476864, "cls_loss": 0.052001953125, "epoch": 0.8168168168168168, "mask_bce_loss": 0.7744531631469727, "mask_dice_loss": 0.0776398703455925, "mask_loss": 0.8520930409431458, "step": 1360 }, { "epoch": 0.8174174174174175, "grad_norm": 19.847305297851562, "learning_rate": 1.7090200035241976e-07, "loss": 0.7312248945236206, "step": 1361 }, { "ce_loss": 5.323101140675135e-05, "cls_loss": 0.059814453125, "epoch": 0.8174174174174175, "mask_bce_loss": 0.9623028039932251, "mask_dice_loss": 0.06276501715183258, "mask_loss": 1.0250678062438965, "step": 1361 }, { "epoch": 0.818018018018018, "grad_norm": 23.21356773376465, "learning_rate": 1.698159691844495e-07, "loss": 0.7373186945915222, "step": 1362 }, { "ce_loss": 0.00014406285481527448, "cls_loss": 0.05615234375, "epoch": 0.818018018018018, "mask_bce_loss": 0.7115849852561951, "mask_dice_loss": 0.1096365824341774, "mask_loss": 0.8212215900421143, "step": 1362 }, { "epoch": 0.8186186186186186, "grad_norm": 22.469955444335938, "learning_rate": 1.6873307945764338e-07, "loss": 0.818580150604248, "step": 1363 }, { "ce_loss": 4.659661863115616e-05, "cls_loss": 0.05224609375, "epoch": 0.8186186186186186, "mask_bce_loss": 1.0772309303283691, "mask_dice_loss": 0.08375679701566696, "mask_loss": 1.1609877347946167, "step": 1363 }, { "epoch": 0.8192192192192193, "grad_norm": 22.016517639160156, "learning_rate": 1.6765333526968827e-07, "loss": 0.8331716060638428, "step": 1364 }, { "ce_loss": 7.792301039444283e-05, "cls_loss": 0.04638671875, "epoch": 0.8192192192192193, "mask_bce_loss": 0.4032435119152069, "mask_dice_loss": 0.04445129632949829, "mask_loss": 0.4476948082447052, "step": 1364 }, { "epoch": 0.8198198198198198, "grad_norm": 19.016935348510742, "learning_rate": 1.6657674070636894e-07, "loss": 0.7507032155990601, "step": 1365 }, { "ce_loss": 4.2515774111961946e-05, "cls_loss": 0.05322265625, "epoch": 0.8198198198198198, "mask_bce_loss": 0.6739796996116638, "mask_dice_loss": 0.1453794240951538, "mask_loss": 0.8193591237068176, "step": 1365 }, { "epoch": 0.8204204204204204, "grad_norm": 21.04986000061035, "learning_rate": 1.655032998415511e-07, "loss": 0.8338176012039185, "step": 1366 }, { "ce_loss": 6.837479304522276e-05, "cls_loss": 0.046630859375, "epoch": 0.8204204204204204, "mask_bce_loss": 0.46384334564208984, "mask_dice_loss": 0.0689573734998703, "mask_loss": 0.5328007340431213, "step": 1366 }, { "epoch": 0.821021021021021, "grad_norm": 17.976539611816406, "learning_rate": 1.644330167371668e-07, "loss": 0.8523483276367188, "step": 1367 }, { "ce_loss": 0.00010690479393815622, "cls_loss": 0.04443359375, "epoch": 0.821021021021021, "mask_bce_loss": 0.6280797123908997, "mask_dice_loss": 0.07633888721466064, "mask_loss": 0.7044185996055603, "step": 1367 }, { "epoch": 0.8216216216216217, "grad_norm": 47.79705047607422, "learning_rate": 1.633658954431999e-07, "loss": 0.8118103742599487, "step": 1368 }, { "ce_loss": 0.0004105781845282763, "cls_loss": 0.0625, "epoch": 0.8216216216216217, "mask_bce_loss": 1.7647497653961182, "mask_dice_loss": 0.1298125684261322, "mask_loss": 1.8945623636245728, "step": 1368 }, { "epoch": 0.8222222222222222, "grad_norm": 58.96717071533203, "learning_rate": 1.6230193999766871e-07, "loss": 0.7829330563545227, "step": 1369 }, { "ce_loss": 3.5199143894715235e-05, "cls_loss": 0.047119140625, "epoch": 0.8222222222222222, "mask_bce_loss": 0.8428727388381958, "mask_dice_loss": 0.05287200212478638, "mask_loss": 0.8957447409629822, "step": 1369 }, { "epoch": 0.8228228228228228, "grad_norm": 28.078916549682617, "learning_rate": 1.6124115442661257e-07, "loss": 0.7095988988876343, "step": 1370 }, { "ce_loss": 3.7514248106162995e-05, "cls_loss": 0.038330078125, "epoch": 0.8228228228228228, "mask_bce_loss": 0.934569776058197, "mask_dice_loss": 0.031780730932950974, "mask_loss": 0.9663504958152771, "step": 1370 }, { "epoch": 0.8234234234234235, "grad_norm": 34.19446563720703, "learning_rate": 1.6018354274407586e-07, "loss": 0.9297623634338379, "step": 1371 }, { "ce_loss": 5.0787821237463504e-05, "cls_loss": 0.036376953125, "epoch": 0.8234234234234235, "mask_bce_loss": 0.2875367999076843, "mask_dice_loss": 0.028374766930937767, "mask_loss": 0.31591156125068665, "step": 1371 }, { "epoch": 0.824024024024024, "grad_norm": 24.180818557739258, "learning_rate": 1.5912910895209252e-07, "loss": 0.7706241011619568, "step": 1372 }, { "ce_loss": 5.8871315559372306e-05, "cls_loss": 0.07177734375, "epoch": 0.824024024024024, "mask_bce_loss": 1.3486509323120117, "mask_dice_loss": 0.175003319978714, "mask_loss": 1.5236542224884033, "step": 1372 }, { "epoch": 0.8246246246246246, "grad_norm": 25.126155853271484, "learning_rate": 1.5807785704067156e-07, "loss": 0.8339285850524902, "step": 1373 }, { "ce_loss": 3.730331445694901e-05, "cls_loss": 0.0546875, "epoch": 0.8246246246246246, "mask_bce_loss": 0.45220860838890076, "mask_dice_loss": 0.06275101751089096, "mask_loss": 0.5149596333503723, "step": 1373 }, { "epoch": 0.8252252252252252, "grad_norm": 43.85594940185547, "learning_rate": 1.570297909877818e-07, "loss": 0.9057456254959106, "step": 1374 }, { "ce_loss": 7.808965165168047e-05, "cls_loss": 0.044921875, "epoch": 0.8252252252252252, "mask_bce_loss": 1.0435446500778198, "mask_dice_loss": 0.07598374038934708, "mask_loss": 1.1195284128189087, "step": 1374 }, { "epoch": 0.8258258258258259, "grad_norm": 22.345062255859375, "learning_rate": 1.559849147593364e-07, "loss": 0.7203041315078735, "step": 1375 }, { "ce_loss": 4.649349648389034e-05, "cls_loss": 0.07275390625, "epoch": 0.8258258258258259, "mask_bce_loss": 0.506213366985321, "mask_dice_loss": 0.13246889412403107, "mask_loss": 0.6386822462081909, "step": 1375 }, { "epoch": 0.8264264264264264, "grad_norm": 34.04449462890625, "learning_rate": 1.5494323230917804e-07, "loss": 0.8148548007011414, "step": 1376 }, { "ce_loss": 9.572482667863369e-05, "cls_loss": 0.0693359375, "epoch": 0.8264264264264264, "mask_bce_loss": 0.37030717730522156, "mask_dice_loss": 0.09565778821706772, "mask_loss": 0.4659649729728699, "step": 1376 }, { "epoch": 0.827027027027027, "grad_norm": 21.92194938659668, "learning_rate": 1.5390474757906448e-07, "loss": 0.8037638664245605, "step": 1377 }, { "ce_loss": 0.0002798844361677766, "cls_loss": 0.037841796875, "epoch": 0.827027027027027, "mask_bce_loss": 0.7251359820365906, "mask_dice_loss": 0.03230902552604675, "mask_loss": 0.7574449777603149, "step": 1377 }, { "epoch": 0.8276276276276276, "grad_norm": 36.787052154541016, "learning_rate": 1.528694644986529e-07, "loss": 0.8611725568771362, "step": 1378 }, { "ce_loss": 0.0001476698089390993, "cls_loss": 0.06103515625, "epoch": 0.8276276276276276, "mask_bce_loss": 0.40964171290397644, "mask_dice_loss": 0.08294593542814255, "mask_loss": 0.4925876557826996, "step": 1378 }, { "epoch": 0.8282282282282283, "grad_norm": 23.210567474365234, "learning_rate": 1.518373869854853e-07, "loss": 0.8099888563156128, "step": 1379 }, { "ce_loss": 3.3279004128417e-05, "cls_loss": 0.0625, "epoch": 0.8282282282282283, "mask_bce_loss": 0.35554173588752747, "mask_dice_loss": 0.10947878658771515, "mask_loss": 0.4650205373764038, "step": 1379 }, { "epoch": 0.8288288288288288, "grad_norm": 17.143640518188477, "learning_rate": 1.5080851894497392e-07, "loss": 0.740979790687561, "step": 1380 }, { "ce_loss": 0.04634387791156769, "cls_loss": 0.021240234375, "epoch": 0.8288288288288288, "mask_bce_loss": 0.37839633226394653, "mask_dice_loss": 0.015012407675385475, "mask_loss": 0.39340874552726746, "step": 1380 }, { "epoch": 0.8294294294294294, "grad_norm": 20.061853408813477, "learning_rate": 1.49782864270386e-07, "loss": 0.7269317507743835, "step": 1381 }, { "ce_loss": 0.00010761415614979342, "cls_loss": 0.06396484375, "epoch": 0.8294294294294294, "mask_bce_loss": 1.0524883270263672, "mask_dice_loss": 0.11892250925302505, "mask_loss": 1.1714107990264893, "step": 1381 }, { "epoch": 0.8300300300300301, "grad_norm": 35.75151824951172, "learning_rate": 1.487604268428293e-07, "loss": 0.9091677665710449, "step": 1382 }, { "ce_loss": 7.71621271269396e-05, "cls_loss": 0.04052734375, "epoch": 0.8300300300300301, "mask_bce_loss": 0.6712149977684021, "mask_dice_loss": 0.05934762954711914, "mask_loss": 0.7305626273155212, "step": 1382 }, { "epoch": 0.8306306306306306, "grad_norm": 25.610952377319336, "learning_rate": 1.4774121053123812e-07, "loss": 0.7413212060928345, "step": 1383 }, { "ce_loss": 2.8903587008244358e-05, "cls_loss": 0.05029296875, "epoch": 0.8306306306306306, "mask_bce_loss": 1.153712511062622, "mask_dice_loss": 0.0628599151968956, "mask_loss": 1.2165724039077759, "step": 1383 }, { "epoch": 0.8312312312312312, "grad_norm": 21.29979133605957, "learning_rate": 1.4672521919235704e-07, "loss": 0.7836052775382996, "step": 1384 }, { "ce_loss": 5.372362647904083e-05, "cls_loss": 0.046875, "epoch": 0.8312312312312312, "mask_bce_loss": 0.38297128677368164, "mask_dice_loss": 0.04685216024518013, "mask_loss": 0.42982345819473267, "step": 1384 }, { "epoch": 0.8318318318318318, "grad_norm": 24.398908615112305, "learning_rate": 1.457124566707274e-07, "loss": 0.9093281030654907, "step": 1385 }, { "ce_loss": 0.00013847145601175725, "cls_loss": 0.03515625, "epoch": 0.8318318318318318, "mask_bce_loss": 0.4179134964942932, "mask_dice_loss": 0.056027382612228394, "mask_loss": 0.4739408791065216, "step": 1385 }, { "epoch": 0.8324324324324325, "grad_norm": 16.368350982666016, "learning_rate": 1.4470292679867347e-07, "loss": 0.6338098049163818, "step": 1386 }, { "ce_loss": 8.765731035964563e-05, "cls_loss": 0.06396484375, "epoch": 0.8324324324324325, "mask_bce_loss": 1.310649037361145, "mask_dice_loss": 0.08243858814239502, "mask_loss": 1.39308762550354, "step": 1386 }, { "epoch": 0.833033033033033, "grad_norm": 44.47443771362305, "learning_rate": 1.436966333962859e-07, "loss": 0.6999610662460327, "step": 1387 }, { "ce_loss": 5.6526150729041547e-05, "cls_loss": 0.0242919921875, "epoch": 0.833033033033033, "mask_bce_loss": 0.2717858850955963, "mask_dice_loss": 0.01745798997581005, "mask_loss": 0.2892438769340515, "step": 1387 }, { "epoch": 0.8336336336336336, "grad_norm": 35.76708221435547, "learning_rate": 1.4269358027140932e-07, "loss": 0.7953603267669678, "step": 1388 }, { "ce_loss": 2.65751968981931e-05, "cls_loss": 0.04345703125, "epoch": 0.8336336336336336, "mask_bce_loss": 0.5251633524894714, "mask_dice_loss": 0.03228113055229187, "mask_loss": 0.5574444532394409, "step": 1388 }, { "epoch": 0.8342342342342343, "grad_norm": 19.752504348754883, "learning_rate": 1.4169377121962655e-07, "loss": 0.7952221632003784, "step": 1389 }, { "ce_loss": 0.00011028642620658502, "cls_loss": 0.056640625, "epoch": 0.8342342342342343, "mask_bce_loss": 0.3799731433391571, "mask_dice_loss": 0.12583518028259277, "mask_loss": 0.5058083534240723, "step": 1389 }, { "epoch": 0.8348348348348348, "grad_norm": 35.158267974853516, "learning_rate": 1.4069721002424484e-07, "loss": 0.9877934455871582, "step": 1390 }, { "ce_loss": 0.00010277178807882592, "cls_loss": 0.04443359375, "epoch": 0.8348348348348348, "mask_bce_loss": 0.6161482930183411, "mask_dice_loss": 0.07918786257505417, "mask_loss": 0.6953361630439758, "step": 1390 }, { "epoch": 0.8354354354354354, "grad_norm": 21.32208824157715, "learning_rate": 1.3970390045628132e-07, "loss": 0.7733190059661865, "step": 1391 }, { "ce_loss": 0.00019024088396690786, "cls_loss": 0.05224609375, "epoch": 0.8354354354354354, "mask_bce_loss": 1.043601632118225, "mask_dice_loss": 0.05620493367314339, "mask_loss": 1.099806547164917, "step": 1391 }, { "epoch": 0.836036036036036, "grad_norm": 21.086576461791992, "learning_rate": 1.3871384627444958e-07, "loss": 0.7859137058258057, "step": 1392 }, { "ce_loss": 2.525357558624819e-05, "cls_loss": 0.055419921875, "epoch": 0.836036036036036, "mask_bce_loss": 0.38217493891716003, "mask_dice_loss": 0.10524073988199234, "mask_loss": 0.4874156713485718, "step": 1392 }, { "epoch": 0.8366366366366367, "grad_norm": 38.671871185302734, "learning_rate": 1.3772705122514395e-07, "loss": 0.8064161539077759, "step": 1393 }, { "ce_loss": 2.8323782316874713e-05, "cls_loss": 0.044677734375, "epoch": 0.8366366366366367, "mask_bce_loss": 0.1837969273328781, "mask_dice_loss": 0.03650451824069023, "mask_loss": 0.22030144929885864, "step": 1393 }, { "epoch": 0.8372372372372372, "grad_norm": 45.64208984375, "learning_rate": 1.3674351904242608e-07, "loss": 0.8461073637008667, "step": 1394 }, { "ce_loss": 3.7823483580723405e-05, "cls_loss": 0.03662109375, "epoch": 0.8372372372372372, "mask_bce_loss": 0.43577343225479126, "mask_dice_loss": 0.03673643618822098, "mask_loss": 0.47250986099243164, "step": 1394 }, { "epoch": 0.8378378378378378, "grad_norm": 35.983673095703125, "learning_rate": 1.3576325344801165e-07, "loss": 0.7326655387878418, "step": 1395 }, { "ce_loss": 6.254521576920524e-05, "cls_loss": 0.03271484375, "epoch": 0.8378378378378378, "mask_bce_loss": 0.21520839631557465, "mask_dice_loss": 0.02471013180911541, "mask_loss": 0.2399185299873352, "step": 1395 }, { "epoch": 0.8384384384384385, "grad_norm": 27.167625427246094, "learning_rate": 1.3478625815125466e-07, "loss": 0.7520112991333008, "step": 1396 }, { "ce_loss": 0.00016112708544824272, "cls_loss": 0.055908203125, "epoch": 0.8384384384384385, "mask_bce_loss": 0.9333621859550476, "mask_dice_loss": 0.08150042593479156, "mask_loss": 1.0148626565933228, "step": 1396 }, { "epoch": 0.8390390390390391, "grad_norm": 17.608814239501953, "learning_rate": 1.338125368491345e-07, "loss": 0.7103894352912903, "step": 1397 }, { "ce_loss": 0.00015178255853243172, "cls_loss": 0.0654296875, "epoch": 0.8390390390390391, "mask_bce_loss": 0.9390491843223572, "mask_dice_loss": 0.17192406952381134, "mask_loss": 1.1109732389450073, "step": 1397 }, { "epoch": 0.8396396396396396, "grad_norm": 43.96320343017578, "learning_rate": 1.328420932262416e-07, "loss": 0.8641735911369324, "step": 1398 }, { "ce_loss": 5.6714899983489886e-05, "cls_loss": 0.06005859375, "epoch": 0.8396396396396396, "mask_bce_loss": 0.6869437098503113, "mask_dice_loss": 0.11625619232654572, "mask_loss": 0.8031998872756958, "step": 1398 }, { "epoch": 0.8402402402402402, "grad_norm": 23.034072875976562, "learning_rate": 1.3187493095476388e-07, "loss": 0.8132345080375671, "step": 1399 }, { "ce_loss": 0.0010654559591785073, "cls_loss": 0.041748046875, "epoch": 0.8402402402402402, "mask_bce_loss": 0.34057602286338806, "mask_dice_loss": 0.047982845455408096, "mask_loss": 0.38855886459350586, "step": 1399 }, { "epoch": 0.8408408408408409, "grad_norm": 17.282915115356445, "learning_rate": 1.3091105369447165e-07, "loss": 0.7845900654792786, "step": 1400 }, { "ce_loss": 5.980750574963167e-05, "cls_loss": 0.05615234375, "epoch": 0.8408408408408409, "mask_bce_loss": 0.8425770998001099, "mask_dice_loss": 0.11328689008951187, "mask_loss": 0.9558640122413635, "step": 1400 }, { "epoch": 0.8414414414414414, "grad_norm": 58.694801330566406, "learning_rate": 1.2995046509270592e-07, "loss": 0.9397027492523193, "step": 1401 }, { "ce_loss": 0.00046162708895280957, "cls_loss": 0.05029296875, "epoch": 0.8414414414414414, "mask_bce_loss": 1.4458246231079102, "mask_dice_loss": 0.057255614548921585, "mask_loss": 1.5030802488327026, "step": 1401 }, { "epoch": 0.842042042042042, "grad_norm": 44.10523223876953, "learning_rate": 1.289931687843624e-07, "loss": 0.887285590171814, "step": 1402 }, { "ce_loss": 0.00011647119390545413, "cls_loss": 0.057861328125, "epoch": 0.842042042042042, "mask_bce_loss": 0.5591073632240295, "mask_dice_loss": 0.08563156425952911, "mask_loss": 0.6447389125823975, "step": 1402 }, { "epoch": 0.8426426426426427, "grad_norm": 52.624473571777344, "learning_rate": 1.2803916839187868e-07, "loss": 0.7856591939926147, "step": 1403 }, { "ce_loss": 0.000853344623465091, "cls_loss": 0.06005859375, "epoch": 0.8426426426426427, "mask_bce_loss": 1.0847575664520264, "mask_dice_loss": 0.08992283791303635, "mask_loss": 1.1746803522109985, "step": 1403 }, { "epoch": 0.8432432432432433, "grad_norm": 24.21363067626953, "learning_rate": 1.2708846752522118e-07, "loss": 0.7441657781600952, "step": 1404 }, { "ce_loss": 4.063563756062649e-05, "cls_loss": 0.05859375, "epoch": 0.8432432432432433, "mask_bce_loss": 0.9403664469718933, "mask_dice_loss": 0.17045705020427704, "mask_loss": 1.1108235120773315, "step": 1404 }, { "epoch": 0.8438438438438438, "grad_norm": 43.79796600341797, "learning_rate": 1.2614106978187012e-07, "loss": 0.8831654787063599, "step": 1405 }, { "ce_loss": 2.9160684789530933e-05, "cls_loss": 0.049560546875, "epoch": 0.8438438438438438, "mask_bce_loss": 0.7409956455230713, "mask_dice_loss": 0.06177929788827896, "mask_loss": 0.802774965763092, "step": 1405 }, { "epoch": 0.8444444444444444, "grad_norm": 28.8966064453125, "learning_rate": 1.251969787468068e-07, "loss": 0.6984028220176697, "step": 1406 }, { "ce_loss": 0.00016491716087330133, "cls_loss": 0.029052734375, "epoch": 0.8444444444444444, "mask_bce_loss": 0.5844823718070984, "mask_dice_loss": 0.02200368233025074, "mask_loss": 0.6064860820770264, "step": 1406 }, { "epoch": 0.8450450450450451, "grad_norm": 15.667647361755371, "learning_rate": 1.2425619799249997e-07, "loss": 0.6823577284812927, "step": 1407 }, { "ce_loss": 0.00016550146392546594, "cls_loss": 0.053955078125, "epoch": 0.8450450450450451, "mask_bce_loss": 1.2320215702056885, "mask_dice_loss": 0.09809726476669312, "mask_loss": 1.3301188945770264, "step": 1407 }, { "epoch": 0.8456456456456456, "grad_norm": 34.85489273071289, "learning_rate": 1.2331873107889212e-07, "loss": 0.9150896668434143, "step": 1408 }, { "ce_loss": 9.941491589415818e-05, "cls_loss": 0.07421875, "epoch": 0.8456456456456456, "mask_bce_loss": 1.3270517587661743, "mask_dice_loss": 0.11456992477178574, "mask_loss": 1.4416216611862183, "step": 1408 }, { "epoch": 0.8462462462462462, "grad_norm": 17.832155227661133, "learning_rate": 1.2238458155338572e-07, "loss": 0.6942142248153687, "step": 1409 }, { "ce_loss": 0.0001359202287858352, "cls_loss": 0.057373046875, "epoch": 0.8462462462462462, "mask_bce_loss": 0.6190802454948425, "mask_dice_loss": 0.11861731857061386, "mask_loss": 0.7376975417137146, "step": 1409 }, { "epoch": 0.8468468468468469, "grad_norm": 43.10861587524414, "learning_rate": 1.2145375295083104e-07, "loss": 0.7560985088348389, "step": 1410 }, { "ce_loss": 6.406986358342692e-05, "cls_loss": 0.06640625, "epoch": 0.8468468468468469, "mask_bce_loss": 0.5497568249702454, "mask_dice_loss": 0.14293447136878967, "mask_loss": 0.6926913261413574, "step": 1410 }, { "epoch": 0.8474474474474475, "grad_norm": 15.433608055114746, "learning_rate": 1.2052624879351103e-07, "loss": 0.7328804731369019, "step": 1411 }, { "ce_loss": 0.0001930307043949142, "cls_loss": 0.06396484375, "epoch": 0.8474474474474475, "mask_bce_loss": 1.026546835899353, "mask_dice_loss": 0.08298888057470322, "mask_loss": 1.1095356941223145, "step": 1411 }, { "epoch": 0.848048048048048, "grad_norm": 38.570037841796875, "learning_rate": 1.1960207259112887e-07, "loss": 0.7885546088218689, "step": 1412 }, { "ce_loss": 4.8691847041482106e-05, "cls_loss": 0.0498046875, "epoch": 0.848048048048048, "mask_bce_loss": 1.1105296611785889, "mask_dice_loss": 0.062074996531009674, "mask_loss": 1.1726046800613403, "step": 1412 }, { "epoch": 0.8486486486486486, "grad_norm": 32.33998489379883, "learning_rate": 1.1868122784079538e-07, "loss": 0.7916386127471924, "step": 1413 }, { "ce_loss": 7.983861723914742e-05, "cls_loss": 0.0498046875, "epoch": 0.8486486486486486, "mask_bce_loss": 0.868384063243866, "mask_dice_loss": 0.04684712365269661, "mask_loss": 0.9152311682701111, "step": 1413 }, { "epoch": 0.8492492492492493, "grad_norm": 37.035953521728516, "learning_rate": 1.1776371802701457e-07, "loss": 0.8171737790107727, "step": 1414 }, { "ce_loss": 6.599210610147566e-05, "cls_loss": 0.059326171875, "epoch": 0.8492492492492493, "mask_bce_loss": 0.5125846862792969, "mask_dice_loss": 0.1519099324941635, "mask_loss": 0.6644946336746216, "step": 1414 }, { "epoch": 0.8498498498498499, "grad_norm": 53.123321533203125, "learning_rate": 1.1684954662167035e-07, "loss": 0.8411667943000793, "step": 1415 }, { "ce_loss": 0.014307687059044838, "cls_loss": 0.040283203125, "epoch": 0.8498498498498499, "mask_bce_loss": 0.9502273797988892, "mask_dice_loss": 0.06565023958683014, "mask_loss": 1.015877604484558, "step": 1415 }, { "epoch": 0.8504504504504504, "grad_norm": 23.640161514282227, "learning_rate": 1.1593871708401526e-07, "loss": 0.8271056413650513, "step": 1416 }, { "ce_loss": 4.8584708565613255e-05, "cls_loss": 0.041259765625, "epoch": 0.8504504504504504, "mask_bce_loss": 0.8229272961616516, "mask_dice_loss": 0.041877277195453644, "mask_loss": 0.8648045659065247, "step": 1416 }, { "epoch": 0.851051051051051, "grad_norm": 38.48598861694336, "learning_rate": 1.150312328606553e-07, "loss": 1.0519952774047852, "step": 1417 }, { "ce_loss": 5.443701593321748e-05, "cls_loss": 0.055419921875, "epoch": 0.851051051051051, "mask_bce_loss": 0.578689694404602, "mask_dice_loss": 0.07595964521169662, "mask_loss": 0.6546493172645569, "step": 1417 }, { "epoch": 0.8516516516516517, "grad_norm": 17.63282585144043, "learning_rate": 1.141270973855375e-07, "loss": 0.7755000591278076, "step": 1418 }, { "ce_loss": 0.03575986251235008, "cls_loss": 0.0439453125, "epoch": 0.8516516516516517, "mask_bce_loss": 0.6664065718650818, "mask_dice_loss": 0.047732990235090256, "mask_loss": 0.7141395807266235, "step": 1418 }, { "epoch": 0.8522522522522522, "grad_norm": 34.2525520324707, "learning_rate": 1.132263140799381e-07, "loss": 0.8152965903282166, "step": 1419 }, { "ce_loss": 4.131539390073158e-05, "cls_loss": 0.0238037109375, "epoch": 0.8522522522522522, "mask_bce_loss": 0.19766734540462494, "mask_dice_loss": 0.019209301099181175, "mask_loss": 0.21687664091587067, "step": 1419 }, { "epoch": 0.8528528528528528, "grad_norm": 25.221542358398438, "learning_rate": 1.1232888635244775e-07, "loss": 0.8059414029121399, "step": 1420 }, { "ce_loss": 0.0006989054963923991, "cls_loss": 0.053955078125, "epoch": 0.8528528528528528, "mask_bce_loss": 0.5947116017341614, "mask_dice_loss": 0.11290694773197174, "mask_loss": 0.7076185345649719, "step": 1420 }, { "epoch": 0.8534534534534535, "grad_norm": 22.869462966918945, "learning_rate": 1.1143481759895957e-07, "loss": 0.7789871692657471, "step": 1421 }, { "ce_loss": 5.045134457759559e-05, "cls_loss": 0.0181884765625, "epoch": 0.8534534534534535, "mask_bce_loss": 0.09558883309364319, "mask_dice_loss": 0.01122849341481924, "mask_loss": 0.106817327439785, "step": 1421 }, { "epoch": 0.8540540540540541, "grad_norm": 15.444385528564453, "learning_rate": 1.1054411120265683e-07, "loss": 0.6455371379852295, "step": 1422 }, { "ce_loss": 0.000394502334529534, "cls_loss": 0.0341796875, "epoch": 0.8540540540540541, "mask_bce_loss": 0.4420309066772461, "mask_dice_loss": 0.028653744608163834, "mask_loss": 0.47068464756011963, "step": 1422 }, { "epoch": 0.8546546546546546, "grad_norm": 27.809680938720703, "learning_rate": 1.0965677053399914e-07, "loss": 0.9074257016181946, "step": 1423 }, { "ce_loss": 6.789325561840087e-05, "cls_loss": 0.053955078125, "epoch": 0.8546546546546546, "mask_bce_loss": 0.7621464133262634, "mask_dice_loss": 0.06898844242095947, "mask_loss": 0.8311348557472229, "step": 1423 }, { "epoch": 0.8552552552552553, "grad_norm": 37.711368560791016, "learning_rate": 1.0877279895070945e-07, "loss": 0.969646692276001, "step": 1424 }, { "ce_loss": 8.210180385503918e-05, "cls_loss": 0.03369140625, "epoch": 0.8552552552552553, "mask_bce_loss": 0.282596230506897, "mask_dice_loss": 0.029718980193138123, "mask_loss": 0.3123152256011963, "step": 1424 }, { "epoch": 0.8558558558558559, "grad_norm": 33.97088623046875, "learning_rate": 1.078921997977632e-07, "loss": 0.9443102478981018, "step": 1425 }, { "ce_loss": 5.306054663378745e-05, "cls_loss": 0.041748046875, "epoch": 0.8558558558558559, "mask_bce_loss": 0.7721970677375793, "mask_dice_loss": 0.042442843317985535, "mask_loss": 0.8146399259567261, "step": 1425 }, { "epoch": 0.8564564564564564, "grad_norm": 24.42403793334961, "learning_rate": 1.0701497640737377e-07, "loss": 0.8956625461578369, "step": 1426 }, { "ce_loss": 0.0001018105394905433, "cls_loss": 0.040283203125, "epoch": 0.8564564564564564, "mask_bce_loss": 0.6893765330314636, "mask_dice_loss": 0.05967165157198906, "mask_loss": 0.7490481734275818, "step": 1426 }, { "epoch": 0.857057057057057, "grad_norm": 34.16425323486328, "learning_rate": 1.0614113209898046e-07, "loss": 0.741096019744873, "step": 1427 }, { "ce_loss": 3.978144741267897e-05, "cls_loss": 0.049072265625, "epoch": 0.857057057057057, "mask_bce_loss": 0.4715305268764496, "mask_dice_loss": 0.05454664304852486, "mask_loss": 0.526077151298523, "step": 1427 }, { "epoch": 0.8576576576576577, "grad_norm": 28.32834243774414, "learning_rate": 1.0527067017923652e-07, "loss": 0.8061798810958862, "step": 1428 }, { "ce_loss": 4.8087076720548794e-05, "cls_loss": 0.05810546875, "epoch": 0.8576576576576577, "mask_bce_loss": 0.9210425615310669, "mask_dice_loss": 0.07014946639537811, "mask_loss": 0.9911920428276062, "step": 1428 }, { "epoch": 0.8582582582582583, "grad_norm": 28.254796981811523, "learning_rate": 1.0440359394199583e-07, "loss": 0.7659914493560791, "step": 1429 }, { "ce_loss": 6.652100273640826e-05, "cls_loss": 0.053466796875, "epoch": 0.8582582582582583, "mask_bce_loss": 0.8817338943481445, "mask_dice_loss": 0.0742102786898613, "mask_loss": 0.9559441804885864, "step": 1429 }, { "epoch": 0.8588588588588588, "grad_norm": 33.62617492675781, "learning_rate": 1.0353990666830037e-07, "loss": 0.8401317000389099, "step": 1430 }, { "ce_loss": 0.0001443733781343326, "cls_loss": 0.046142578125, "epoch": 0.8588588588588588, "mask_bce_loss": 0.6803104281425476, "mask_dice_loss": 0.07214287668466568, "mask_loss": 0.7524533271789551, "step": 1430 }, { "epoch": 0.8594594594594595, "grad_norm": 40.40386199951172, "learning_rate": 1.0267961162636917e-07, "loss": 0.7890908718109131, "step": 1431 }, { "ce_loss": 3.127412128378637e-05, "cls_loss": 0.0654296875, "epoch": 0.8594594594594595, "mask_bce_loss": 0.27009841799736023, "mask_dice_loss": 0.02280229888856411, "mask_loss": 0.2929007112979889, "step": 1431 }, { "epoch": 0.8600600600600601, "grad_norm": 24.919208526611328, "learning_rate": 1.0182271207158422e-07, "loss": 0.8315411806106567, "step": 1432 }, { "ce_loss": 0.0017564985901117325, "cls_loss": 0.045654296875, "epoch": 0.8600600600600601, "mask_bce_loss": 1.1634472608566284, "mask_dice_loss": 0.10979855060577393, "mask_loss": 1.2732458114624023, "step": 1432 }, { "epoch": 0.8606606606606607, "grad_norm": 21.84977149963379, "learning_rate": 1.0096921124647839e-07, "loss": 0.7819463610649109, "step": 1433 }, { "ce_loss": 3.0643423087894917e-05, "cls_loss": 0.05322265625, "epoch": 0.8606606606606607, "mask_bce_loss": 0.8434557914733887, "mask_dice_loss": 0.1003446951508522, "mask_loss": 0.9438005089759827, "step": 1433 }, { "epoch": 0.8612612612612612, "grad_norm": 32.25675582885742, "learning_rate": 1.0011911238072501e-07, "loss": 0.773833155632019, "step": 1434 }, { "ce_loss": 3.004424797836691e-05, "cls_loss": 0.1201171875, "epoch": 0.8612612612612612, "mask_bce_loss": 0.7325303554534912, "mask_dice_loss": 0.05858555808663368, "mask_loss": 0.791115939617157, "step": 1434 }, { "epoch": 0.8618618618618619, "grad_norm": 41.98572540283203, "learning_rate": 9.927241869112335e-08, "loss": 0.8355801105499268, "step": 1435 }, { "ce_loss": 3.2530657335883006e-05, "cls_loss": 0.040771484375, "epoch": 0.8618618618618619, "mask_bce_loss": 0.459612101316452, "mask_dice_loss": 0.05523398518562317, "mask_loss": 0.5148460865020752, "step": 1435 }, { "epoch": 0.8624624624624625, "grad_norm": 27.44304656982422, "learning_rate": 9.842913338158732e-08, "loss": 0.7923280596733093, "step": 1436 }, { "ce_loss": 0.0007203268469311297, "cls_loss": 0.0546875, "epoch": 0.8624624624624625, "mask_bce_loss": 0.8075595498085022, "mask_dice_loss": 0.06282901018857956, "mask_loss": 0.8703885674476624, "step": 1436 }, { "epoch": 0.863063063063063, "grad_norm": 44.70570755004883, "learning_rate": 9.758925964313425e-08, "loss": 0.8945378661155701, "step": 1437 }, { "ce_loss": 3.928448859369382e-05, "cls_loss": 0.0172119140625, "epoch": 0.863063063063063, "mask_bce_loss": 0.11498415470123291, "mask_dice_loss": 0.010504433885216713, "mask_loss": 0.12548859417438507, "step": 1437 }, { "epoch": 0.8636636636636636, "grad_norm": 27.31764793395996, "learning_rate": 9.675280065387114e-08, "loss": 0.699474573135376, "step": 1438 }, { "ce_loss": 4.0517323213862255e-05, "cls_loss": 0.0693359375, "epoch": 0.8636636636636636, "mask_bce_loss": 0.20862965285778046, "mask_dice_loss": 0.03202565759420395, "mask_loss": 0.2406553030014038, "step": 1438 }, { "epoch": 0.8642642642642643, "grad_norm": 64.54093170166016, "learning_rate": 9.591975957898368e-08, "loss": 0.9351232051849365, "step": 1439 }, { "ce_loss": 3.287885192548856e-05, "cls_loss": 0.041015625, "epoch": 0.8642642642642643, "mask_bce_loss": 0.4015236496925354, "mask_dice_loss": 0.03724448010325432, "mask_loss": 0.4387681186199188, "step": 1439 }, { "epoch": 0.8648648648648649, "grad_norm": 28.02349281311035, "learning_rate": 9.509013957072454e-08, "loss": 0.7979149222373962, "step": 1440 }, { "ce_loss": 0.0001705248432699591, "cls_loss": 0.05615234375, "epoch": 0.8648648648648649, "mask_bce_loss": 0.40207529067993164, "mask_dice_loss": 0.10510312765836716, "mask_loss": 0.5071784257888794, "step": 1440 }, { "epoch": 0.8654654654654654, "grad_norm": 17.732114791870117, "learning_rate": 9.426394376840052e-08, "loss": 0.7445254921913147, "step": 1441 }, { "ce_loss": 6.403733277693391e-05, "cls_loss": 0.08740234375, "epoch": 0.8654654654654654, "mask_bce_loss": 0.6744541525840759, "mask_dice_loss": 0.10609617084264755, "mask_loss": 0.7805503010749817, "step": 1441 }, { "epoch": 0.8660660660660661, "grad_norm": 54.20353317260742, "learning_rate": 9.344117529836071e-08, "loss": 1.0360933542251587, "step": 1442 }, { "ce_loss": 3.771562478505075e-05, "cls_loss": 0.042236328125, "epoch": 0.8660660660660661, "mask_bce_loss": 0.4245412051677704, "mask_dice_loss": 0.04640280827879906, "mask_loss": 0.47094401717185974, "step": 1442 }, { "epoch": 0.8666666666666667, "grad_norm": 23.437734603881836, "learning_rate": 9.26218372739862e-08, "loss": 0.8145735263824463, "step": 1443 }, { "ce_loss": 0.003345792181789875, "cls_loss": 0.048095703125, "epoch": 0.8666666666666667, "mask_bce_loss": 1.0255321264266968, "mask_dice_loss": 0.1201433464884758, "mask_loss": 1.1456754207611084, "step": 1443 }, { "epoch": 0.8672672672672672, "grad_norm": 25.593276977539062, "learning_rate": 9.180593279567616e-08, "loss": 0.8883010745048523, "step": 1444 }, { "ce_loss": 3.764987559407018e-05, "cls_loss": 0.04248046875, "epoch": 0.8672672672672672, "mask_bce_loss": 0.7537776827812195, "mask_dice_loss": 0.07983852922916412, "mask_loss": 0.8336161971092224, "step": 1444 }, { "epoch": 0.8678678678678678, "grad_norm": 22.042314529418945, "learning_rate": 9.099346495083749e-08, "loss": 0.6684699058532715, "step": 1445 }, { "ce_loss": 0.0008300761110149324, "cls_loss": 0.04443359375, "epoch": 0.8678678678678678, "mask_bce_loss": 0.3667539954185486, "mask_dice_loss": 0.0800490528345108, "mask_loss": 0.4468030333518982, "step": 1445 }, { "epoch": 0.8684684684684685, "grad_norm": 28.319377899169922, "learning_rate": 9.01844368138731e-08, "loss": 0.7803988456726074, "step": 1446 }, { "ce_loss": 0.00016204947314690799, "cls_loss": 0.031982421875, "epoch": 0.8684684684684685, "mask_bce_loss": 0.27118033170700073, "mask_dice_loss": 0.024069149047136307, "mask_loss": 0.29524949193000793, "step": 1446 }, { "epoch": 0.8690690690690691, "grad_norm": 21.47161102294922, "learning_rate": 8.937885144616919e-08, "loss": 0.8667815923690796, "step": 1447 }, { "ce_loss": 4.29160863859579e-05, "cls_loss": 0.059814453125, "epoch": 0.8690690690690691, "mask_bce_loss": 0.7875692844390869, "mask_dice_loss": 0.0831700935959816, "mask_loss": 0.8707394003868103, "step": 1447 }, { "epoch": 0.8696696696696696, "grad_norm": 71.59803009033203, "learning_rate": 8.85767118960854e-08, "loss": 0.876186728477478, "step": 1448 }, { "ce_loss": 0.0018228047993034124, "cls_loss": 0.05712890625, "epoch": 0.8696696696696696, "mask_bce_loss": 0.867039144039154, "mask_dice_loss": 0.08221359550952911, "mask_loss": 0.949252724647522, "step": 1448 }, { "epoch": 0.8702702702702703, "grad_norm": 24.587900161743164, "learning_rate": 8.777802119894151e-08, "loss": 0.8742303848266602, "step": 1449 }, { "ce_loss": 2.4731360099394806e-05, "cls_loss": 0.0634765625, "epoch": 0.8702702702702703, "mask_bce_loss": 0.4952019155025482, "mask_dice_loss": 0.09696133434772491, "mask_loss": 0.5921632647514343, "step": 1449 }, { "epoch": 0.8708708708708709, "grad_norm": 22.80773162841797, "learning_rate": 8.698278237700651e-08, "loss": 0.8147166967391968, "step": 1450 }, { "ce_loss": 5.681602851836942e-05, "cls_loss": 0.060546875, "epoch": 0.8708708708708709, "mask_bce_loss": 0.9694210290908813, "mask_dice_loss": 0.1074555441737175, "mask_loss": 1.0768765211105347, "step": 1450 }, { "epoch": 0.8714714714714715, "grad_norm": 20.700674057006836, "learning_rate": 8.61909984394884e-08, "loss": 0.8065765500068665, "step": 1451 }, { "ce_loss": 0.0006613194127567112, "cls_loss": 0.04541015625, "epoch": 0.8714714714714715, "mask_bce_loss": 0.3203727900981903, "mask_dice_loss": 0.038225140422582626, "mask_loss": 0.35859793424606323, "step": 1451 }, { "epoch": 0.872072072072072, "grad_norm": 50.99140167236328, "learning_rate": 8.540267238252107e-08, "loss": 0.894781768321991, "step": 1452 }, { "ce_loss": 2.5810140868998133e-05, "cls_loss": 0.048828125, "epoch": 0.872072072072072, "mask_bce_loss": 0.7758011221885681, "mask_dice_loss": 0.060968536883592606, "mask_loss": 0.8367696404457092, "step": 1452 }, { "epoch": 0.8726726726726727, "grad_norm": 41.51217269897461, "learning_rate": 8.461780718915378e-08, "loss": 0.7477806806564331, "step": 1453 }, { "ce_loss": 0.000251174031291157, "cls_loss": 0.045166015625, "epoch": 0.8726726726726727, "mask_bce_loss": 0.6815659403800964, "mask_dice_loss": 0.0752914696931839, "mask_loss": 0.7568573951721191, "step": 1453 }, { "epoch": 0.8732732732732733, "grad_norm": 27.411083221435547, "learning_rate": 8.383640582934004e-08, "loss": 0.7172006368637085, "step": 1454 }, { "ce_loss": 4.875648664892651e-05, "cls_loss": 0.0400390625, "epoch": 0.8732732732732733, "mask_bce_loss": 0.8277225494384766, "mask_dice_loss": 0.06208086758852005, "mask_loss": 0.889803409576416, "step": 1454 }, { "epoch": 0.8738738738738738, "grad_norm": 28.49555778503418, "learning_rate": 8.305847125992593e-08, "loss": 0.714962899684906, "step": 1455 }, { "ce_loss": 7.744233880657703e-05, "cls_loss": 0.048095703125, "epoch": 0.8738738738738738, "mask_bce_loss": 0.4944915771484375, "mask_dice_loss": 0.04477328062057495, "mask_loss": 0.5392648577690125, "step": 1455 }, { "epoch": 0.8744744744744745, "grad_norm": 24.97199058532715, "learning_rate": 8.228400642463873e-08, "loss": 0.7244499921798706, "step": 1456 }, { "ce_loss": 8.738750329939649e-05, "cls_loss": 0.055419921875, "epoch": 0.8744744744744745, "mask_bce_loss": 0.8585386276245117, "mask_dice_loss": 0.12580250203609467, "mask_loss": 0.9843411445617676, "step": 1456 }, { "epoch": 0.8750750750750751, "grad_norm": 49.739837646484375, "learning_rate": 8.151301425407698e-08, "loss": 0.7675367593765259, "step": 1457 }, { "ce_loss": 5.936877278145403e-05, "cls_loss": 0.052734375, "epoch": 0.8750750750750751, "mask_bce_loss": 0.6766008138656616, "mask_dice_loss": 0.03792889416217804, "mask_loss": 0.7145296931266785, "step": 1457 }, { "epoch": 0.8756756756756757, "grad_norm": 44.93984603881836, "learning_rate": 8.074549766569794e-08, "loss": 0.8760418891906738, "step": 1458 }, { "ce_loss": 0.00010011818812927231, "cls_loss": 0.05615234375, "epoch": 0.8756756756756757, "mask_bce_loss": 0.5712503790855408, "mask_dice_loss": 0.10792437940835953, "mask_loss": 0.6791747808456421, "step": 1458 }, { "epoch": 0.8762762762762762, "grad_norm": 30.762155532836914, "learning_rate": 7.998145956380697e-08, "loss": 0.8219060897827148, "step": 1459 }, { "ce_loss": 3.578913310775533e-05, "cls_loss": 0.038330078125, "epoch": 0.8762762762762762, "mask_bce_loss": 0.6318856477737427, "mask_dice_loss": 0.06494387239217758, "mask_loss": 0.6968294978141785, "step": 1459 }, { "epoch": 0.8768768768768769, "grad_norm": 21.060518264770508, "learning_rate": 7.922090283954753e-08, "loss": 0.7785722017288208, "step": 1460 }, { "ce_loss": 2.318689257663209e-05, "cls_loss": 0.04541015625, "epoch": 0.8768768768768769, "mask_bce_loss": 0.6030166745185852, "mask_dice_loss": 0.057217251509428024, "mask_loss": 0.6602339148521423, "step": 1460 }, { "epoch": 0.8774774774774775, "grad_norm": 18.223283767700195, "learning_rate": 7.846383037088866e-08, "loss": 0.8690675497055054, "step": 1461 }, { "ce_loss": 2.5337873012176715e-05, "cls_loss": 0.061279296875, "epoch": 0.8774774774774775, "mask_bce_loss": 0.5407604575157166, "mask_dice_loss": 0.06905729323625565, "mask_loss": 0.6098177433013916, "step": 1461 }, { "epoch": 0.878078078078078, "grad_norm": 17.599353790283203, "learning_rate": 7.771024502261525e-08, "loss": 0.8275448083877563, "step": 1462 }, { "ce_loss": 6.069150913390331e-05, "cls_loss": 0.037841796875, "epoch": 0.878078078078078, "mask_bce_loss": 0.5381034016609192, "mask_dice_loss": 0.03736668825149536, "mask_loss": 0.5754700899124146, "step": 1462 }, { "epoch": 0.8786786786786787, "grad_norm": 28.835786819458008, "learning_rate": 7.696014964631692e-08, "loss": 0.857324481010437, "step": 1463 }, { "ce_loss": 6.343862332869321e-05, "cls_loss": 0.0634765625, "epoch": 0.8786786786786787, "mask_bce_loss": 1.4218941926956177, "mask_dice_loss": 0.14816491305828094, "mask_loss": 1.570059061050415, "step": 1463 }, { "epoch": 0.8792792792792793, "grad_norm": 17.767173767089844, "learning_rate": 7.621354708037686e-08, "loss": 0.8849021792411804, "step": 1464 }, { "ce_loss": 7.226313027786091e-05, "cls_loss": 0.04638671875, "epoch": 0.8792792792792793, "mask_bce_loss": 0.4102761447429657, "mask_dice_loss": 0.1450365036725998, "mask_loss": 0.5553126335144043, "step": 1464 }, { "epoch": 0.8798798798798799, "grad_norm": 60.94428634643555, "learning_rate": 7.547044014996129e-08, "loss": 0.9250492453575134, "step": 1465 }, { "ce_loss": 0.0010835077846422791, "cls_loss": 0.11181640625, "epoch": 0.8798798798798799, "mask_bce_loss": 0.9804744124412537, "mask_dice_loss": 0.10854743421077728, "mask_loss": 1.0890218019485474, "step": 1465 }, { "epoch": 0.8804804804804804, "grad_norm": 19.78199577331543, "learning_rate": 7.473083166700944e-08, "loss": 0.7684871554374695, "step": 1466 }, { "ce_loss": 4.984487532055937e-05, "cls_loss": 0.03857421875, "epoch": 0.8804804804804804, "mask_bce_loss": 0.6671496629714966, "mask_dice_loss": 0.0474063865840435, "mask_loss": 0.7145560383796692, "step": 1466 }, { "epoch": 0.8810810810810811, "grad_norm": 31.587064743041992, "learning_rate": 7.399472443022181e-08, "loss": 0.9144370555877686, "step": 1467 }, { "ce_loss": 5.600286021945067e-05, "cls_loss": 0.056640625, "epoch": 0.8810810810810811, "mask_bce_loss": 2.3294286727905273, "mask_dice_loss": 0.08150940388441086, "mask_loss": 2.410938024520874, "step": 1467 }, { "epoch": 0.8816816816816817, "grad_norm": 24.72409439086914, "learning_rate": 7.32621212250497e-08, "loss": 0.7204768657684326, "step": 1468 }, { "ce_loss": 0.0001667636533966288, "cls_loss": 0.06005859375, "epoch": 0.8816816816816817, "mask_bce_loss": 0.9405710101127625, "mask_dice_loss": 0.07764836400747299, "mask_loss": 1.0182193517684937, "step": 1468 }, { "epoch": 0.8822822822822823, "grad_norm": 23.96383285522461, "learning_rate": 7.253302482368606e-08, "loss": 0.8332866430282593, "step": 1469 }, { "ce_loss": 4.6910990931792185e-05, "cls_loss": 0.056640625, "epoch": 0.8822822822822823, "mask_bce_loss": 0.5976237654685974, "mask_dice_loss": 0.06789106875658035, "mask_loss": 0.6655148267745972, "step": 1469 }, { "epoch": 0.8828828828828829, "grad_norm": 53.108211517333984, "learning_rate": 7.180743798505318e-08, "loss": 0.7947278618812561, "step": 1470 }, { "ce_loss": 0.005938768852502108, "cls_loss": 0.05712890625, "epoch": 0.8828828828828829, "mask_bce_loss": 1.090506911277771, "mask_dice_loss": 0.09587442874908447, "mask_loss": 1.1863813400268555, "step": 1470 }, { "epoch": 0.8834834834834835, "grad_norm": 27.69301986694336, "learning_rate": 7.108536345479322e-08, "loss": 0.8693608045578003, "step": 1471 }, { "ce_loss": 0.0001825192739488557, "cls_loss": 0.044189453125, "epoch": 0.8834834834834835, "mask_bce_loss": 0.2591492533683777, "mask_dice_loss": 0.039551831781864166, "mask_loss": 0.29870107769966125, "step": 1471 }, { "epoch": 0.8840840840840841, "grad_norm": 28.68438720703125, "learning_rate": 7.036680396525797e-08, "loss": 0.8226268291473389, "step": 1472 }, { "ce_loss": 9.050319931702688e-05, "cls_loss": 0.0419921875, "epoch": 0.8840840840840841, "mask_bce_loss": 0.7398526072502136, "mask_dice_loss": 0.055422376841306686, "mask_loss": 0.7952749729156494, "step": 1472 }, { "epoch": 0.8846846846846846, "grad_norm": 21.006837844848633, "learning_rate": 6.965176223549807e-08, "loss": 0.8775839805603027, "step": 1473 }, { "ce_loss": 4.888102921540849e-05, "cls_loss": 0.038818359375, "epoch": 0.8846846846846846, "mask_bce_loss": 0.3287965953350067, "mask_dice_loss": 0.0359472781419754, "mask_loss": 0.3647438883781433, "step": 1473 }, { "epoch": 0.8852852852852853, "grad_norm": 26.449954986572266, "learning_rate": 6.89402409712525e-08, "loss": 0.8134360313415527, "step": 1474 }, { "ce_loss": 0.0002965311286970973, "cls_loss": 0.05224609375, "epoch": 0.8852852852852853, "mask_bce_loss": 0.9005023837089539, "mask_dice_loss": 0.08065340667963028, "mask_loss": 0.9811558127403259, "step": 1474 }, { "epoch": 0.8858858858858859, "grad_norm": 19.62240219116211, "learning_rate": 6.823224286493923e-08, "loss": 0.7394833564758301, "step": 1475 }, { "ce_loss": 0.00011829563300125301, "cls_loss": 0.06494140625, "epoch": 0.8858858858858859, "mask_bce_loss": 0.47094589471817017, "mask_dice_loss": 0.1044914498925209, "mask_loss": 0.5754373669624329, "step": 1475 }, { "epoch": 0.8864864864864865, "grad_norm": 38.2655029296875, "learning_rate": 6.75277705956443e-08, "loss": 0.8488963842391968, "step": 1476 }, { "ce_loss": 0.00017820709035731852, "cls_loss": 0.060546875, "epoch": 0.8864864864864865, "mask_bce_loss": 0.6477115154266357, "mask_dice_loss": 0.11024868488311768, "mask_loss": 0.7579602003097534, "step": 1476 }, { "epoch": 0.8870870870870871, "grad_norm": 20.57881736755371, "learning_rate": 6.682682682911134e-08, "loss": 0.8609458804130554, "step": 1477 }, { "ce_loss": 2.8876438591396436e-05, "cls_loss": 0.057861328125, "epoch": 0.8870870870870871, "mask_bce_loss": 0.7011016607284546, "mask_dice_loss": 0.07069584727287292, "mask_loss": 0.7717975378036499, "step": 1477 }, { "epoch": 0.8876876876876877, "grad_norm": 27.239360809326172, "learning_rate": 6.6129414217733e-08, "loss": 0.8218224048614502, "step": 1478 }, { "ce_loss": 7.946416008053347e-05, "cls_loss": 0.05615234375, "epoch": 0.8876876876876877, "mask_bce_loss": 0.48752984404563904, "mask_dice_loss": 0.10703971236944199, "mask_loss": 0.5945695638656616, "step": 1478 }, { "epoch": 0.8882882882882883, "grad_norm": 27.479137420654297, "learning_rate": 6.543553540053926e-08, "loss": 0.9849551320075989, "step": 1479 }, { "ce_loss": 4.042333966935985e-05, "cls_loss": 0.056640625, "epoch": 0.8882882882882883, "mask_bce_loss": 0.27920660376548767, "mask_dice_loss": 0.11828930675983429, "mask_loss": 0.39749592542648315, "step": 1479 }, { "epoch": 0.8888888888888888, "grad_norm": 36.49138259887695, "learning_rate": 6.474519300318803e-08, "loss": 0.7646536827087402, "step": 1480 }, { "ce_loss": 0.00029182728030718863, "cls_loss": 0.04833984375, "epoch": 0.8888888888888888, "mask_bce_loss": 1.282234787940979, "mask_dice_loss": 0.10606440156698227, "mask_loss": 1.3882992267608643, "step": 1480 }, { "epoch": 0.8894894894894895, "grad_norm": 19.531360626220703, "learning_rate": 6.405838963795595e-08, "loss": 0.7431117296218872, "step": 1481 }, { "ce_loss": 3.294753696536645e-05, "cls_loss": 0.036865234375, "epoch": 0.8894894894894895, "mask_bce_loss": 0.6357871294021606, "mask_dice_loss": 0.05292070657014847, "mask_loss": 0.6887078285217285, "step": 1481 }, { "epoch": 0.8900900900900901, "grad_norm": 27.406166076660156, "learning_rate": 6.337512790372734e-08, "loss": 0.8298591375350952, "step": 1482 }, { "ce_loss": 6.2768034695182e-05, "cls_loss": 0.06494140625, "epoch": 0.8900900900900901, "mask_bce_loss": 0.7455635070800781, "mask_dice_loss": 0.11598608642816544, "mask_loss": 0.8615496158599854, "step": 1482 }, { "epoch": 0.8906906906906907, "grad_norm": 60.570533752441406, "learning_rate": 6.26954103859847e-08, "loss": 0.8544765710830688, "step": 1483 }, { "ce_loss": 5.5600110499653965e-05, "cls_loss": 0.053955078125, "epoch": 0.8906906906906907, "mask_bce_loss": 0.426084965467453, "mask_dice_loss": 0.07789216190576553, "mask_loss": 0.5039771199226379, "step": 1483 }, { "epoch": 0.8912912912912913, "grad_norm": 29.19913673400879, "learning_rate": 6.201923965680012e-08, "loss": 0.796912431716919, "step": 1484 }, { "ce_loss": 4.7580710088368505e-05, "cls_loss": 0.052001953125, "epoch": 0.8912912912912913, "mask_bce_loss": 0.8439043164253235, "mask_dice_loss": 0.050077129155397415, "mask_loss": 0.8939814567565918, "step": 1484 }, { "epoch": 0.8918918918918919, "grad_norm": 19.8299617767334, "learning_rate": 6.134661827482313e-08, "loss": 0.8064650893211365, "step": 1485 }, { "ce_loss": 5.848078217240982e-05, "cls_loss": 0.056640625, "epoch": 0.8918918918918919, "mask_bce_loss": 0.5986294150352478, "mask_dice_loss": 0.0676342248916626, "mask_loss": 0.6662636399269104, "step": 1485 }, { "epoch": 0.8924924924924925, "grad_norm": 26.35797119140625, "learning_rate": 6.067754878527331e-08, "loss": 0.8425136208534241, "step": 1486 }, { "ce_loss": 0.00012847621110267937, "cls_loss": 0.061767578125, "epoch": 0.8924924924924925, "mask_bce_loss": 0.38128790259361267, "mask_dice_loss": 0.09892036765813828, "mask_loss": 0.48020827770233154, "step": 1486 }, { "epoch": 0.893093093093093, "grad_norm": 29.111900329589844, "learning_rate": 6.00120337199298e-08, "loss": 0.8954738974571228, "step": 1487 }, { "ce_loss": 9.019792196340859e-05, "cls_loss": 0.0478515625, "epoch": 0.893093093093093, "mask_bce_loss": 0.7665874361991882, "mask_dice_loss": 0.08745003491640091, "mask_loss": 0.8540374636650085, "step": 1487 }, { "epoch": 0.8936936936936937, "grad_norm": 26.773273468017578, "learning_rate": 5.9350075597121355e-08, "loss": 0.8328904509544373, "step": 1488 }, { "ce_loss": 0.00014074107457417995, "cls_loss": 0.05810546875, "epoch": 0.8936936936936937, "mask_bce_loss": 1.0096220970153809, "mask_dice_loss": 0.13717932999134064, "mask_loss": 1.146801471710205, "step": 1488 }, { "epoch": 0.8942942942942943, "grad_norm": 33.74367141723633, "learning_rate": 5.8691676921717056e-08, "loss": 0.7767205238342285, "step": 1489 }, { "ce_loss": 0.0003591676941141486, "cls_loss": 0.027587890625, "epoch": 0.8942942942942943, "mask_bce_loss": 0.21250246465206146, "mask_dice_loss": 0.02206135354936123, "mask_loss": 0.23456381261348724, "step": 1489 }, { "epoch": 0.8948948948948949, "grad_norm": 24.766149520874023, "learning_rate": 5.803684018511756e-08, "loss": 0.7433240413665771, "step": 1490 }, { "ce_loss": 3.160516280331649e-05, "cls_loss": 0.030029296875, "epoch": 0.8948948948948949, "mask_bce_loss": 0.4015624523162842, "mask_dice_loss": 0.0659525915980339, "mask_loss": 0.4675150513648987, "step": 1490 }, { "epoch": 0.8954954954954955, "grad_norm": 27.757640838623047, "learning_rate": 5.738556786524451e-08, "loss": 0.7778077125549316, "step": 1491 }, { "ce_loss": 0.0655263289809227, "cls_loss": 0.027587890625, "epoch": 0.8954954954954955, "mask_bce_loss": 0.3159782588481903, "mask_dice_loss": 0.019713355228304863, "mask_loss": 0.3356916010379791, "step": 1491 }, { "epoch": 0.8960960960960961, "grad_norm": 14.156587600708008, "learning_rate": 5.673786242653156e-08, "loss": 0.6520524024963379, "step": 1492 }, { "ce_loss": 0.00011153319792356342, "cls_loss": 0.04052734375, "epoch": 0.8960960960960961, "mask_bce_loss": 0.35214701294898987, "mask_dice_loss": 0.035993289202451706, "mask_loss": 0.3881402909755707, "step": 1492 }, { "epoch": 0.8966966966966967, "grad_norm": 62.07066345214844, "learning_rate": 5.6093726319915825e-08, "loss": 0.905558168888092, "step": 1493 }, { "ce_loss": 0.00012665928807109594, "cls_loss": 0.037109375, "epoch": 0.8966966966966967, "mask_bce_loss": 0.4091016948223114, "mask_dice_loss": 0.03204189985990524, "mask_loss": 0.44114360213279724, "step": 1493 }, { "epoch": 0.8972972972972973, "grad_norm": 20.260793685913086, "learning_rate": 5.5453161982827346e-08, "loss": 0.8202299475669861, "step": 1494 }, { "ce_loss": 5.6586082791909575e-05, "cls_loss": 0.0498046875, "epoch": 0.8972972972972973, "mask_bce_loss": 0.405562162399292, "mask_dice_loss": 0.12770281732082367, "mask_loss": 0.5332649946212769, "step": 1494 }, { "epoch": 0.8978978978978979, "grad_norm": 25.9221134185791, "learning_rate": 5.4816171839180526e-08, "loss": 0.8854976892471313, "step": 1495 }, { "ce_loss": 3.824204395641573e-05, "cls_loss": 0.06494140625, "epoch": 0.8978978978978979, "mask_bce_loss": 0.5540070533752441, "mask_dice_loss": 0.12281203269958496, "mask_loss": 0.6768190860748291, "step": 1495 }, { "epoch": 0.8984984984984985, "grad_norm": 27.536636352539062, "learning_rate": 5.4182758299365364e-08, "loss": 0.8292391896247864, "step": 1496 }, { "ce_loss": 6.350185140036047e-05, "cls_loss": 0.057373046875, "epoch": 0.8984984984984985, "mask_bce_loss": 0.9635055661201477, "mask_dice_loss": 0.07673230022192001, "mask_loss": 1.0402379035949707, "step": 1496 }, { "epoch": 0.8990990990990991, "grad_norm": 45.639564514160156, "learning_rate": 5.355292376023757e-08, "loss": 0.920868456363678, "step": 1497 }, { "ce_loss": 6.199390190886334e-05, "cls_loss": 0.05615234375, "epoch": 0.8990990990990991, "mask_bce_loss": 0.5660491585731506, "mask_dice_loss": 0.10024376958608627, "mask_loss": 0.6662929058074951, "step": 1497 }, { "epoch": 0.8996996996996997, "grad_norm": 16.634614944458008, "learning_rate": 5.292667060510969e-08, "loss": 0.7018892765045166, "step": 1498 }, { "ce_loss": 0.00013097978080622852, "cls_loss": 0.06005859375, "epoch": 0.8996996996996997, "mask_bce_loss": 1.2727994918823242, "mask_dice_loss": 0.1060144305229187, "mask_loss": 1.3788139820098877, "step": 1498 }, { "epoch": 0.9003003003003003, "grad_norm": 24.397146224975586, "learning_rate": 5.2304001203742566e-08, "loss": 0.7709173560142517, "step": 1499 }, { "ce_loss": 0.012713726609945297, "cls_loss": 0.0712890625, "epoch": 0.9003003003003003, "mask_bce_loss": 0.230611190199852, "mask_dice_loss": 0.10715966671705246, "mask_loss": 0.33777084946632385, "step": 1499 }, { "epoch": 0.9009009009009009, "grad_norm": 36.10843276977539, "learning_rate": 5.168491791233576e-08, "loss": 0.9643365740776062, "step": 1500 }, { "ce_loss": 5.112071812618524e-05, "cls_loss": 0.0498046875, "epoch": 0.9009009009009009, "mask_bce_loss": 0.5317806601524353, "mask_dice_loss": 0.05231615528464317, "mask_loss": 0.5840967893600464, "step": 1500 }, { "epoch": 0.9015015015015015, "grad_norm": 22.72440528869629, "learning_rate": 5.1069423073518805e-08, "loss": 0.8544532656669617, "step": 1501 }, { "ce_loss": 0.01964358612895012, "cls_loss": 0.068359375, "epoch": 0.9015015015015015, "mask_bce_loss": 1.3447977304458618, "mask_dice_loss": 0.10864772647619247, "mask_loss": 1.4534454345703125, "step": 1501 }, { "epoch": 0.9021021021021021, "grad_norm": 41.76002502441406, "learning_rate": 5.0457519016342875e-08, "loss": 0.7679380774497986, "step": 1502 }, { "ce_loss": 6.309741002041847e-05, "cls_loss": 0.05126953125, "epoch": 0.9021021021021021, "mask_bce_loss": 0.4748557209968567, "mask_dice_loss": 0.06769328564405441, "mask_loss": 0.5425490140914917, "step": 1502 }, { "epoch": 0.9027027027027027, "grad_norm": 22.325716018676758, "learning_rate": 4.98492080562708e-08, "loss": 0.7330664396286011, "step": 1503 }, { "ce_loss": 4.416501906234771e-05, "cls_loss": 0.052001953125, "epoch": 0.9027027027027027, "mask_bce_loss": 0.4092044532299042, "mask_dice_loss": 0.07690097391605377, "mask_loss": 0.48610544204711914, "step": 1503 }, { "epoch": 0.9033033033033033, "grad_norm": 20.715564727783203, "learning_rate": 4.9244492495169484e-08, "loss": 0.6801574230194092, "step": 1504 }, { "ce_loss": 9.306106949225068e-05, "cls_loss": 0.037109375, "epoch": 0.9033033033033033, "mask_bce_loss": 0.3893992006778717, "mask_dice_loss": 0.03338026627898216, "mask_loss": 0.42277947068214417, "step": 1504 }, { "epoch": 0.9039039039039038, "grad_norm": 31.760190963745117, "learning_rate": 4.864337462130097e-08, "loss": 0.7656285166740417, "step": 1505 }, { "ce_loss": 5.8829249610425904e-05, "cls_loss": 0.04248046875, "epoch": 0.9039039039039038, "mask_bce_loss": 0.5591796636581421, "mask_dice_loss": 0.09759991616010666, "mask_loss": 0.6567795872688293, "step": 1505 }, { "epoch": 0.9045045045045045, "grad_norm": 22.103443145751953, "learning_rate": 4.804585670931294e-08, "loss": 0.8161717653274536, "step": 1506 }, { "ce_loss": 0.017073003575205803, "cls_loss": 0.1064453125, "epoch": 0.9045045045045045, "mask_bce_loss": 1.2049061059951782, "mask_dice_loss": 0.1270868480205536, "mask_loss": 1.3319929838180542, "step": 1506 }, { "epoch": 0.9051051051051051, "grad_norm": 39.75365447998047, "learning_rate": 4.7451941020230976e-08, "loss": 0.7380903959274292, "step": 1507 }, { "ce_loss": 0.001270450185984373, "cls_loss": 0.05712890625, "epoch": 0.9051051051051051, "mask_bce_loss": 1.2911540269851685, "mask_dice_loss": 0.15444521605968475, "mask_loss": 1.4455991983413696, "step": 1507 }, { "epoch": 0.9057057057057057, "grad_norm": 30.35955238342285, "learning_rate": 4.6861629801450144e-08, "loss": 0.6802016496658325, "step": 1508 }, { "ce_loss": 5.9858266467927024e-05, "cls_loss": 0.043701171875, "epoch": 0.9057057057057057, "mask_bce_loss": 0.4340159595012665, "mask_dice_loss": 0.043502677232027054, "mask_loss": 0.47751864790916443, "step": 1508 }, { "epoch": 0.9063063063063063, "grad_norm": 29.87144660949707, "learning_rate": 4.6274925286725386e-08, "loss": 0.6867000460624695, "step": 1509 }, { "ce_loss": 3.604713492677547e-05, "cls_loss": 0.043701171875, "epoch": 0.9063063063063063, "mask_bce_loss": 0.5156400203704834, "mask_dice_loss": 0.0551295280456543, "mask_loss": 0.5707695484161377, "step": 1509 }, { "epoch": 0.9069069069069069, "grad_norm": 19.008201599121094, "learning_rate": 4.5691829696164254e-08, "loss": 0.7450835108757019, "step": 1510 }, { "ce_loss": 2.935484008048661e-05, "cls_loss": 0.05126953125, "epoch": 0.9069069069069069, "mask_bce_loss": 0.5523175597190857, "mask_dice_loss": 0.05623403936624527, "mask_loss": 0.6085516214370728, "step": 1510 }, { "epoch": 0.9075075075075075, "grad_norm": 15.659805297851562, "learning_rate": 4.511234523621799e-08, "loss": 0.6383728981018066, "step": 1511 }, { "ce_loss": 6.982875493122265e-05, "cls_loss": 0.030517578125, "epoch": 0.9075075075075075, "mask_bce_loss": 0.5109832882881165, "mask_dice_loss": 0.02446889691054821, "mask_loss": 0.5354521870613098, "step": 1511 }, { "epoch": 0.9081081081081082, "grad_norm": 36.97061538696289, "learning_rate": 4.453647409967287e-08, "loss": 0.9768787622451782, "step": 1512 }, { "ce_loss": 5.884995334781706e-05, "cls_loss": 0.05712890625, "epoch": 0.9081081081081082, "mask_bce_loss": 0.558671772480011, "mask_dice_loss": 0.07150488346815109, "mask_loss": 0.6301766633987427, "step": 1512 }, { "epoch": 0.9087087087087087, "grad_norm": 47.74964904785156, "learning_rate": 4.396421846564235e-08, "loss": 0.8794848918914795, "step": 1513 }, { "ce_loss": 8.20883215055801e-05, "cls_loss": 0.05859375, "epoch": 0.9087087087087087, "mask_bce_loss": 0.8648447394371033, "mask_dice_loss": 0.14646589756011963, "mask_loss": 1.0113105773925781, "step": 1513 }, { "epoch": 0.9093093093093093, "grad_norm": 35.92013168334961, "learning_rate": 4.339558049955927e-08, "loss": 0.7940784692764282, "step": 1514 }, { "ce_loss": 4.929930219077505e-05, "cls_loss": 0.04931640625, "epoch": 0.9093093093093093, "mask_bce_loss": 0.3219853639602661, "mask_dice_loss": 0.054804395884275436, "mask_loss": 0.37678974866867065, "step": 1514 }, { "epoch": 0.9099099099099099, "grad_norm": 15.955143928527832, "learning_rate": 4.28305623531664e-08, "loss": 0.6999301910400391, "step": 1515 }, { "ce_loss": 0.0006027781637385488, "cls_loss": 0.057861328125, "epoch": 0.9099099099099099, "mask_bce_loss": 1.1686530113220215, "mask_dice_loss": 0.13322018086910248, "mask_loss": 1.3018732070922852, "step": 1515 }, { "epoch": 0.9105105105105105, "grad_norm": 40.47359848022461, "learning_rate": 4.226916616450915e-08, "loss": 0.7374730706214905, "step": 1516 }, { "ce_loss": 7.317030394915491e-05, "cls_loss": 0.0634765625, "epoch": 0.9105105105105105, "mask_bce_loss": 0.6775946617126465, "mask_dice_loss": 0.10905170440673828, "mask_loss": 0.7866463661193848, "step": 1516 }, { "epoch": 0.9111111111111111, "grad_norm": 20.342050552368164, "learning_rate": 4.17113940579279e-08, "loss": 0.8216186761856079, "step": 1517 }, { "ce_loss": 8.978542609838769e-05, "cls_loss": 0.030029296875, "epoch": 0.9111111111111111, "mask_bce_loss": 0.5075815916061401, "mask_dice_loss": 0.03901445493102074, "mask_loss": 0.5465960502624512, "step": 1517 }, { "epoch": 0.9117117117117117, "grad_norm": 32.023441314697266, "learning_rate": 4.115724814404853e-08, "loss": 0.8200154900550842, "step": 1518 }, { "ce_loss": 5.1895818614866585e-05, "cls_loss": 0.040283203125, "epoch": 0.9117117117117117, "mask_bce_loss": 0.3679666817188263, "mask_dice_loss": 0.05918074771761894, "mask_loss": 0.42714741826057434, "step": 1518 }, { "epoch": 0.9123123123123124, "grad_norm": 38.506038665771484, "learning_rate": 4.060673051977603e-08, "loss": 0.9147262573242188, "step": 1519 }, { "ce_loss": 5.257935481495224e-05, "cls_loss": 0.0908203125, "epoch": 0.9123123123123124, "mask_bce_loss": 0.9965611696243286, "mask_dice_loss": 0.09256532043218613, "mask_loss": 1.089126467704773, "step": 1519 }, { "epoch": 0.9129129129129129, "grad_norm": 39.04480743408203, "learning_rate": 4.005984326828527e-08, "loss": 0.826669454574585, "step": 1520 }, { "ce_loss": 0.0007449384429492056, "cls_loss": 0.04736328125, "epoch": 0.9129129129129129, "mask_bce_loss": 1.2004377841949463, "mask_dice_loss": 0.06249229237437248, "mask_loss": 1.2629300355911255, "step": 1520 }, { "epoch": 0.9135135135135135, "grad_norm": 41.59500503540039, "learning_rate": 3.951658845901418e-08, "loss": 0.9849637746810913, "step": 1521 }, { "ce_loss": 9.232106822310016e-05, "cls_loss": 0.05078125, "epoch": 0.9135135135135135, "mask_bce_loss": 1.0043367147445679, "mask_dice_loss": 0.07588615268468857, "mask_loss": 1.0802228450775146, "step": 1521 }, { "epoch": 0.9141141141141141, "grad_norm": 15.334593772888184, "learning_rate": 3.8976968147654944e-08, "loss": 0.7528819441795349, "step": 1522 }, { "ce_loss": 6.372362986439839e-05, "cls_loss": 0.052001953125, "epoch": 0.9141141141141141, "mask_bce_loss": 0.49921485781669617, "mask_dice_loss": 0.06320616602897644, "mask_loss": 0.5624210238456726, "step": 1522 }, { "epoch": 0.9147147147147147, "grad_norm": 22.229686737060547, "learning_rate": 3.844098437614707e-08, "loss": 0.7750124335289001, "step": 1523 }, { "ce_loss": 9.249179129255936e-05, "cls_loss": 0.057861328125, "epoch": 0.9147147147147147, "mask_bce_loss": 0.7786630988121033, "mask_dice_loss": 0.1299050897359848, "mask_loss": 0.9085682034492493, "step": 1523 }, { "epoch": 0.9153153153153153, "grad_norm": 29.177949905395508, "learning_rate": 3.7908639172669175e-08, "loss": 0.7665049433708191, "step": 1524 }, { "ce_loss": 3.0305322070489638e-05, "cls_loss": 0.037841796875, "epoch": 0.9153153153153153, "mask_bce_loss": 0.25757452845573425, "mask_dice_loss": 0.055811382830142975, "mask_loss": 0.31338590383529663, "step": 1524 }, { "epoch": 0.9159159159159159, "grad_norm": 32.75572204589844, "learning_rate": 3.737993455163135e-08, "loss": 0.957131028175354, "step": 1525 }, { "ce_loss": 4.29169085691683e-05, "cls_loss": 0.06689453125, "epoch": 0.9159159159159159, "mask_bce_loss": 0.5184838771820068, "mask_dice_loss": 0.0925060361623764, "mask_loss": 0.6109899282455444, "step": 1525 }, { "epoch": 0.9165165165165166, "grad_norm": 25.25627899169922, "learning_rate": 3.68548725136677e-08, "loss": 0.6753145456314087, "step": 1526 }, { "ce_loss": 0.0002863250847440213, "cls_loss": 0.039794921875, "epoch": 0.9165165165165166, "mask_bce_loss": 0.3119577467441559, "mask_dice_loss": 0.0337664857506752, "mask_loss": 0.3457242250442505, "step": 1526 }, { "epoch": 0.9171171171171171, "grad_norm": 46.28623580932617, "learning_rate": 3.633345504562868e-08, "loss": 0.780651330947876, "step": 1527 }, { "ce_loss": 4.7749304940225556e-05, "cls_loss": 0.053466796875, "epoch": 0.9171171171171171, "mask_bce_loss": 1.3822101354599, "mask_dice_loss": 0.07438760995864868, "mask_loss": 1.4565978050231934, "step": 1527 }, { "epoch": 0.9177177177177177, "grad_norm": 34.47573471069336, "learning_rate": 3.5815684120573565e-08, "loss": 0.7626335620880127, "step": 1528 }, { "ce_loss": 2.8721800845232792e-05, "cls_loss": 0.05859375, "epoch": 0.9177177177177177, "mask_bce_loss": 0.5798426866531372, "mask_dice_loss": 0.10446671396493912, "mask_loss": 0.6843094229698181, "step": 1528 }, { "epoch": 0.9183183183183183, "grad_norm": 48.87440872192383, "learning_rate": 3.5301561697762994e-08, "loss": 0.8881040811538696, "step": 1529 }, { "ce_loss": 2.9309807359823026e-05, "cls_loss": 0.040771484375, "epoch": 0.9183183183183183, "mask_bce_loss": 0.2676038146018982, "mask_dice_loss": 0.037094563245773315, "mask_loss": 0.3046983778476715, "step": 1529 }, { "epoch": 0.918918918918919, "grad_norm": 23.128887176513672, "learning_rate": 3.4791089722651435e-08, "loss": 0.6570672392845154, "step": 1530 }, { "ce_loss": 3.6177574656903744e-05, "cls_loss": 0.05126953125, "epoch": 0.918918918918919, "mask_bce_loss": 0.541578471660614, "mask_dice_loss": 0.06338665634393692, "mask_loss": 0.6049651503562927, "step": 1530 }, { "epoch": 0.9195195195195195, "grad_norm": 68.14988708496094, "learning_rate": 3.428427012688007e-08, "loss": 0.9235873222351074, "step": 1531 }, { "ce_loss": 0.00012907171912956983, "cls_loss": 0.0693359375, "epoch": 0.9195195195195195, "mask_bce_loss": 0.633693516254425, "mask_dice_loss": 0.077341228723526, "mask_loss": 0.7110347747802734, "step": 1531 }, { "epoch": 0.9201201201201201, "grad_norm": 21.21222496032715, "learning_rate": 3.378110482826968e-08, "loss": 0.871117353439331, "step": 1532 }, { "ce_loss": 0.00010685437155188993, "cls_loss": 0.027099609375, "epoch": 0.9201201201201201, "mask_bce_loss": 0.2643570601940155, "mask_dice_loss": 0.027939412742853165, "mask_loss": 0.29229646921157837, "step": 1532 }, { "epoch": 0.9207207207207208, "grad_norm": 22.78925323486328, "learning_rate": 3.3281595730812575e-08, "loss": 0.8108116388320923, "step": 1533 }, { "ce_loss": 0.00014729602844454348, "cls_loss": 0.03662109375, "epoch": 0.9207207207207208, "mask_bce_loss": 0.22479262948036194, "mask_dice_loss": 0.03151752054691315, "mask_loss": 0.2563101649284363, "step": 1533 }, { "epoch": 0.9213213213213213, "grad_norm": 22.249422073364258, "learning_rate": 3.2785744724666e-08, "loss": 0.8073433041572571, "step": 1534 }, { "ce_loss": 0.0001508357236161828, "cls_loss": 0.059326171875, "epoch": 0.9213213213213213, "mask_bce_loss": 0.632057785987854, "mask_dice_loss": 0.06667758524417877, "mask_loss": 0.6987353563308716, "step": 1534 }, { "epoch": 0.9219219219219219, "grad_norm": 21.76515007019043, "learning_rate": 3.2293553686145283e-08, "loss": 0.8532515168190002, "step": 1535 }, { "ce_loss": 3.329514220240526e-05, "cls_loss": 0.03564453125, "epoch": 0.9219219219219219, "mask_bce_loss": 0.29216626286506653, "mask_dice_loss": 0.07303894311189651, "mask_loss": 0.36520519852638245, "step": 1535 }, { "epoch": 0.9225225225225225, "grad_norm": 31.040925979614258, "learning_rate": 3.180502447771572e-08, "loss": 0.6860508918762207, "step": 1536 }, { "ce_loss": 0.00025506707606837153, "cls_loss": 0.04833984375, "epoch": 0.9225225225225225, "mask_bce_loss": 0.6926127672195435, "mask_dice_loss": 0.05702151730656624, "mask_loss": 0.7496342658996582, "step": 1536 }, { "epoch": 0.9231231231231232, "grad_norm": 29.11676597595215, "learning_rate": 3.1320158947986565e-08, "loss": 0.7651721835136414, "step": 1537 }, { "ce_loss": 8.10171986813657e-05, "cls_loss": 0.04443359375, "epoch": 0.9231231231231232, "mask_bce_loss": 0.4308958947658539, "mask_dice_loss": 0.042818207293748856, "mask_loss": 0.47371411323547363, "step": 1537 }, { "epoch": 0.9237237237237237, "grad_norm": 42.03389358520508, "learning_rate": 3.083895893170352e-08, "loss": 0.8197970986366272, "step": 1538 }, { "ce_loss": 3.9316746551776305e-05, "cls_loss": 0.057861328125, "epoch": 0.9237237237237237, "mask_bce_loss": 0.604704737663269, "mask_dice_loss": 0.13407132029533386, "mask_loss": 0.7387760877609253, "step": 1538 }, { "epoch": 0.9243243243243243, "grad_norm": 15.945862770080566, "learning_rate": 3.036142624974158e-08, "loss": 0.8274660110473633, "step": 1539 }, { "ce_loss": 6.735159695381299e-05, "cls_loss": 0.055419921875, "epoch": 0.9243243243243243, "mask_bce_loss": 0.7736730575561523, "mask_dice_loss": 0.0868002250790596, "mask_loss": 0.8604732751846313, "step": 1539 }, { "epoch": 0.924924924924925, "grad_norm": 53.96303939819336, "learning_rate": 2.9887562709098756e-08, "loss": 0.8859108686447144, "step": 1540 }, { "ce_loss": 0.0002523729926906526, "cls_loss": 0.052001953125, "epoch": 0.924924924924925, "mask_bce_loss": 1.4678267240524292, "mask_dice_loss": 0.1499355584383011, "mask_loss": 1.6177623271942139, "step": 1540 }, { "epoch": 0.9255255255255255, "grad_norm": 36.456504821777344, "learning_rate": 2.9417370102889026e-08, "loss": 0.7440961003303528, "step": 1541 }, { "ce_loss": 3.3006825105985627e-05, "cls_loss": 0.060546875, "epoch": 0.9255255255255255, "mask_bce_loss": 0.7691754102706909, "mask_dice_loss": 0.11629237979650497, "mask_loss": 0.8854677677154541, "step": 1541 }, { "epoch": 0.9261261261261261, "grad_norm": 60.882728576660156, "learning_rate": 2.8950850210335054e-08, "loss": 0.7301994562149048, "step": 1542 }, { "ce_loss": 3.0758859793422744e-05, "cls_loss": 0.06201171875, "epoch": 0.9261261261261261, "mask_bce_loss": 0.5935294032096863, "mask_dice_loss": 0.09749650955200195, "mask_loss": 0.6910259127616882, "step": 1542 }, { "epoch": 0.9267267267267267, "grad_norm": 43.94486618041992, "learning_rate": 2.8488004796762056e-08, "loss": 0.795333743095398, "step": 1543 }, { "ce_loss": 4.551565143628977e-05, "cls_loss": 0.06982421875, "epoch": 0.9267267267267267, "mask_bce_loss": 1.1943305730819702, "mask_dice_loss": 0.1554306000471115, "mask_loss": 1.3497611284255981, "step": 1543 }, { "epoch": 0.9273273273273274, "grad_norm": 30.56834602355957, "learning_rate": 2.802883561359104e-08, "loss": 0.8282744884490967, "step": 1544 }, { "ce_loss": 9.642812074162066e-05, "cls_loss": 0.05908203125, "epoch": 0.9273273273273274, "mask_bce_loss": 1.1248102188110352, "mask_dice_loss": 0.10480425506830215, "mask_loss": 1.229614496231079, "step": 1544 }, { "epoch": 0.9279279279279279, "grad_norm": 36.64910125732422, "learning_rate": 2.7573344398331923e-08, "loss": 0.8055034875869751, "step": 1545 }, { "ce_loss": 0.00012349382450338453, "cls_loss": 0.03369140625, "epoch": 0.9279279279279279, "mask_bce_loss": 0.5057388544082642, "mask_dice_loss": 0.03725442290306091, "mask_loss": 0.5429933071136475, "step": 1545 }, { "epoch": 0.9285285285285285, "grad_norm": 25.841278076171875, "learning_rate": 2.7121532874576857e-08, "loss": 0.7745062708854675, "step": 1546 }, { "ce_loss": 0.0017640716396272182, "cls_loss": 0.05615234375, "epoch": 0.9285285285285285, "mask_bce_loss": 0.7181331515312195, "mask_dice_loss": 0.0782233253121376, "mask_loss": 0.7963564991950989, "step": 1546 }, { "epoch": 0.9291291291291291, "grad_norm": 36.0497932434082, "learning_rate": 2.6673402751994255e-08, "loss": 0.8510030508041382, "step": 1547 }, { "ce_loss": 0.0003987188683822751, "cls_loss": 0.036376953125, "epoch": 0.9291291291291291, "mask_bce_loss": 0.4745262563228607, "mask_dice_loss": 0.07107516378164291, "mask_loss": 0.5456014275550842, "step": 1547 }, { "epoch": 0.9297297297297298, "grad_norm": 15.765971183776855, "learning_rate": 2.622895572632189e-08, "loss": 0.7438912391662598, "step": 1548 }, { "ce_loss": 8.925548172555864e-05, "cls_loss": 0.0546875, "epoch": 0.9297297297297298, "mask_bce_loss": 1.5923882722854614, "mask_dice_loss": 0.11383042484521866, "mask_loss": 1.7062187194824219, "step": 1548 }, { "epoch": 0.9303303303303303, "grad_norm": 120.87538146972656, "learning_rate": 2.5788193479360454e-08, "loss": 0.7210407853126526, "step": 1549 }, { "ce_loss": 5.1691782573470846e-05, "cls_loss": 0.06201171875, "epoch": 0.9303303303303303, "mask_bce_loss": 0.46359825134277344, "mask_dice_loss": 0.08088832348585129, "mask_loss": 0.5444865822792053, "step": 1549 }, { "epoch": 0.9309309309309309, "grad_norm": 30.982898712158203, "learning_rate": 2.535111767896769e-08, "loss": 0.8216704726219177, "step": 1550 }, { "ce_loss": 7.135787745937705e-05, "cls_loss": 0.04931640625, "epoch": 0.9309309309309309, "mask_bce_loss": 0.3847613036632538, "mask_dice_loss": 0.07519594579935074, "mask_loss": 0.4599572420120239, "step": 1550 }, { "epoch": 0.9315315315315316, "grad_norm": 27.38432502746582, "learning_rate": 2.4917729979051395e-08, "loss": 0.8204172849655151, "step": 1551 }, { "ce_loss": 5.2527910156641155e-05, "cls_loss": 0.0181884765625, "epoch": 0.9315315315315316, "mask_bce_loss": 0.2759120464324951, "mask_dice_loss": 0.011933499947190285, "mask_loss": 0.28784555196762085, "step": 1551 }, { "epoch": 0.9321321321321321, "grad_norm": 34.893768310546875, "learning_rate": 2.4488032019563398e-08, "loss": 0.8969225883483887, "step": 1552 }, { "ce_loss": 6.560191104654223e-05, "cls_loss": 0.055908203125, "epoch": 0.9321321321321321, "mask_bce_loss": 0.504584789276123, "mask_dice_loss": 0.11571375280618668, "mask_loss": 0.6202985644340515, "step": 1552 }, { "epoch": 0.9327327327327327, "grad_norm": 26.789012908935547, "learning_rate": 2.4062025426493714e-08, "loss": 0.8798518180847168, "step": 1553 }, { "ce_loss": 0.00014738665777258575, "cls_loss": 0.05126953125, "epoch": 0.9327327327327327, "mask_bce_loss": 0.9346825480461121, "mask_dice_loss": 0.0803920328617096, "mask_loss": 1.015074610710144, "step": 1553 }, { "epoch": 0.9333333333333333, "grad_norm": 44.035560607910156, "learning_rate": 2.3639711811864195e-08, "loss": 0.8078632354736328, "step": 1554 }, { "ce_loss": 7.664029544685036e-05, "cls_loss": 0.0625, "epoch": 0.9333333333333333, "mask_bce_loss": 0.7338165640830994, "mask_dice_loss": 0.1336192637681961, "mask_loss": 0.8674358129501343, "step": 1554 }, { "epoch": 0.933933933933934, "grad_norm": 30.321685791015625, "learning_rate": 2.3221092773721974e-08, "loss": 0.6330522894859314, "step": 1555 }, { "ce_loss": 0.0007776633137837052, "cls_loss": 0.059326171875, "epoch": 0.933933933933934, "mask_bce_loss": 1.1675046682357788, "mask_dice_loss": 0.1364275962114334, "mask_loss": 1.3039323091506958, "step": 1555 }, { "epoch": 0.9345345345345345, "grad_norm": 23.951297760009766, "learning_rate": 2.280616989613404e-08, "loss": 0.7052963972091675, "step": 1556 }, { "ce_loss": 6.263786053750664e-05, "cls_loss": 0.057861328125, "epoch": 0.9345345345345345, "mask_bce_loss": 0.7326751947402954, "mask_dice_loss": 0.07089219987392426, "mask_loss": 0.8035674095153809, "step": 1556 }, { "epoch": 0.9351351351351351, "grad_norm": 21.98650360107422, "learning_rate": 2.239494474918091e-08, "loss": 0.9100497961044312, "step": 1557 }, { "ce_loss": 0.0005657183355651796, "cls_loss": 0.061767578125, "epoch": 0.9351351351351351, "mask_bce_loss": 1.4049500226974487, "mask_dice_loss": 0.1315498799085617, "mask_loss": 1.5364998579025269, "step": 1557 }, { "epoch": 0.9357357357357358, "grad_norm": 49.80411911010742, "learning_rate": 2.1987418888950726e-08, "loss": 0.8463922142982483, "step": 1558 }, { "ce_loss": 0.00012276340567041188, "cls_loss": 0.038818359375, "epoch": 0.9357357357357358, "mask_bce_loss": 0.39654651284217834, "mask_dice_loss": 0.06693916767835617, "mask_loss": 0.4634856879711151, "step": 1558 }, { "epoch": 0.9363363363363363, "grad_norm": 23.959348678588867, "learning_rate": 2.1583593857533832e-08, "loss": 0.7880276441574097, "step": 1559 }, { "ce_loss": 4.141007593716495e-05, "cls_loss": 0.04833984375, "epoch": 0.9363363363363363, "mask_bce_loss": 0.5821470022201538, "mask_dice_loss": 0.04459959268569946, "mask_loss": 0.6267465949058533, "step": 1559 }, { "epoch": 0.9369369369369369, "grad_norm": 21.28049087524414, "learning_rate": 2.1183471183015887e-08, "loss": 0.7376804351806641, "step": 1560 }, { "ce_loss": 8.482321572955698e-05, "cls_loss": 0.041259765625, "epoch": 0.9369369369369369, "mask_bce_loss": 0.5149620175361633, "mask_dice_loss": 0.041787952184677124, "mask_loss": 0.5567499399185181, "step": 1560 }, { "epoch": 0.9375375375375375, "grad_norm": 30.845626831054688, "learning_rate": 2.0787052379473202e-08, "loss": 0.7151897549629211, "step": 1561 }, { "ce_loss": 3.235182157368399e-05, "cls_loss": 0.04541015625, "epoch": 0.9375375375375375, "mask_bce_loss": 0.4815486967563629, "mask_dice_loss": 0.08036940544843674, "mask_loss": 0.5619180798530579, "step": 1561 }, { "epoch": 0.9381381381381382, "grad_norm": 17.26018714904785, "learning_rate": 2.0394338946966294e-08, "loss": 0.7610337138175964, "step": 1562 }, { "ce_loss": 0.008221813477575779, "cls_loss": 0.0654296875, "epoch": 0.9381381381381382, "mask_bce_loss": 0.95159512758255, "mask_dice_loss": 0.09893514961004257, "mask_loss": 1.0505303144454956, "step": 1562 }, { "epoch": 0.9387387387387387, "grad_norm": 51.759639739990234, "learning_rate": 2.0005332371534677e-08, "loss": 0.7311298847198486, "step": 1563 }, { "ce_loss": 0.017826983705163002, "cls_loss": 0.06591796875, "epoch": 0.9387387387387387, "mask_bce_loss": 1.116805911064148, "mask_dice_loss": 0.16303656995296478, "mask_loss": 1.279842495918274, "step": 1563 }, { "epoch": 0.9393393393393393, "grad_norm": 47.38961410522461, "learning_rate": 1.9620034125190643e-08, "loss": 0.8524361848831177, "step": 1564 }, { "ce_loss": 6.274530460359529e-05, "cls_loss": 0.0419921875, "epoch": 0.9393393393393393, "mask_bce_loss": 0.39635273814201355, "mask_dice_loss": 0.049489643424749374, "mask_loss": 0.4458423852920532, "step": 1564 }, { "epoch": 0.93993993993994, "grad_norm": 26.860157012939453, "learning_rate": 1.9238445665914482e-08, "loss": 0.745236873626709, "step": 1565 }, { "ce_loss": 0.00010000413021771237, "cls_loss": 0.0250244140625, "epoch": 0.93993993993994, "mask_bce_loss": 0.1783696860074997, "mask_dice_loss": 0.03986528888344765, "mask_loss": 0.21823497116565704, "step": 1565 }, { "epoch": 0.9405405405405406, "grad_norm": 18.761964797973633, "learning_rate": 1.886056843764805e-08, "loss": 0.7951424717903137, "step": 1566 }, { "ce_loss": 4.1073635657085106e-05, "cls_loss": 0.057373046875, "epoch": 0.9405405405405406, "mask_bce_loss": 0.47098469734191895, "mask_dice_loss": 0.0837055966258049, "mask_loss": 0.5546903014183044, "step": 1566 }, { "epoch": 0.9411411411411411, "grad_norm": 91.85110473632812, "learning_rate": 1.8486403870290212e-08, "loss": 0.8985646367073059, "step": 1567 }, { "ce_loss": 8.864192204782739e-05, "cls_loss": 0.048583984375, "epoch": 0.9411411411411411, "mask_bce_loss": 0.6857842803001404, "mask_dice_loss": 0.047277357429265976, "mask_loss": 0.7330616116523743, "step": 1567 }, { "epoch": 0.9417417417417417, "grad_norm": 19.725610733032227, "learning_rate": 1.8115953379690853e-08, "loss": 0.7293225526809692, "step": 1568 }, { "ce_loss": 4.271792568033561e-05, "cls_loss": 0.0546875, "epoch": 0.9417417417417417, "mask_bce_loss": 0.6501469016075134, "mask_dice_loss": 0.05414588004350662, "mask_loss": 0.7042927742004395, "step": 1568 }, { "epoch": 0.9423423423423424, "grad_norm": 32.87376022338867, "learning_rate": 1.7749218367645756e-08, "loss": 0.824661135673523, "step": 1569 }, { "ce_loss": 4.213268402963877e-05, "cls_loss": 0.03515625, "epoch": 0.9423423423423424, "mask_bce_loss": 0.39959266781806946, "mask_dice_loss": 0.029557039961218834, "mask_loss": 0.42914971709251404, "step": 1569 }, { "epoch": 0.9429429429429429, "grad_norm": 38.26142120361328, "learning_rate": 1.7386200221890966e-08, "loss": 0.7840945720672607, "step": 1570 }, { "ce_loss": 0.00013150200538802892, "cls_loss": 0.0260009765625, "epoch": 0.9429429429429429, "mask_bce_loss": 0.3400367498397827, "mask_dice_loss": 0.03959008678793907, "mask_loss": 0.3796268403530121, "step": 1570 }, { "epoch": 0.9435435435435435, "grad_norm": 27.650007247924805, "learning_rate": 1.7026900316098212e-08, "loss": 0.9100169539451599, "step": 1571 }, { "ce_loss": 0.0001471090508857742, "cls_loss": 0.03759765625, "epoch": 0.9435435435435435, "mask_bce_loss": 0.6810230612754822, "mask_dice_loss": 0.053782492876052856, "mask_loss": 0.7348055839538574, "step": 1571 }, { "epoch": 0.9441441441441442, "grad_norm": 15.049201011657715, "learning_rate": 1.667132000986904e-08, "loss": 0.7340018153190613, "step": 1572 }, { "ce_loss": 9.83338468358852e-05, "cls_loss": 0.05810546875, "epoch": 0.9441441441441442, "mask_bce_loss": 1.1060285568237305, "mask_dice_loss": 0.1085146814584732, "mask_loss": 1.2145432233810425, "step": 1572 }, { "epoch": 0.9447447447447448, "grad_norm": 35.575531005859375, "learning_rate": 1.63194606487298e-08, "loss": 0.9791551828384399, "step": 1573 }, { "ce_loss": 8.615489059593529e-05, "cls_loss": 0.04296875, "epoch": 0.9447447447447448, "mask_bce_loss": 0.6676496267318726, "mask_dice_loss": 0.10208622366189957, "mask_loss": 0.7697358727455139, "step": 1573 }, { "epoch": 0.9453453453453453, "grad_norm": 39.485191345214844, "learning_rate": 1.5971323564126893e-08, "loss": 0.7201567888259888, "step": 1574 }, { "ce_loss": 7.72428757045418e-05, "cls_loss": 0.0308837890625, "epoch": 0.9453453453453453, "mask_bce_loss": 0.2661878764629364, "mask_dice_loss": 0.023349203169345856, "mask_loss": 0.28953707218170166, "step": 1574 }, { "epoch": 0.9459459459459459, "grad_norm": 28.714553833007812, "learning_rate": 1.5626910073421427e-08, "loss": 0.8244661092758179, "step": 1575 }, { "ce_loss": 0.00012201982463011518, "cls_loss": 0.056640625, "epoch": 0.9459459459459459, "mask_bce_loss": 0.40219759941101074, "mask_dice_loss": 0.16290400922298431, "mask_loss": 0.5651016235351562, "step": 1575 }, { "epoch": 0.9465465465465466, "grad_norm": 35.77933120727539, "learning_rate": 1.5286221479884342e-08, "loss": 0.811741828918457, "step": 1576 }, { "ce_loss": 5.6791879615047947e-05, "cls_loss": 0.033935546875, "epoch": 0.9465465465465466, "mask_bce_loss": 0.3595145046710968, "mask_dice_loss": 0.028562655672430992, "mask_loss": 0.38807716965675354, "step": 1576 }, { "epoch": 0.9471471471471471, "grad_norm": 39.16961669921875, "learning_rate": 1.4949259072691623e-08, "loss": 0.8317311406135559, "step": 1577 }, { "ce_loss": 0.0004405466897878796, "cls_loss": 0.04345703125, "epoch": 0.9471471471471471, "mask_bce_loss": 0.6730270385742188, "mask_dice_loss": 0.07646065205335617, "mask_loss": 0.7494876980781555, "step": 1577 }, { "epoch": 0.9477477477477477, "grad_norm": 23.309368133544922, "learning_rate": 1.4616024126918758e-08, "loss": 0.8379514217376709, "step": 1578 }, { "ce_loss": 5.196661368245259e-05, "cls_loss": 0.04248046875, "epoch": 0.9477477477477477, "mask_bce_loss": 0.31446656584739685, "mask_dice_loss": 0.06670059263706207, "mask_loss": 0.3811671733856201, "step": 1578 }, { "epoch": 0.9483483483483484, "grad_norm": 44.04023742675781, "learning_rate": 1.4286517903536966e-08, "loss": 0.7881731986999512, "step": 1579 }, { "ce_loss": 9.152097482001409e-05, "cls_loss": 0.030517578125, "epoch": 0.9483483483483484, "mask_bce_loss": 0.22404395043849945, "mask_dice_loss": 0.02350161224603653, "mask_loss": 0.24754557013511658, "step": 1579 }, { "epoch": 0.948948948948949, "grad_norm": 20.518329620361328, "learning_rate": 1.3960741649407527e-08, "loss": 0.7387527227401733, "step": 1580 }, { "ce_loss": 0.0001042051735566929, "cls_loss": 0.0673828125, "epoch": 0.948948948948949, "mask_bce_loss": 1.2015630006790161, "mask_dice_loss": 0.08483012765645981, "mask_loss": 1.286393165588379, "step": 1580 }, { "epoch": 0.9495495495495495, "grad_norm": 30.15326499938965, "learning_rate": 1.3638696597277677e-08, "loss": 0.8919235467910767, "step": 1581 }, { "ce_loss": 0.00011179089779034257, "cls_loss": 0.0299072265625, "epoch": 0.9495495495495495, "mask_bce_loss": 0.4027073085308075, "mask_dice_loss": 0.04544064402580261, "mask_loss": 0.4481479525566101, "step": 1581 }, { "epoch": 0.9501501501501501, "grad_norm": 27.248502731323242, "learning_rate": 1.3320383965775284e-08, "loss": 0.7970267534255981, "step": 1582 }, { "ce_loss": 4.779655864695087e-05, "cls_loss": 0.048828125, "epoch": 0.9501501501501501, "mask_bce_loss": 0.8800055384635925, "mask_dice_loss": 0.06314685195684433, "mask_loss": 0.9431523680686951, "step": 1582 }, { "epoch": 0.9507507507507508, "grad_norm": 37.25521469116211, "learning_rate": 1.3005804959405064e-08, "loss": 0.8645075559616089, "step": 1583 }, { "ce_loss": 9.032934030983597e-05, "cls_loss": 0.12255859375, "epoch": 0.9507507507507508, "mask_bce_loss": 0.5410715937614441, "mask_dice_loss": 0.06032479926943779, "mask_loss": 0.601396381855011, "step": 1583 }, { "epoch": 0.9513513513513514, "grad_norm": 45.495521545410156, "learning_rate": 1.269496076854315e-08, "loss": 0.8768759369850159, "step": 1584 }, { "ce_loss": 3.659651702037081e-05, "cls_loss": 0.02734375, "epoch": 0.9513513513513514, "mask_bce_loss": 0.30474141240119934, "mask_dice_loss": 0.041806962341070175, "mask_loss": 0.3465483784675598, "step": 1584 }, { "epoch": 0.9519519519519519, "grad_norm": 26.616363525390625, "learning_rate": 1.238785256943342e-08, "loss": 0.7764886617660522, "step": 1585 }, { "ce_loss": 4.8042293201433495e-05, "cls_loss": 0.04833984375, "epoch": 0.9519519519519519, "mask_bce_loss": 0.33670496940612793, "mask_dice_loss": 0.06455674767494202, "mask_loss": 0.40126171708106995, "step": 1585 }, { "epoch": 0.9525525525525526, "grad_norm": 29.970678329467773, "learning_rate": 1.2084481524182399e-08, "loss": 0.8583704233169556, "step": 1586 }, { "ce_loss": 5.91832103964407e-05, "cls_loss": 0.045166015625, "epoch": 0.9525525525525526, "mask_bce_loss": 0.6930677890777588, "mask_dice_loss": 0.06645725667476654, "mask_loss": 0.7595250606536865, "step": 1586 }, { "epoch": 0.9531531531531532, "grad_norm": 32.537506103515625, "learning_rate": 1.1784848780755252e-08, "loss": 0.7926861047744751, "step": 1587 }, { "ce_loss": 8.01755886641331e-05, "cls_loss": 0.0546875, "epoch": 0.9531531531531532, "mask_bce_loss": 0.911344051361084, "mask_dice_loss": 0.06616406887769699, "mask_loss": 0.9775081276893616, "step": 1587 }, { "epoch": 0.9537537537537537, "grad_norm": 22.14620590209961, "learning_rate": 1.1488955472971018e-08, "loss": 0.7843003273010254, "step": 1588 }, { "ce_loss": 0.004111407790333033, "cls_loss": 0.060546875, "epoch": 0.9537537537537537, "mask_bce_loss": 0.6273741126060486, "mask_dice_loss": 0.10214600712060928, "mask_loss": 0.7295201420783997, "step": 1588 }, { "epoch": 0.9543543543543543, "grad_norm": 47.173423767089844, "learning_rate": 1.1196802720499055e-08, "loss": 0.8340620398521423, "step": 1589 }, { "ce_loss": 0.004844349343329668, "cls_loss": 0.04296875, "epoch": 0.9543543543543543, "mask_bce_loss": 0.7437102198600769, "mask_dice_loss": 0.08455999940633774, "mask_loss": 0.8282701969146729, "step": 1589 }, { "epoch": 0.954954954954955, "grad_norm": 13.489806175231934, "learning_rate": 1.0908391628854042e-08, "loss": 0.6685543060302734, "step": 1590 }, { "ce_loss": 5.341435826267116e-05, "cls_loss": 0.06298828125, "epoch": 0.954954954954955, "mask_bce_loss": 0.6763090491294861, "mask_dice_loss": 0.13278444111347198, "mask_loss": 0.8090934753417969, "step": 1590 }, { "epoch": 0.9555555555555556, "grad_norm": 29.576879501342773, "learning_rate": 1.0623723289392094e-08, "loss": 0.8436827659606934, "step": 1591 }, { "ce_loss": 0.0008982768631540239, "cls_loss": 0.0625, "epoch": 0.9555555555555556, "mask_bce_loss": 0.7548280358314514, "mask_dice_loss": 0.11226110905408859, "mask_loss": 0.8670891523361206, "step": 1591 }, { "epoch": 0.9561561561561561, "grad_norm": 13.098740577697754, "learning_rate": 1.0342798779306772e-08, "loss": 0.715121865272522, "step": 1592 }, { "ce_loss": 4.6583678340539336e-05, "cls_loss": 0.05810546875, "epoch": 0.9561561561561561, "mask_bce_loss": 0.5628347992897034, "mask_dice_loss": 0.15837503969669342, "mask_loss": 0.7212098240852356, "step": 1592 }, { "epoch": 0.9567567567567568, "grad_norm": 21.82835578918457, "learning_rate": 1.0065619161624739e-08, "loss": 0.7150670289993286, "step": 1593 }, { "ce_loss": 5.713346763513982e-05, "cls_loss": 0.056640625, "epoch": 0.9567567567567568, "mask_bce_loss": 1.1639090776443481, "mask_dice_loss": 0.10707942396402359, "mask_loss": 1.2709884643554688, "step": 1593 }, { "epoch": 0.9573573573573574, "grad_norm": 28.96108627319336, "learning_rate": 9.792185485201997e-09, "loss": 0.8925516605377197, "step": 1594 }, { "ce_loss": 6.331736949505284e-05, "cls_loss": 0.0654296875, "epoch": 0.9573573573573574, "mask_bce_loss": 1.7858842611312866, "mask_dice_loss": 0.1322135180234909, "mask_loss": 1.918097734451294, "step": 1594 }, { "epoch": 0.9579579579579579, "grad_norm": 125.34935760498047, "learning_rate": 9.52249878471978e-09, "loss": 1.1060335636138916, "step": 1595 }, { "ce_loss": 2.5347260816488415e-05, "cls_loss": 0.10791015625, "epoch": 0.9579579579579579, "mask_bce_loss": 0.8641033172607422, "mask_dice_loss": 0.09998566657304764, "mask_loss": 0.9640889763832092, "step": 1595 }, { "epoch": 0.9585585585585585, "grad_norm": 32.3657341003418, "learning_rate": 9.256560080680542e-09, "loss": 0.9141820669174194, "step": 1596 }, { "ce_loss": 8.618370338808745e-05, "cls_loss": 0.0498046875, "epoch": 0.9585585585585585, "mask_bce_loss": 0.5164405703544617, "mask_dice_loss": 0.07415208965539932, "mask_loss": 0.5905926823616028, "step": 1596 }, { "epoch": 0.9591591591591592, "grad_norm": 41.87116241455078, "learning_rate": 8.994370379404314e-09, "loss": 0.764319658279419, "step": 1597 }, { "ce_loss": 0.00030053441878408194, "cls_loss": 0.06591796875, "epoch": 0.9591591591591592, "mask_bce_loss": 2.165649175643921, "mask_dice_loss": 0.09814687818288803, "mask_loss": 2.263796091079712, "step": 1597 }, { "epoch": 0.9597597597597598, "grad_norm": 27.224864959716797, "learning_rate": 8.735930673024805e-09, "loss": 0.7928612232208252, "step": 1598 }, { "ce_loss": 0.0008777435286901891, "cls_loss": 0.047119140625, "epoch": 0.9597597597597598, "mask_bce_loss": 0.6977686882019043, "mask_dice_loss": 0.10021164268255234, "mask_loss": 0.7979803085327148, "step": 1598 }, { "epoch": 0.9603603603603603, "grad_norm": 24.372745513916016, "learning_rate": 8.481241939485517e-09, "loss": 0.7345318794250488, "step": 1599 }, { "ce_loss": 0.0001279336865991354, "cls_loss": 0.06103515625, "epoch": 0.9603603603603603, "mask_bce_loss": 0.8343752026557922, "mask_dice_loss": 0.07262479513883591, "mask_loss": 0.9070000052452087, "step": 1599 }, { "epoch": 0.960960960960961, "grad_norm": 34.42945861816406, "learning_rate": 8.230305142536198e-09, "loss": 0.8956494331359863, "step": 1600 }, { "ce_loss": 0.00020141989807598293, "cls_loss": 0.0267333984375, "epoch": 0.960960960960961, "mask_bce_loss": 0.13342133164405823, "mask_dice_loss": 0.018031572923064232, "mask_loss": 0.151452898979187, "step": 1600 }, { "epoch": 0.9615615615615616, "grad_norm": 27.496246337890625, "learning_rate": 7.983121231729173e-09, "loss": 0.8195914030075073, "step": 1601 }, { "ce_loss": 0.0012169991387054324, "cls_loss": 0.05126953125, "epoch": 0.9615615615615616, "mask_bce_loss": 0.7585832476615906, "mask_dice_loss": 0.06481310725212097, "mask_loss": 0.8233963251113892, "step": 1601 }, { "epoch": 0.9621621621621622, "grad_norm": 43.350406646728516, "learning_rate": 7.739691142415795e-09, "loss": 0.8211070895195007, "step": 1602 }, { "ce_loss": 0.00026354100555181503, "cls_loss": 0.0693359375, "epoch": 0.9621621621621622, "mask_bce_loss": 0.5636121034622192, "mask_dice_loss": 0.1020369678735733, "mask_loss": 0.6656490564346313, "step": 1602 }, { "epoch": 0.9627627627627627, "grad_norm": 24.76934814453125, "learning_rate": 7.500015795742886e-09, "loss": 0.7266354560852051, "step": 1603 }, { "ce_loss": 4.841851114179008e-05, "cls_loss": 0.031005859375, "epoch": 0.9627627627627627, "mask_bce_loss": 0.3154594898223877, "mask_dice_loss": 0.055403243750333786, "mask_loss": 0.3708627223968506, "step": 1603 }, { "epoch": 0.9633633633633634, "grad_norm": 37.72730255126953, "learning_rate": 7.264096098648976e-09, "loss": 0.8977245092391968, "step": 1604 }, { "ce_loss": 6.951730756554753e-05, "cls_loss": 0.0634765625, "epoch": 0.9633633633633634, "mask_bce_loss": 1.0039279460906982, "mask_dice_loss": 0.09648922830820084, "mask_loss": 1.100417137145996, "step": 1604 }, { "epoch": 0.963963963963964, "grad_norm": 26.30194091796875, "learning_rate": 7.031932943861285e-09, "loss": 0.8993555903434753, "step": 1605 }, { "ce_loss": 3.9297196053666994e-05, "cls_loss": 0.044189453125, "epoch": 0.963963963963964, "mask_bce_loss": 0.29496827721595764, "mask_dice_loss": 0.11262967437505722, "mask_loss": 0.40759795904159546, "step": 1605 }, { "epoch": 0.9645645645645645, "grad_norm": 25.778676986694336, "learning_rate": 6.803527209892079e-09, "loss": 0.8774750232696533, "step": 1606 }, { "ce_loss": 7.751179509796202e-05, "cls_loss": 0.060546875, "epoch": 0.9645645645645645, "mask_bce_loss": 1.1207903623580933, "mask_dice_loss": 0.1326717734336853, "mask_loss": 1.2534620761871338, "step": 1606 }, { "epoch": 0.9651651651651652, "grad_norm": 36.925106048583984, "learning_rate": 6.578879761035549e-09, "loss": 0.9138369560241699, "step": 1607 }, { "ce_loss": 4.153602276346646e-05, "cls_loss": 0.057861328125, "epoch": 0.9651651651651652, "mask_bce_loss": 0.6795647740364075, "mask_dice_loss": 0.11353176087141037, "mask_loss": 0.7930965423583984, "step": 1607 }, { "epoch": 0.9657657657657658, "grad_norm": 35.84165573120117, "learning_rate": 6.357991447364264e-09, "loss": 0.7801684737205505, "step": 1608 }, { "ce_loss": 0.0003484618791844696, "cls_loss": 0.031982421875, "epoch": 0.9657657657657658, "mask_bce_loss": 0.5159086585044861, "mask_dice_loss": 0.04877268150448799, "mask_loss": 0.564681351184845, "step": 1608 }, { "epoch": 0.9663663663663664, "grad_norm": 65.13900756835938, "learning_rate": 6.1408631047263905e-09, "loss": 0.757876455783844, "step": 1609 }, { "ce_loss": 0.00013490853598341346, "cls_loss": 0.061767578125, "epoch": 0.9663663663663664, "mask_bce_loss": 0.5346238613128662, "mask_dice_loss": 0.13176636397838593, "mask_loss": 0.6663902401924133, "step": 1609 }, { "epoch": 0.9669669669669669, "grad_norm": 37.01641082763672, "learning_rate": 5.927495554741813e-09, "loss": 0.7977418899536133, "step": 1610 }, { "ce_loss": 0.00011762756912503392, "cls_loss": 0.04150390625, "epoch": 0.9669669669669669, "mask_bce_loss": 0.6725409030914307, "mask_dice_loss": 0.05244023725390434, "mask_loss": 0.7249811291694641, "step": 1610 }, { "epoch": 0.9675675675675676, "grad_norm": 24.31170654296875, "learning_rate": 5.717889604799797e-09, "loss": 0.7867835760116577, "step": 1611 }, { "ce_loss": 3.78478471247945e-05, "cls_loss": 0.052734375, "epoch": 0.9675675675675676, "mask_bce_loss": 0.35884809494018555, "mask_dice_loss": 0.1068471223115921, "mask_loss": 0.46569520235061646, "step": 1611 }, { "epoch": 0.9681681681681682, "grad_norm": 39.13132095336914, "learning_rate": 5.512046048055663e-09, "loss": 0.7650428414344788, "step": 1612 }, { "ce_loss": 3.5599958209786564e-05, "cls_loss": 0.0458984375, "epoch": 0.9681681681681682, "mask_bce_loss": 0.6274210214614868, "mask_dice_loss": 0.05716317892074585, "mask_loss": 0.6845842003822327, "step": 1612 }, { "epoch": 0.9687687687687687, "grad_norm": 31.03400421142578, "learning_rate": 5.309965663427341e-09, "loss": 0.9781216382980347, "step": 1613 }, { "ce_loss": 8.501196862198412e-05, "cls_loss": 0.045654296875, "epoch": 0.9687687687687687, "mask_bce_loss": 0.44783076643943787, "mask_dice_loss": 0.0613582618534565, "mask_loss": 0.5091890096664429, "step": 1613 }, { "epoch": 0.9693693693693693, "grad_norm": 20.579124450683594, "learning_rate": 5.11164921559315e-09, "loss": 0.8029670715332031, "step": 1614 }, { "ce_loss": 7.200403342721984e-05, "cls_loss": 0.026123046875, "epoch": 0.9693693693693693, "mask_bce_loss": 0.1996050328016281, "mask_dice_loss": 0.019933681935071945, "mask_loss": 0.21953871846199036, "step": 1614 }, { "epoch": 0.96996996996997, "grad_norm": 42.38483810424805, "learning_rate": 4.917097454988584e-09, "loss": 0.8167329430580139, "step": 1615 }, { "ce_loss": 0.005575280636548996, "cls_loss": 0.05517578125, "epoch": 0.96996996996997, "mask_bce_loss": 1.0749714374542236, "mask_dice_loss": 0.08025097846984863, "mask_loss": 1.1552224159240723, "step": 1615 }, { "epoch": 0.9705705705705706, "grad_norm": 18.726369857788086, "learning_rate": 4.7263111178030836e-09, "loss": 0.7072327136993408, "step": 1616 }, { "ce_loss": 4.2829175072256476e-05, "cls_loss": 0.03662109375, "epoch": 0.9705705705705706, "mask_bce_loss": 0.5731096267700195, "mask_dice_loss": 0.10514700412750244, "mask_loss": 0.678256630897522, "step": 1616 }, { "epoch": 0.9711711711711711, "grad_norm": 33.942222595214844, "learning_rate": 4.539290925978046e-09, "loss": 0.7844276428222656, "step": 1617 }, { "ce_loss": 0.0003849603235721588, "cls_loss": 0.057861328125, "epoch": 0.9711711711711711, "mask_bce_loss": 0.6437317132949829, "mask_dice_loss": 0.07128915935754776, "mask_loss": 0.7150208950042725, "step": 1617 }, { "epoch": 0.9717717717717718, "grad_norm": 17.790407180786133, "learning_rate": 4.356037587203376e-09, "loss": 0.7351356744766235, "step": 1618 }, { "ce_loss": 7.444948278134689e-05, "cls_loss": 0.03955078125, "epoch": 0.9717717717717718, "mask_bce_loss": 0.561252236366272, "mask_dice_loss": 0.05925344303250313, "mask_loss": 0.620505690574646, "step": 1618 }, { "epoch": 0.9723723723723724, "grad_norm": 20.323299407958984, "learning_rate": 4.1765517949150505e-09, "loss": 0.7449316382408142, "step": 1619 }, { "ce_loss": 5.398847861215472e-05, "cls_loss": 0.053955078125, "epoch": 0.9723723723723724, "mask_bce_loss": 0.5807098150253296, "mask_dice_loss": 0.09995049983263016, "mask_loss": 0.6806603074073792, "step": 1619 }, { "epoch": 0.972972972972973, "grad_norm": 21.898880004882812, "learning_rate": 4.000834228292893e-09, "loss": 0.7281391620635986, "step": 1620 }, { "ce_loss": 0.00034587542177177966, "cls_loss": 0.057373046875, "epoch": 0.972972972972973, "mask_bce_loss": 1.1458083391189575, "mask_dice_loss": 0.1058284193277359, "mask_loss": 1.2516367435455322, "step": 1620 }, { "epoch": 0.9735735735735735, "grad_norm": 33.455955505371094, "learning_rate": 3.828885552257244e-09, "loss": 0.8755297660827637, "step": 1621 }, { "ce_loss": 0.000255509716225788, "cls_loss": 0.05615234375, "epoch": 0.9735735735735735, "mask_bce_loss": 1.2683019638061523, "mask_dice_loss": 0.08599896728992462, "mask_loss": 1.3543009757995605, "step": 1621 }, { "epoch": 0.9741741741741742, "grad_norm": 26.104389190673828, "learning_rate": 3.6607064174671854e-09, "loss": 0.6933647394180298, "step": 1622 }, { "ce_loss": 0.00011425679986132309, "cls_loss": 0.0380859375, "epoch": 0.9741741741741742, "mask_bce_loss": 0.6696191430091858, "mask_dice_loss": 0.06781736761331558, "mask_loss": 0.7374365329742432, "step": 1622 }, { "epoch": 0.9747747747747748, "grad_norm": 34.484413146972656, "learning_rate": 3.496297460317432e-09, "loss": 0.8652414679527283, "step": 1623 }, { "ce_loss": 8.269976387964562e-05, "cls_loss": 0.033203125, "epoch": 0.9747747747747748, "mask_bce_loss": 0.4814628064632416, "mask_dice_loss": 0.0626474991440773, "mask_loss": 0.5441102981567383, "step": 1623 }, { "epoch": 0.9753753753753753, "grad_norm": 28.991239547729492, "learning_rate": 3.3356593029364443e-09, "loss": 0.9882416725158691, "step": 1624 }, { "ce_loss": 6.868410127935931e-05, "cls_loss": 0.050048828125, "epoch": 0.9753753753753753, "mask_bce_loss": 0.1548873335123062, "mask_dice_loss": 0.06276144832372665, "mask_loss": 0.21764877438545227, "step": 1624 }, { "epoch": 0.975975975975976, "grad_norm": 33.077754974365234, "learning_rate": 3.178792553183873e-09, "loss": 0.9707802534103394, "step": 1625 }, { "ce_loss": 2.28148255700944e-05, "cls_loss": 0.0634765625, "epoch": 0.975975975975976, "mask_bce_loss": 0.4692535400390625, "mask_dice_loss": 0.11273135244846344, "mask_loss": 0.5819848775863647, "step": 1625 }, { "epoch": 0.9765765765765766, "grad_norm": 22.15627098083496, "learning_rate": 3.0256978046481195e-09, "loss": 0.7357218861579895, "step": 1626 }, { "ce_loss": 0.0022498618345707655, "cls_loss": 0.0654296875, "epoch": 0.9765765765765766, "mask_bce_loss": 1.4524056911468506, "mask_dice_loss": 0.10882651805877686, "mask_loss": 1.5612322092056274, "step": 1626 }, { "epoch": 0.9771771771771772, "grad_norm": 26.219682693481445, "learning_rate": 2.8763756366443346e-09, "loss": 0.7299023866653442, "step": 1627 }, { "ce_loss": 5.911682455916889e-05, "cls_loss": 0.03271484375, "epoch": 0.9771771771771772, "mask_bce_loss": 0.291803240776062, "mask_dice_loss": 0.028284410014748573, "mask_loss": 0.32008764147758484, "step": 1627 }, { "epoch": 0.9777777777777777, "grad_norm": 18.2640380859375, "learning_rate": 2.7308266142119786e-09, "loss": 0.7569894790649414, "step": 1628 }, { "ce_loss": 5.197579594096169e-05, "cls_loss": 0.05078125, "epoch": 0.9777777777777777, "mask_bce_loss": 0.7570606470108032, "mask_dice_loss": 0.08266349136829376, "mask_loss": 0.8397241234779358, "step": 1628 }, { "epoch": 0.9783783783783784, "grad_norm": 34.01309585571289, "learning_rate": 2.589051288112931e-09, "loss": 0.7909774780273438, "step": 1629 }, { "ce_loss": 5.177401544642635e-05, "cls_loss": 0.05224609375, "epoch": 0.9783783783783784, "mask_bce_loss": 1.7754029035568237, "mask_dice_loss": 0.09698721021413803, "mask_loss": 1.8723901510238647, "step": 1629 }, { "epoch": 0.978978978978979, "grad_norm": 25.046932220458984, "learning_rate": 2.451050194829274e-09, "loss": 0.9948714971542358, "step": 1630 }, { "ce_loss": 0.12167977541685104, "cls_loss": 0.06640625, "epoch": 0.978978978978979, "mask_bce_loss": 1.5163732767105103, "mask_dice_loss": 0.13454066216945648, "mask_loss": 1.650913953781128, "step": 1630 }, { "epoch": 0.9795795795795795, "grad_norm": 24.607177734375, "learning_rate": 2.3168238565612895e-09, "loss": 0.706508994102478, "step": 1631 }, { "ce_loss": 5.0330727390246466e-05, "cls_loss": 0.0257568359375, "epoch": 0.9795795795795795, "mask_bce_loss": 0.22494564950466156, "mask_dice_loss": 0.04277375340461731, "mask_loss": 0.2677193880081177, "step": 1631 }, { "epoch": 0.9801801801801802, "grad_norm": 37.1655387878418, "learning_rate": 2.186372781225465e-09, "loss": 1.0101134777069092, "step": 1632 }, { "ce_loss": 4.298337080399506e-05, "cls_loss": 0.05859375, "epoch": 0.9801801801801802, "mask_bce_loss": 1.0827252864837646, "mask_dice_loss": 0.07664652168750763, "mask_loss": 1.1593718528747559, "step": 1632 }, { "epoch": 0.9807807807807808, "grad_norm": 31.394142150878906, "learning_rate": 2.059697462452492e-09, "loss": 0.7137796878814697, "step": 1633 }, { "ce_loss": 0.0001304203033214435, "cls_loss": 0.05615234375, "epoch": 0.9807807807807808, "mask_bce_loss": 0.6871821880340576, "mask_dice_loss": 0.1718904823064804, "mask_loss": 0.8590726852416992, "step": 1633 }, { "epoch": 0.9813813813813814, "grad_norm": 21.73752784729004, "learning_rate": 1.9367983795856024e-09, "loss": 0.7467273473739624, "step": 1634 }, { "ce_loss": 0.00012721371604129672, "cls_loss": 0.037109375, "epoch": 0.9813813813813814, "mask_bce_loss": 0.28815945982933044, "mask_dice_loss": 0.031028253957629204, "mask_loss": 0.3191877007484436, "step": 1634 }, { "epoch": 0.9819819819819819, "grad_norm": 15.822310447692871, "learning_rate": 1.8176759976785694e-09, "loss": 0.7179657220840454, "step": 1635 }, { "ce_loss": 0.00048197293654084206, "cls_loss": 0.045166015625, "epoch": 0.9819819819819819, "mask_bce_loss": 0.6053842902183533, "mask_dice_loss": 0.046318281441926956, "mask_loss": 0.6517025828361511, "step": 1635 }, { "epoch": 0.9825825825825826, "grad_norm": 21.132211685180664, "learning_rate": 1.7023307674940424e-09, "loss": 0.7637456059455872, "step": 1636 }, { "ce_loss": 0.0007969992584548891, "cls_loss": 0.04736328125, "epoch": 0.9825825825825826, "mask_bce_loss": 0.9846293330192566, "mask_dice_loss": 0.08963202685117722, "mask_loss": 1.0742613077163696, "step": 1636 }, { "epoch": 0.9831831831831832, "grad_norm": 35.797855377197266, "learning_rate": 1.5907631255017706e-09, "loss": 0.7219981551170349, "step": 1637 }, { "ce_loss": 9.963300544768572e-05, "cls_loss": 0.036865234375, "epoch": 0.9831831831831832, "mask_bce_loss": 0.4907074570655823, "mask_dice_loss": 0.04391677305102348, "mask_loss": 0.5346242189407349, "step": 1637 }, { "epoch": 0.9837837837837838, "grad_norm": 20.950166702270508, "learning_rate": 1.482973493877271e-09, "loss": 0.6707118153572083, "step": 1638 }, { "ce_loss": 0.0005717252497561276, "cls_loss": 0.05517578125, "epoch": 0.9837837837837838, "mask_bce_loss": 0.8602733612060547, "mask_dice_loss": 0.05874377116560936, "mask_loss": 0.9190171360969543, "step": 1638 }, { "epoch": 0.9843843843843844, "grad_norm": 23.591798782348633, "learning_rate": 1.378962280499385e-09, "loss": 0.7744800448417664, "step": 1639 }, { "ce_loss": 0.00015401719429064542, "cls_loss": 0.052734375, "epoch": 0.9843843843843844, "mask_bce_loss": 0.7097653746604919, "mask_dice_loss": 0.08986286073923111, "mask_loss": 0.7996282577514648, "step": 1639 }, { "epoch": 0.984984984984985, "grad_norm": 25.181955337524414, "learning_rate": 1.2787298789499468e-09, "loss": 0.8926490545272827, "step": 1640 }, { "ce_loss": 9.285274427384138e-05, "cls_loss": 0.0478515625, "epoch": 0.984984984984985, "mask_bce_loss": 0.6949083805084229, "mask_dice_loss": 0.12337058037519455, "mask_loss": 0.818278968334198, "step": 1640 }, { "epoch": 0.9855855855855856, "grad_norm": 25.740598678588867, "learning_rate": 1.1822766685112284e-09, "loss": 0.8583257794380188, "step": 1641 }, { "ce_loss": 7.238476973725483e-05, "cls_loss": 0.04931640625, "epoch": 0.9855855855855856, "mask_bce_loss": 0.9311676025390625, "mask_dice_loss": 0.08342214673757553, "mask_loss": 1.014589786529541, "step": 1641 }, { "epoch": 0.9861861861861861, "grad_norm": 22.637046813964844, "learning_rate": 1.0896030141650525e-09, "loss": 0.8707566261291504, "step": 1642 }, { "ce_loss": 3.8190821214811876e-05, "cls_loss": 0.039794921875, "epoch": 0.9861861861861861, "mask_bce_loss": 0.2733311057090759, "mask_dice_loss": 0.055062007158994675, "mask_loss": 0.3283931016921997, "step": 1642 }, { "epoch": 0.9867867867867868, "grad_norm": 49.5020866394043, "learning_rate": 1.0007092665913486e-09, "loss": 0.856677770614624, "step": 1643 }, { "ce_loss": 0.0002485712175257504, "cls_loss": 0.03369140625, "epoch": 0.9867867867867868, "mask_bce_loss": 0.2426331341266632, "mask_dice_loss": 0.025604525581002235, "mask_loss": 0.2682376503944397, "step": 1643 }, { "epoch": 0.9873873873873874, "grad_norm": 30.986974716186523, "learning_rate": 9.155957621665989e-10, "loss": 0.8673272728919983, "step": 1644 }, { "ce_loss": 4.199123941361904e-05, "cls_loss": 0.04052734375, "epoch": 0.9873873873873874, "mask_bce_loss": 0.4738418757915497, "mask_dice_loss": 0.03937236964702606, "mask_loss": 0.5132142305374146, "step": 1644 }, { "epoch": 0.987987987987988, "grad_norm": 20.474843978881836, "learning_rate": 8.342628229630611e-10, "loss": 0.7110438346862793, "step": 1645 }, { "ce_loss": 3.640655995695852e-05, "cls_loss": 0.04150390625, "epoch": 0.987987987987988, "mask_bce_loss": 0.3536900281906128, "mask_dice_loss": 0.04434345290064812, "mask_loss": 0.39803346991539, "step": 1645 }, { "epoch": 0.9885885885885886, "grad_norm": 32.842735290527344, "learning_rate": 7.567107567467701e-10, "loss": 0.7591404914855957, "step": 1646 }, { "ce_loss": 0.00011371563596185297, "cls_loss": 0.05908203125, "epoch": 0.9885885885885886, "mask_bce_loss": 0.9053813815116882, "mask_dice_loss": 0.09058945626020432, "mask_loss": 0.9959708452224731, "step": 1646 }, { "epoch": 0.9891891891891892, "grad_norm": 18.907939910888672, "learning_rate": 6.829398569770939e-10, "loss": 0.8598214387893677, "step": 1647 }, { "ce_loss": 4.327698115957901e-05, "cls_loss": 0.051513671875, "epoch": 0.9891891891891892, "mask_bce_loss": 0.7864032983779907, "mask_dice_loss": 0.11548838764429092, "mask_loss": 0.9018917083740234, "step": 1647 }, { "epoch": 0.9897897897897898, "grad_norm": 22.802234649658203, "learning_rate": 6.129504028055122e-10, "loss": 0.7844939231872559, "step": 1648 }, { "ce_loss": 2.709869295358658e-05, "cls_loss": 0.048095703125, "epoch": 0.9897897897897898, "mask_bce_loss": 0.40327683091163635, "mask_dice_loss": 0.04995683580636978, "mask_loss": 0.45323365926742554, "step": 1648 }, { "epoch": 0.9903903903903903, "grad_norm": 38.573490142822266, "learning_rate": 5.46742659073951e-10, "loss": 0.9672501683235168, "step": 1649 }, { "ce_loss": 8.156995318131521e-05, "cls_loss": 0.029052734375, "epoch": 0.9903903903903903, "mask_bce_loss": 0.20236502587795258, "mask_dice_loss": 0.021960845217108727, "mask_loss": 0.22432586550712585, "step": 1649 }, { "epoch": 0.990990990990991, "grad_norm": 43.77049255371094, "learning_rate": 4.843168763146721e-10, "loss": 0.9081733822822571, "step": 1650 }, { "ce_loss": 7.357908179983497e-05, "cls_loss": 0.045654296875, "epoch": 0.990990990990991, "mask_bce_loss": 0.9161426424980164, "mask_dice_loss": 0.10489790886640549, "mask_loss": 1.0210405588150024, "step": 1650 }, { "epoch": 0.9915915915915916, "grad_norm": 29.044273376464844, "learning_rate": 4.2567329074871815e-10, "loss": 0.7002263069152832, "step": 1651 }, { "ce_loss": 2.236923683085479e-05, "cls_loss": 0.031005859375, "epoch": 0.9915915915915916, "mask_bce_loss": 0.5421926379203796, "mask_dice_loss": 0.027758324518799782, "mask_loss": 0.5699509382247925, "step": 1651 }, { "epoch": 0.9921921921921922, "grad_norm": 20.167566299438477, "learning_rate": 3.708121242850248e-10, "loss": 0.6870443820953369, "step": 1652 }, { "ce_loss": 9.874461102299392e-05, "cls_loss": 0.023681640625, "epoch": 0.9921921921921922, "mask_bce_loss": 0.22335851192474365, "mask_dice_loss": 0.015578952617943287, "mask_loss": 0.23893746733665466, "step": 1652 }, { "epoch": 0.9927927927927928, "grad_norm": 21.79425048828125, "learning_rate": 3.197335845199767e-10, "loss": 0.8779474496841431, "step": 1653 }, { "ce_loss": 9.888368367683142e-05, "cls_loss": 0.03955078125, "epoch": 0.9927927927927928, "mask_bce_loss": 0.6702622771263123, "mask_dice_loss": 0.0789983943104744, "mask_loss": 0.749260663986206, "step": 1653 }, { "epoch": 0.9933933933933934, "grad_norm": 38.699649810791016, "learning_rate": 2.7243786473629683e-10, "loss": 0.8875558376312256, "step": 1654 }, { "ce_loss": 0.0004427247622516006, "cls_loss": 0.040283203125, "epoch": 0.9933933933933934, "mask_bce_loss": 0.6869449019432068, "mask_dice_loss": 0.033831637352705, "mask_loss": 0.7207765579223633, "step": 1654 }, { "epoch": 0.993993993993994, "grad_norm": 18.30940055847168, "learning_rate": 2.2892514390238095e-10, "loss": 0.7713621854782104, "step": 1655 }, { "ce_loss": 0.00020927174773532897, "cls_loss": 0.06787109375, "epoch": 0.993993993993994, "mask_bce_loss": 0.9336949586868286, "mask_dice_loss": 0.1387874037027359, "mask_loss": 1.0724823474884033, "step": 1655 }, { "epoch": 0.9945945945945946, "grad_norm": 16.653846740722656, "learning_rate": 1.8919558667174208e-10, "loss": 0.9222574234008789, "step": 1656 }, { "ce_loss": 6.705724081257358e-05, "cls_loss": 0.0322265625, "epoch": 0.9945945945945946, "mask_bce_loss": 0.3443571627140045, "mask_dice_loss": 0.022797169163823128, "mask_loss": 0.3671543300151825, "step": 1656 }, { "epoch": 0.9951951951951952, "grad_norm": 35.063270568847656, "learning_rate": 1.5324934338201145e-10, "loss": 0.7920852303504944, "step": 1657 }, { "ce_loss": 5.2357012464199215e-05, "cls_loss": 0.047607421875, "epoch": 0.9951951951951952, "mask_bce_loss": 0.8331848978996277, "mask_dice_loss": 0.110460065305233, "mask_loss": 0.9436449408531189, "step": 1657 }, { "epoch": 0.9957957957957958, "grad_norm": 38.798675537109375, "learning_rate": 1.210865500549385e-10, "loss": 0.8934506177902222, "step": 1658 }, { "ce_loss": 0.00017051378381438553, "cls_loss": 0.06640625, "epoch": 0.9957957957957958, "mask_bce_loss": 1.217143177986145, "mask_dice_loss": 0.11300019919872284, "mask_loss": 1.3301433324813843, "step": 1658 }, { "epoch": 0.9963963963963964, "grad_norm": 37.07097625732422, "learning_rate": 9.270732839561369e-11, "loss": 0.8357969522476196, "step": 1659 }, { "ce_loss": 4.119363075005822e-05, "cls_loss": 0.06298828125, "epoch": 0.9963963963963964, "mask_bce_loss": 0.7603009343147278, "mask_dice_loss": 0.17426824569702148, "mask_loss": 0.9345691800117493, "step": 1659 }, { "epoch": 0.996996996996997, "grad_norm": 74.0802001953125, "learning_rate": 6.811178579180232e-11, "loss": 0.793846845626831, "step": 1660 }, { "ce_loss": 3.5445620596874505e-05, "cls_loss": 0.06494140625, "epoch": 0.996996996996997, "mask_bce_loss": 1.2373102903366089, "mask_dice_loss": 0.1250511109828949, "mask_loss": 1.3623614311218262, "step": 1660 }, { "epoch": 0.9975975975975976, "grad_norm": 22.89891815185547, "learning_rate": 4.7300015313611606e-11, "loss": 1.020329236984253, "step": 1661 }, { "ce_loss": 0.00010735722025856376, "cls_loss": 0.02734375, "epoch": 0.9975975975975976, "mask_bce_loss": 0.5593875646591187, "mask_dice_loss": 0.045536261051893234, "mask_loss": 0.6049238443374634, "step": 1661 }, { "epoch": 0.9981981981981982, "grad_norm": 43.89223098754883, "learning_rate": 3.0272095713601564e-11, "loss": 0.8221509456634521, "step": 1662 }, { "ce_loss": 3.9080976421246305e-05, "cls_loss": 0.06005859375, "epoch": 0.9981981981981982, "mask_bce_loss": 1.2856882810592651, "mask_dice_loss": 0.09206981211900711, "mask_loss": 1.3777581453323364, "step": 1662 }, { "epoch": 0.9987987987987988, "grad_norm": 31.941431045532227, "learning_rate": 1.7028091425896894e-11, "loss": 0.9134284257888794, "step": 1663 }, { "ce_loss": 0.03644444793462753, "cls_loss": 0.05908203125, "epoch": 0.9987987987987988, "mask_bce_loss": 2.1531119346618652, "mask_dice_loss": 0.09433292597532272, "mask_loss": 2.2474448680877686, "step": 1663 }, { "epoch": 0.9993993993993994, "grad_norm": 24.129201889038086, "learning_rate": 7.568052566075956e-12, "loss": 0.8448575735092163, "step": 1664 }, { "ce_loss": 0.00010907428804785013, "cls_loss": 0.04931640625, "epoch": 0.9993993993993994, "mask_bce_loss": 0.5701794028282166, "mask_dice_loss": 0.04910232499241829, "mask_loss": 0.6192817091941833, "step": 1664 }, { "epoch": 1.0, "grad_norm": 17.456424713134766, "learning_rate": 1.8920149313927936e-12, "loss": 0.69564288854599, "step": 1665 } ], "logging_steps": 1.0, "max_steps": 1665, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.4797965143887053e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }