{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.8678446210990385, "global_step": 200000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "accuracy": 10.791, "active_queue_size": 16384.0, "cl_loss": 200.9172, "doc_norm": 8.4175, "encoder_q-embeddings": 55437.1836, "encoder_q-layer.0": 66370.9844, "encoder_q-layer.1": 53287.2891, "encoder_q-layer.10": 154976.8438, "encoder_q-layer.11": 101511.0, "encoder_q-layer.2": 62755.918, "encoder_q-layer.3": 66636.25, "encoder_q-layer.4": 77426.4375, "encoder_q-layer.5": 87966.7188, "encoder_q-layer.6": 115355.9531, "encoder_q-layer.7": 137456.0938, "encoder_q-layer.8": 170578.2969, "encoder_q-layer.9": 134351.7344, "epoch": 0.0, "inbatch_neg_score": 39.3983, "inbatch_pos_score": 47.1562, "learning_rate": 5.000000000000001e-07, "loss": 200.9172, "norm_diff": 0.3088, "norm_loss": 0.0, "num_token_doc": 66.7654, "num_token_overlap": 11.6294, "num_token_query": 31.2578, "num_token_union": 65.0871, "num_word_context": 202.3035, "num_word_doc": 49.822, "num_word_query": 23.1951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 144184.8461, "preclip_grad_norm_avg": 0.0013, "q@queue_neg_score": 39.4375, "query_norm": 8.1087, "queue_k_norm": 8.4247, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2578, "sent_len_1": 66.7654, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.4137, "stdk": 0.1805, "stdq": 0.1932, "stdqueue_k": 0.1804, "stdqueue_q": 0.0, "step": 100 }, { "accuracy": 11.1816, "active_queue_size": 16384.0, "cl_loss": 124.8438, "doc_norm": 8.345, "encoder_q-embeddings": 12813.1895, "encoder_q-layer.0": 12976.9443, "encoder_q-layer.1": 17972.5723, "encoder_q-layer.10": 37473.9141, "encoder_q-layer.11": 40268.9141, "encoder_q-layer.2": 17961.0273, "encoder_q-layer.3": 17090.4766, "encoder_q-layer.4": 19060.9609, "encoder_q-layer.5": 21451.4648, "encoder_q-layer.6": 26098.6855, "encoder_q-layer.7": 28392.4531, "encoder_q-layer.8": 35797.7227, "encoder_q-layer.9": 28150.7227, "epoch": 0.0, "inbatch_neg_score": 36.1652, "inbatch_pos_score": 40.6562, "learning_rate": 1.0000000000000002e-06, "loss": 124.8438, "norm_diff": 1.0323, "norm_loss": 0.0, "num_token_doc": 66.7068, "num_token_overlap": 11.6792, "num_token_query": 31.4401, "num_token_union": 65.1096, "num_word_context": 202.2005, "num_word_doc": 49.7441, "num_word_query": 23.347, "postclip_grad_norm": 1.0, "preclip_grad_norm": 36050.5753, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 36.2188, "query_norm": 7.3126, "queue_k_norm": 8.3548, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4401, "sent_len_1": 66.7068, "sent_len_max_0": 128.0, "sent_len_max_1": 208.5875, "stdk": 0.1778, "stdq": 0.1433, "stdqueue_k": 0.1781, "stdqueue_q": 0.0, "step": 200 }, { "accuracy": 10.2051, "active_queue_size": 16384.0, "cl_loss": 72.7976, "doc_norm": 8.1958, "encoder_q-embeddings": 6725.1069, "encoder_q-layer.0": 6506.3452, "encoder_q-layer.1": 6297.2163, "encoder_q-layer.10": 19143.2871, "encoder_q-layer.11": 31101.4023, "encoder_q-layer.2": 7227.1138, "encoder_q-layer.3": 7853.5107, "encoder_q-layer.4": 8821.5254, "encoder_q-layer.5": 8913.7354, "encoder_q-layer.6": 10392.0557, "encoder_q-layer.7": 10815.458, "encoder_q-layer.8": 12893.9424, "encoder_q-layer.9": 11737.0762, "epoch": 0.0, "inbatch_neg_score": 34.4461, "inbatch_pos_score": 37.125, "learning_rate": 1.5e-06, "loss": 72.7976, "norm_diff": 1.1867, "norm_loss": 0.0, "num_token_doc": 66.7181, "num_token_overlap": 11.6515, "num_token_query": 31.3642, "num_token_union": 65.1086, "num_word_context": 202.5296, "num_word_doc": 49.7846, "num_word_query": 23.2914, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20728.9341, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 34.4375, "query_norm": 7.009, "queue_k_norm": 8.2154, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3642, "sent_len_1": 66.7181, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.7637, "stdk": 0.1715, "stdq": 0.115, "stdqueue_k": 0.1727, "stdqueue_q": 0.0, "step": 300 }, { "accuracy": 12.1094, "active_queue_size": 16384.0, "cl_loss": 52.5094, "doc_norm": 8.0452, "encoder_q-embeddings": 5721.5093, "encoder_q-layer.0": 5076.9121, "encoder_q-layer.1": 5483.3975, "encoder_q-layer.10": 14232.9805, "encoder_q-layer.11": 23520.0156, "encoder_q-layer.2": 6100.3325, "encoder_q-layer.3": 6910.5547, "encoder_q-layer.4": 7162.4468, "encoder_q-layer.5": 7358.7861, "encoder_q-layer.6": 7566.4653, "encoder_q-layer.7": 7998.3252, "encoder_q-layer.8": 9886.9971, "encoder_q-layer.9": 8714.792, "epoch": 0.0, "inbatch_neg_score": 32.2121, "inbatch_pos_score": 34.2812, "learning_rate": 2.0000000000000003e-06, "loss": 52.5094, "norm_diff": 1.0948, "norm_loss": 0.0, "num_token_doc": 66.8055, "num_token_overlap": 11.6772, "num_token_query": 31.4119, "num_token_union": 65.1658, "num_word_context": 202.3638, "num_word_doc": 49.8353, "num_word_query": 23.3323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 15810.1168, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 32.1875, "query_norm": 6.9504, "queue_k_norm": 8.049, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4119, "sent_len_1": 66.8055, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.1675, "stdk": 0.1662, "stdq": 0.1034, "stdqueue_k": 0.1667, "stdqueue_q": 0.0, "step": 400 }, { "accuracy": 12.5, "active_queue_size": 16384.0, "cl_loss": 39.8046, "doc_norm": 7.8774, "encoder_q-embeddings": 8220.8779, "encoder_q-layer.0": 8347.9072, "encoder_q-layer.1": 10069.5674, "encoder_q-layer.10": 10900.5215, "encoder_q-layer.11": 18055.1816, "encoder_q-layer.2": 11652.0195, "encoder_q-layer.3": 12056.2021, "encoder_q-layer.4": 13327.7012, "encoder_q-layer.5": 13382.4307, "encoder_q-layer.6": 11243.7441, "encoder_q-layer.7": 9869.6602, "encoder_q-layer.8": 8554.4775, "encoder_q-layer.9": 7143.8477, "epoch": 0.0, "inbatch_neg_score": 29.9416, "inbatch_pos_score": 31.5625, "learning_rate": 2.5e-06, "loss": 39.8046, "norm_diff": 0.8268, "norm_loss": 0.0, "num_token_doc": 66.6789, "num_token_overlap": 11.6597, "num_token_query": 31.3736, "num_token_union": 65.0968, "num_word_context": 202.3139, "num_word_doc": 49.7718, "num_word_query": 23.3034, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16921.1265, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 29.8906, "query_norm": 7.0506, "queue_k_norm": 7.8819, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3736, "sent_len_1": 66.6789, "sent_len_max_0": 127.9975, "sent_len_max_1": 206.65, "stdk": 0.1606, "stdq": 0.0976, "stdqueue_k": 0.1612, "stdqueue_q": 0.0, "step": 500 }, { "accuracy": 13.9648, "active_queue_size": 16384.0, "cl_loss": 32.782, "doc_norm": 7.7061, "encoder_q-embeddings": 2271.4744, "encoder_q-layer.0": 1795.7006, "encoder_q-layer.1": 2264.2131, "encoder_q-layer.10": 4847.8726, "encoder_q-layer.11": 8261.6514, "encoder_q-layer.2": 2631.9209, "encoder_q-layer.3": 2897.4346, "encoder_q-layer.4": 3302.825, "encoder_q-layer.5": 3550.6133, "encoder_q-layer.6": 3448.8936, "encoder_q-layer.7": 3667.8865, "encoder_q-layer.8": 4265.4648, "encoder_q-layer.9": 3029.0752, "epoch": 0.0, "inbatch_neg_score": 27.6951, "inbatch_pos_score": 29.1406, "learning_rate": 3e-06, "loss": 32.782, "norm_diff": 0.9358, "norm_loss": 0.0, "num_token_doc": 66.6412, "num_token_overlap": 11.6639, "num_token_query": 31.3526, "num_token_union": 65.0543, "num_word_context": 202.2271, "num_word_doc": 49.7642, "num_word_query": 23.3089, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5758.4462, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 27.6562, "query_norm": 6.7703, "queue_k_norm": 7.717, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3526, "sent_len_1": 66.6412, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.3375, "stdk": 0.1551, "stdq": 0.092, "stdqueue_k": 0.1558, "stdqueue_q": 0.0, "step": 600 }, { "accuracy": 13.2812, "active_queue_size": 16384.0, "cl_loss": 27.6497, "doc_norm": 7.5395, "encoder_q-embeddings": 2476.6882, "encoder_q-layer.0": 2119.3091, "encoder_q-layer.1": 2673.3184, "encoder_q-layer.10": 7942.2153, "encoder_q-layer.11": 9212.7051, "encoder_q-layer.2": 3024.4937, "encoder_q-layer.3": 3474.4324, "encoder_q-layer.4": 3725.2273, "encoder_q-layer.5": 3782.9102, "encoder_q-layer.6": 3834.6204, "encoder_q-layer.7": 3849.0513, "encoder_q-layer.8": 4408.2192, "encoder_q-layer.9": 3741.1411, "epoch": 0.0, "inbatch_neg_score": 23.5596, "inbatch_pos_score": 24.75, "learning_rate": 3.5000000000000004e-06, "loss": 27.6497, "norm_diff": 1.5231, "norm_loss": 0.0, "num_token_doc": 66.7853, "num_token_overlap": 11.6852, "num_token_query": 31.3972, "num_token_union": 65.1018, "num_word_context": 202.2911, "num_word_doc": 49.8109, "num_word_query": 23.3189, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6483.3165, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 23.5312, "query_norm": 6.0163, "queue_k_norm": 7.5491, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3972, "sent_len_1": 66.7853, "sent_len_max_0": 127.975, "sent_len_max_1": 211.4187, "stdk": 0.1495, "stdq": 0.0851, "stdqueue_k": 0.15, "stdqueue_q": 0.0, "step": 700 }, { "accuracy": 13.5742, "active_queue_size": 16384.0, "cl_loss": 22.7845, "doc_norm": 7.3694, "encoder_q-embeddings": 2614.2778, "encoder_q-layer.0": 2096.7385, "encoder_q-layer.1": 2678.8691, "encoder_q-layer.10": 5936.835, "encoder_q-layer.11": 9043.5078, "encoder_q-layer.2": 3133.1533, "encoder_q-layer.3": 3477.8523, "encoder_q-layer.4": 3764.5903, "encoder_q-layer.5": 3867.9712, "encoder_q-layer.6": 3323.792, "encoder_q-layer.7": 3073.5315, "encoder_q-layer.8": 3499.7227, "encoder_q-layer.9": 2742.0015, "epoch": 0.0, "inbatch_neg_score": 17.5153, "inbatch_pos_score": 18.4688, "learning_rate": 4.000000000000001e-06, "loss": 22.7845, "norm_diff": 2.5057, "norm_loss": 0.0, "num_token_doc": 66.7692, "num_token_overlap": 11.6921, "num_token_query": 31.4429, "num_token_union": 65.1397, "num_word_context": 202.3105, "num_word_doc": 49.804, "num_word_query": 23.362, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6028.9476, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 17.5, "query_norm": 4.8637, "queue_k_norm": 7.3808, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4429, "sent_len_1": 66.7692, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.1687, "stdk": 0.1425, "stdq": 0.0789, "stdqueue_k": 0.1437, "stdqueue_q": 0.0, "step": 800 }, { "accuracy": 13.1348, "active_queue_size": 16384.0, "cl_loss": 18.5794, "doc_norm": 7.2151, "encoder_q-embeddings": 2972.7585, "encoder_q-layer.0": 2597.916, "encoder_q-layer.1": 3182.2065, "encoder_q-layer.10": 5800.5835, "encoder_q-layer.11": 8803.6875, "encoder_q-layer.2": 3573.6064, "encoder_q-layer.3": 3908.7888, "encoder_q-layer.4": 4417.6323, "encoder_q-layer.5": 4974.7563, "encoder_q-layer.6": 4218.5474, "encoder_q-layer.7": 3622.4602, "encoder_q-layer.8": 3933.6577, "encoder_q-layer.9": 2803.8181, "epoch": 0.0, "inbatch_neg_score": 11.7514, "inbatch_pos_score": 12.4922, "learning_rate": 4.5e-06, "loss": 18.5794, "norm_diff": 3.5065, "norm_loss": 0.0, "num_token_doc": 66.8885, "num_token_overlap": 11.7031, "num_token_query": 31.3915, "num_token_union": 65.1839, "num_word_context": 202.6105, "num_word_doc": 49.9004, "num_word_query": 23.3236, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6402.8775, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 11.7266, "query_norm": 3.7087, "queue_k_norm": 7.2292, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3915, "sent_len_1": 66.8885, "sent_len_max_0": 127.985, "sent_len_max_1": 209.2413, "stdk": 0.1364, "stdq": 0.0752, "stdqueue_k": 0.1378, "stdqueue_q": 0.0, "step": 900 }, { "accuracy": 12.6953, "active_queue_size": 16384.0, "cl_loss": 15.7051, "doc_norm": 7.084, "encoder_q-embeddings": 3822.9514, "encoder_q-layer.0": 3648.0667, "encoder_q-layer.1": 4033.2578, "encoder_q-layer.10": 3866.0879, "encoder_q-layer.11": 6754.9741, "encoder_q-layer.2": 4718.3027, "encoder_q-layer.3": 5056.7803, "encoder_q-layer.4": 5791.5591, "encoder_q-layer.5": 5969.8955, "encoder_q-layer.6": 4708.7261, "encoder_q-layer.7": 3678.6348, "encoder_q-layer.8": 3275.8811, "encoder_q-layer.9": 1947.9097, "epoch": 0.0, "inbatch_neg_score": 7.6627, "inbatch_pos_score": 8.2812, "learning_rate": 5e-06, "loss": 15.7051, "norm_diff": 4.2549, "norm_loss": 0.0, "num_token_doc": 66.7378, "num_token_overlap": 11.665, "num_token_query": 31.3144, "num_token_union": 65.0856, "num_word_context": 202.4127, "num_word_doc": 49.801, "num_word_query": 23.2487, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6623.2243, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 7.6641, "query_norm": 2.8292, "queue_k_norm": 7.0932, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3144, "sent_len_1": 66.7378, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.0662, "stdk": 0.1315, "stdq": 0.0691, "stdqueue_k": 0.1319, "stdqueue_q": 0.0, "step": 1000 }, { "accuracy": 13.0371, "active_queue_size": 16384.0, "cl_loss": 13.8059, "doc_norm": 6.9657, "encoder_q-embeddings": 4213.6172, "encoder_q-layer.0": 3769.0312, "encoder_q-layer.1": 4578.5742, "encoder_q-layer.10": 3513.6946, "encoder_q-layer.11": 6707.4658, "encoder_q-layer.2": 5414.8159, "encoder_q-layer.3": 5846.5166, "encoder_q-layer.4": 6546.9912, "encoder_q-layer.5": 7215.6133, "encoder_q-layer.6": 6010.6748, "encoder_q-layer.7": 6220.396, "encoder_q-layer.8": 6168.9082, "encoder_q-layer.9": 2070.1182, "epoch": 0.0, "inbatch_neg_score": 5.6476, "inbatch_pos_score": 6.2148, "learning_rate": 5.500000000000001e-06, "loss": 13.8059, "norm_diff": 4.5564, "norm_loss": 0.0, "num_token_doc": 66.8225, "num_token_overlap": 11.6372, "num_token_query": 31.2592, "num_token_union": 65.1267, "num_word_context": 202.7045, "num_word_doc": 49.8696, "num_word_query": 23.2287, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7811.6892, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.6484, "query_norm": 2.4093, "queue_k_norm": 6.9619, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2592, "sent_len_1": 66.8225, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.7688, "stdk": 0.1268, "stdq": 0.0642, "stdqueue_k": 0.1265, "stdqueue_q": 0.0, "step": 1100 }, { "accuracy": 13.1348, "active_queue_size": 16384.0, "cl_loss": 12.5114, "doc_norm": 6.8377, "encoder_q-embeddings": 3474.7734, "encoder_q-layer.0": 3289.24, "encoder_q-layer.1": 3705.5508, "encoder_q-layer.10": 3977.2559, "encoder_q-layer.11": 6326.1641, "encoder_q-layer.2": 4401.748, "encoder_q-layer.3": 4686.5127, "encoder_q-layer.4": 5150.1416, "encoder_q-layer.5": 5138.4565, "encoder_q-layer.6": 4462.5088, "encoder_q-layer.7": 3655.3223, "encoder_q-layer.8": 3582.5198, "encoder_q-layer.9": 1826.275, "epoch": 0.01, "inbatch_neg_score": 4.8346, "inbatch_pos_score": 5.3438, "learning_rate": 6e-06, "loss": 12.5114, "norm_diff": 4.6343, "norm_loss": 0.0, "num_token_doc": 66.8304, "num_token_overlap": 11.6574, "num_token_query": 31.2618, "num_token_union": 65.0975, "num_word_context": 202.3404, "num_word_doc": 49.8822, "num_word_query": 23.2159, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6142.6891, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8203, "query_norm": 2.2034, "queue_k_norm": 6.8453, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2618, "sent_len_1": 66.8304, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.5888, "stdk": 0.1198, "stdq": 0.0611, "stdqueue_k": 0.1211, "stdqueue_q": 0.0, "step": 1200 }, { "accuracy": 13.9648, "active_queue_size": 16384.0, "cl_loss": 11.7577, "doc_norm": 6.7344, "encoder_q-embeddings": 3395.2688, "encoder_q-layer.0": 3023.6484, "encoder_q-layer.1": 3422.0483, "encoder_q-layer.10": 5496.5952, "encoder_q-layer.11": 7979.3442, "encoder_q-layer.2": 4207.439, "encoder_q-layer.3": 4650.687, "encoder_q-layer.4": 4674.0425, "encoder_q-layer.5": 4556.8901, "encoder_q-layer.6": 4182.0874, "encoder_q-layer.7": 3098.594, "encoder_q-layer.8": 3347.1848, "encoder_q-layer.9": 2434.5051, "epoch": 0.01, "inbatch_neg_score": 3.6323, "inbatch_pos_score": 4.1133, "learning_rate": 6.5000000000000004e-06, "loss": 11.7577, "norm_diff": 4.7201, "norm_loss": 0.0, "num_token_doc": 66.7739, "num_token_overlap": 11.6321, "num_token_query": 31.2365, "num_token_union": 65.0916, "num_word_context": 202.4879, "num_word_doc": 49.8466, "num_word_query": 23.202, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6217.2092, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6191, "query_norm": 2.0143, "queue_k_norm": 6.7356, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2365, "sent_len_1": 66.7739, "sent_len_max_0": 127.955, "sent_len_max_1": 209.2488, "stdk": 0.1152, "stdq": 0.0573, "stdqueue_k": 0.1161, "stdqueue_q": 0.0, "step": 1300 }, { "accuracy": 14.8438, "active_queue_size": 16384.0, "cl_loss": 11.2712, "doc_norm": 6.6293, "encoder_q-embeddings": 1941.1903, "encoder_q-layer.0": 1611.1165, "encoder_q-layer.1": 2086.2622, "encoder_q-layer.10": 2653.4133, "encoder_q-layer.11": 5515.1265, "encoder_q-layer.2": 2589.7747, "encoder_q-layer.3": 2712.9116, "encoder_q-layer.4": 3021.9619, "encoder_q-layer.5": 3559.4641, "encoder_q-layer.6": 3319.5254, "encoder_q-layer.7": 2993.9656, "encoder_q-layer.8": 2938.3389, "encoder_q-layer.9": 1433.5894, "epoch": 0.01, "inbatch_neg_score": 3.1023, "inbatch_pos_score": 3.6094, "learning_rate": 7.000000000000001e-06, "loss": 11.2712, "norm_diff": 4.6306, "norm_loss": 0.0, "num_token_doc": 66.7933, "num_token_overlap": 11.7295, "num_token_query": 31.4674, "num_token_union": 65.1649, "num_word_context": 202.4377, "num_word_doc": 49.8688, "num_word_query": 23.3689, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4177.9626, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.1035, "query_norm": 1.9987, "queue_k_norm": 6.6372, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4674, "sent_len_1": 66.7933, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.0112, "stdk": 0.1107, "stdq": 0.0582, "stdqueue_k": 0.1108, "stdqueue_q": 0.0, "step": 1400 }, { "accuracy": 14.8438, "active_queue_size": 16384.0, "cl_loss": 10.9157, "doc_norm": 6.5311, "encoder_q-embeddings": 2006.0846, "encoder_q-layer.0": 1754.8059, "encoder_q-layer.1": 1927.9735, "encoder_q-layer.10": 4748.2573, "encoder_q-layer.11": 8440.1523, "encoder_q-layer.2": 2237.9712, "encoder_q-layer.3": 2377.1956, "encoder_q-layer.4": 2556.8003, "encoder_q-layer.5": 2633.9404, "encoder_q-layer.6": 2432.8093, "encoder_q-layer.7": 2473.7942, "encoder_q-layer.8": 2739.6965, "encoder_q-layer.9": 2178.8933, "epoch": 0.01, "inbatch_neg_score": 2.8473, "inbatch_pos_score": 3.332, "learning_rate": 7.5e-06, "loss": 10.9157, "norm_diff": 4.5481, "norm_loss": 0.0, "num_token_doc": 66.8964, "num_token_overlap": 11.6679, "num_token_query": 31.3311, "num_token_union": 65.1871, "num_word_context": 202.4967, "num_word_doc": 49.9058, "num_word_query": 23.2619, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4669.4215, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.8379, "query_norm": 1.9829, "queue_k_norm": 6.536, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3311, "sent_len_1": 66.8964, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.1538, "stdk": 0.1053, "stdq": 0.0573, "stdqueue_k": 0.106, "stdqueue_q": 0.0, "step": 1500 }, { "accuracy": 14.3066, "active_queue_size": 16384.0, "cl_loss": 10.6903, "doc_norm": 6.4329, "encoder_q-embeddings": 3932.4836, "encoder_q-layer.0": 3435.7422, "encoder_q-layer.1": 3760.801, "encoder_q-layer.10": 4658.2383, "encoder_q-layer.11": 7225.5464, "encoder_q-layer.2": 4072.5657, "encoder_q-layer.3": 4164.3833, "encoder_q-layer.4": 4256.0234, "encoder_q-layer.5": 3860.4458, "encoder_q-layer.6": 3050.646, "encoder_q-layer.7": 2676.2617, "encoder_q-layer.8": 2891.2224, "encoder_q-layer.9": 2277.3997, "epoch": 0.01, "inbatch_neg_score": 2.7306, "inbatch_pos_score": 3.1992, "learning_rate": 8.000000000000001e-06, "loss": 10.6903, "norm_diff": 4.4501, "norm_loss": 0.0, "num_token_doc": 66.6533, "num_token_overlap": 11.6506, "num_token_query": 31.3314, "num_token_union": 65.0264, "num_word_context": 202.2089, "num_word_doc": 49.7238, "num_word_query": 23.2666, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5702.744, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7266, "query_norm": 1.9828, "queue_k_norm": 6.4343, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3314, "sent_len_1": 66.6533, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.1738, "stdk": 0.1014, "stdq": 0.0586, "stdqueue_k": 0.1016, "stdqueue_q": 0.0, "step": 1600 }, { "accuracy": 16.6504, "active_queue_size": 16384.0, "cl_loss": 10.4135, "doc_norm": 6.3197, "encoder_q-embeddings": 11218.9541, "encoder_q-layer.0": 8949.6572, "encoder_q-layer.1": 10991.8555, "encoder_q-layer.10": 2413.1724, "encoder_q-layer.11": 5071.8623, "encoder_q-layer.2": 12907.04, "encoder_q-layer.3": 12567.1318, "encoder_q-layer.4": 11464.166, "encoder_q-layer.5": 9742.5645, "encoder_q-layer.6": 6732.7539, "encoder_q-layer.7": 4757.0991, "encoder_q-layer.8": 4173.1299, "encoder_q-layer.9": 1487.0763, "epoch": 0.01, "inbatch_neg_score": 2.3593, "inbatch_pos_score": 2.8145, "learning_rate": 8.500000000000002e-06, "loss": 10.4135, "norm_diff": 4.3283, "norm_loss": 0.0, "num_token_doc": 67.0091, "num_token_overlap": 11.6683, "num_token_query": 31.3253, "num_token_union": 65.2372, "num_word_context": 202.386, "num_word_doc": 49.9567, "num_word_query": 23.2635, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13296.1154, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3555, "query_norm": 1.9913, "queue_k_norm": 6.3324, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3253, "sent_len_1": 67.0091, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.0213, "stdk": 0.0964, "stdq": 0.0582, "stdqueue_k": 0.097, "stdqueue_q": 0.0, "step": 1700 }, { "accuracy": 16.4551, "active_queue_size": 16384.0, "cl_loss": 10.1942, "doc_norm": 6.2133, "encoder_q-embeddings": 2626.5427, "encoder_q-layer.0": 2256.5159, "encoder_q-layer.1": 2655.144, "encoder_q-layer.10": 2098.4697, "encoder_q-layer.11": 4539.9814, "encoder_q-layer.2": 3032.45, "encoder_q-layer.3": 3287.8657, "encoder_q-layer.4": 3678.1294, "encoder_q-layer.5": 4197.5918, "encoder_q-layer.6": 3668.665, "encoder_q-layer.7": 3768.9832, "encoder_q-layer.8": 3571.4731, "encoder_q-layer.9": 1382.9806, "epoch": 0.01, "inbatch_neg_score": 1.4071, "inbatch_pos_score": 1.8574, "learning_rate": 9e-06, "loss": 10.1942, "norm_diff": 4.199, "norm_loss": 0.0, "num_token_doc": 66.8128, "num_token_overlap": 11.6977, "num_token_query": 31.421, "num_token_union": 65.1168, "num_word_context": 202.428, "num_word_doc": 49.836, "num_word_query": 23.3364, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4714.8283, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4062, "query_norm": 2.0143, "queue_k_norm": 6.2205, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.421, "sent_len_1": 66.8128, "sent_len_max_0": 127.975, "sent_len_max_1": 209.79, "stdk": 0.0929, "stdq": 0.0594, "stdqueue_k": 0.0931, "stdqueue_q": 0.0, "step": 1800 }, { "accuracy": 15.7227, "active_queue_size": 16384.0, "cl_loss": 9.9819, "doc_norm": 6.0734, "encoder_q-embeddings": 3034.051, "encoder_q-layer.0": 2373.8088, "encoder_q-layer.1": 2719.429, "encoder_q-layer.10": 4390.5835, "encoder_q-layer.11": 6068.6621, "encoder_q-layer.2": 3154.9985, "encoder_q-layer.3": 3428.7666, "encoder_q-layer.4": 3754.167, "encoder_q-layer.5": 4164.2578, "encoder_q-layer.6": 3735.7068, "encoder_q-layer.7": 3969.6819, "encoder_q-layer.8": 4013.9514, "encoder_q-layer.9": 2205.0127, "epoch": 0.01, "inbatch_neg_score": 1.1527, "inbatch_pos_score": 1.5986, "learning_rate": 9.5e-06, "loss": 9.9819, "norm_diff": 4.0381, "norm_loss": 0.0, "num_token_doc": 66.8393, "num_token_overlap": 11.6676, "num_token_query": 31.3344, "num_token_union": 65.0991, "num_word_context": 202.4138, "num_word_doc": 49.8677, "num_word_query": 23.2677, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5248.5342, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.1553, "query_norm": 2.0352, "queue_k_norm": 6.0809, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3344, "sent_len_1": 66.8393, "sent_len_max_0": 127.97, "sent_len_max_1": 209.6413, "stdk": 0.0886, "stdq": 0.0624, "stdqueue_k": 0.0888, "stdqueue_q": 0.0, "step": 1900 }, { "accuracy": 17.7246, "active_queue_size": 16384.0, "cl_loss": 9.5574, "doc_norm": 5.9131, "encoder_q-embeddings": 1043.048, "encoder_q-layer.0": 972.0192, "encoder_q-layer.1": 1042.0078, "encoder_q-layer.10": 1604.2937, "encoder_q-layer.11": 3035.3728, "encoder_q-layer.2": 1164.843, "encoder_q-layer.3": 1195.7158, "encoder_q-layer.4": 1239.647, "encoder_q-layer.5": 1155.9531, "encoder_q-layer.6": 1044.1185, "encoder_q-layer.7": 971.8439, "encoder_q-layer.8": 1177.0261, "encoder_q-layer.9": 943.3303, "epoch": 0.01, "inbatch_neg_score": 1.0575, "inbatch_pos_score": 1.5146, "learning_rate": 1e-05, "loss": 9.5574, "norm_diff": 3.8778, "norm_loss": 0.0, "num_token_doc": 66.9615, "num_token_overlap": 11.7162, "num_token_query": 31.4662, "num_token_union": 65.2779, "num_word_context": 202.4937, "num_word_doc": 49.9801, "num_word_query": 23.3771, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1939.5769, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.0537, "query_norm": 2.0353, "queue_k_norm": 5.9256, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4662, "sent_len_1": 66.9615, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.0987, "stdk": 0.0846, "stdq": 0.0626, "stdqueue_k": 0.0852, "stdqueue_q": 0.0, "step": 2000 }, { "accuracy": 19.043, "active_queue_size": 16384.0, "cl_loss": 9.1205, "doc_norm": 5.7382, "encoder_q-embeddings": 944.1782, "encoder_q-layer.0": 829.7369, "encoder_q-layer.1": 985.3349, "encoder_q-layer.10": 3499.3914, "encoder_q-layer.11": 4228.3911, "encoder_q-layer.2": 1240.6857, "encoder_q-layer.3": 1370.2714, "encoder_q-layer.4": 1575.2197, "encoder_q-layer.5": 1894.2675, "encoder_q-layer.6": 2042.4446, "encoder_q-layer.7": 2269.0771, "encoder_q-layer.8": 2380.0156, "encoder_q-layer.9": 2224.8242, "epoch": 0.01, "inbatch_neg_score": 1.202, "inbatch_pos_score": 1.6396, "learning_rate": 1.05e-05, "loss": 9.1205, "norm_diff": 3.7509, "norm_loss": 0.0, "num_token_doc": 66.6218, "num_token_overlap": 11.6768, "num_token_query": 31.3668, "num_token_union": 65.0358, "num_word_context": 202.1016, "num_word_doc": 49.7191, "num_word_query": 23.289, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2878.676, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.1973, "query_norm": 1.9872, "queue_k_norm": 5.7486, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3668, "sent_len_1": 66.6218, "sent_len_max_0": 127.97, "sent_len_max_1": 208.81, "stdk": 0.0813, "stdq": 0.0603, "stdqueue_k": 0.0818, "stdqueue_q": 0.0, "step": 2100 }, { "accuracy": 16.5039, "active_queue_size": 16384.0, "cl_loss": 8.8429, "doc_norm": 5.5414, "encoder_q-embeddings": 1193.7643, "encoder_q-layer.0": 1083.324, "encoder_q-layer.1": 1684.4711, "encoder_q-layer.10": 6908.6738, "encoder_q-layer.11": 7738.7134, "encoder_q-layer.2": 2538.551, "encoder_q-layer.3": 3092.8323, "encoder_q-layer.4": 4487.834, "encoder_q-layer.5": 6317.8789, "encoder_q-layer.6": 7807.9653, "encoder_q-layer.7": 8516.1904, "encoder_q-layer.8": 8120.6611, "encoder_q-layer.9": 5461.7256, "epoch": 0.01, "inbatch_neg_score": 1.5509, "inbatch_pos_score": 1.9893, "learning_rate": 1.1000000000000001e-05, "loss": 8.8429, "norm_diff": 3.5061, "norm_loss": 0.0, "num_token_doc": 66.8914, "num_token_overlap": 11.7017, "num_token_query": 31.3939, "num_token_union": 65.2102, "num_word_context": 202.3652, "num_word_doc": 49.8995, "num_word_query": 23.3057, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7643.7817, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5498, "query_norm": 2.0353, "queue_k_norm": 5.5496, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3939, "sent_len_1": 66.8914, "sent_len_max_0": 127.975, "sent_len_max_1": 209.8313, "stdk": 0.0779, "stdq": 0.0614, "stdqueue_k": 0.0782, "stdqueue_q": 0.0, "step": 2200 }, { "accuracy": 16.3086, "active_queue_size": 16384.0, "cl_loss": 8.7086, "doc_norm": 5.3191, "encoder_q-embeddings": 2442.5627, "encoder_q-layer.0": 2269.5015, "encoder_q-layer.1": 3589.5884, "encoder_q-layer.10": 16017.6797, "encoder_q-layer.11": 15507.4199, "encoder_q-layer.2": 5504.2856, "encoder_q-layer.3": 6611.0146, "encoder_q-layer.4": 9501.4385, "encoder_q-layer.5": 13519.0039, "encoder_q-layer.6": 16639.7559, "encoder_q-layer.7": 18193.918, "encoder_q-layer.8": 17162.0996, "encoder_q-layer.9": 12697.9326, "epoch": 0.01, "inbatch_neg_score": 0.4734, "inbatch_pos_score": 0.9053, "learning_rate": 1.1500000000000002e-05, "loss": 8.7086, "norm_diff": 3.3003, "norm_loss": 0.0, "num_token_doc": 66.8563, "num_token_overlap": 11.7427, "num_token_query": 31.4781, "num_token_union": 65.1911, "num_word_context": 202.2858, "num_word_doc": 49.8984, "num_word_query": 23.3797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16463.0753, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.4709, "query_norm": 2.0188, "queue_k_norm": 5.335, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4781, "sent_len_1": 66.8563, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.76, "stdk": 0.0754, "stdq": 0.0624, "stdqueue_k": 0.0755, "stdqueue_q": 0.0, "step": 2300 }, { "accuracy": 15.9668, "active_queue_size": 16384.0, "cl_loss": 8.52, "doc_norm": 5.1114, "encoder_q-embeddings": 1540.6632, "encoder_q-layer.0": 1466.1143, "encoder_q-layer.1": 2218.8921, "encoder_q-layer.10": 10648.7881, "encoder_q-layer.11": 10580.502, "encoder_q-layer.2": 3351.2083, "encoder_q-layer.3": 4162.5337, "encoder_q-layer.4": 6343.4863, "encoder_q-layer.5": 9131.8525, "encoder_q-layer.6": 11697.8799, "encoder_q-layer.7": 13032.7852, "encoder_q-layer.8": 12188.9678, "encoder_q-layer.9": 9598.3955, "epoch": 0.01, "inbatch_neg_score": 0.365, "inbatch_pos_score": 0.7676, "learning_rate": 1.2e-05, "loss": 8.52, "norm_diff": 3.1481, "norm_loss": 0.0, "num_token_doc": 66.7791, "num_token_overlap": 11.6572, "num_token_query": 31.3254, "num_token_union": 65.0927, "num_word_context": 202.0539, "num_word_doc": 49.8298, "num_word_query": 23.2623, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11459.3925, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3625, "query_norm": 1.9633, "queue_k_norm": 5.1103, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3254, "sent_len_1": 66.7791, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.465, "stdk": 0.0723, "stdq": 0.0581, "stdqueue_k": 0.0727, "stdqueue_q": 0.0, "step": 2400 }, { "accuracy": 19.4336, "active_queue_size": 16384.0, "cl_loss": 8.139, "doc_norm": 4.8836, "encoder_q-embeddings": 994.6444, "encoder_q-layer.0": 934.4149, "encoder_q-layer.1": 1334.2285, "encoder_q-layer.10": 6812.2954, "encoder_q-layer.11": 7083.4146, "encoder_q-layer.2": 1990.8895, "encoder_q-layer.3": 2493.8997, "encoder_q-layer.4": 3856.6311, "encoder_q-layer.5": 5808.0659, "encoder_q-layer.6": 7158.1328, "encoder_q-layer.7": 7850.3018, "encoder_q-layer.8": 7473.8579, "encoder_q-layer.9": 5823.6172, "epoch": 0.01, "inbatch_neg_score": 0.9426, "inbatch_pos_score": 1.374, "learning_rate": 1.25e-05, "loss": 8.139, "norm_diff": 2.9806, "norm_loss": 0.0, "num_token_doc": 66.7065, "num_token_overlap": 11.687, "num_token_query": 31.4528, "num_token_union": 65.1404, "num_word_context": 202.4393, "num_word_doc": 49.7987, "num_word_query": 23.3645, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7064.9847, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.939, "query_norm": 1.9029, "queue_k_norm": 4.883, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4528, "sent_len_1": 66.7065, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.3013, "stdk": 0.0698, "stdq": 0.0583, "stdqueue_k": 0.0703, "stdqueue_q": 0.0, "step": 2500 }, { "accuracy": 17.4805, "active_queue_size": 16384.0, "cl_loss": 7.8796, "doc_norm": 4.6319, "encoder_q-embeddings": 4522.4658, "encoder_q-layer.0": 4308.4878, "encoder_q-layer.1": 6666.6543, "encoder_q-layer.10": 33176.0469, "encoder_q-layer.11": 30608.9492, "encoder_q-layer.2": 10398.1152, "encoder_q-layer.3": 13541.1807, "encoder_q-layer.4": 21262.9238, "encoder_q-layer.5": 32569.7324, "encoder_q-layer.6": 41583.3164, "encoder_q-layer.7": 46779.9102, "encoder_q-layer.8": 43281.25, "encoder_q-layer.9": 32010.4688, "epoch": 0.01, "inbatch_neg_score": 0.3075, "inbatch_pos_score": 0.731, "learning_rate": 1.3000000000000001e-05, "loss": 7.8796, "norm_diff": 2.7317, "norm_loss": 0.0, "num_token_doc": 66.8854, "num_token_overlap": 11.6584, "num_token_query": 31.2795, "num_token_union": 65.1283, "num_word_context": 202.5218, "num_word_doc": 49.9031, "num_word_query": 23.2263, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39131.1189, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 0.302, "query_norm": 1.9002, "queue_k_norm": 4.661, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2795, "sent_len_1": 66.8854, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7775, "stdk": 0.0679, "stdq": 0.0577, "stdqueue_k": 0.0682, "stdqueue_q": 0.0, "step": 2600 }, { "accuracy": 18.8965, "active_queue_size": 16384.0, "cl_loss": 7.8461, "doc_norm": 4.4476, "encoder_q-embeddings": 2758.6819, "encoder_q-layer.0": 2570.8098, "encoder_q-layer.1": 4013.2871, "encoder_q-layer.10": 17532.1328, "encoder_q-layer.11": 17909.5625, "encoder_q-layer.2": 6446.0894, "encoder_q-layer.3": 7927.6147, "encoder_q-layer.4": 11835.3643, "encoder_q-layer.5": 18333.0488, "encoder_q-layer.6": 22753.377, "encoder_q-layer.7": 24638.2539, "encoder_q-layer.8": 23242.5176, "encoder_q-layer.9": 16849.2363, "epoch": 0.01, "inbatch_neg_score": 1.0879, "inbatch_pos_score": 1.5176, "learning_rate": 1.3500000000000001e-05, "loss": 7.8461, "norm_diff": 2.555, "norm_loss": 0.0, "num_token_doc": 66.7837, "num_token_overlap": 11.6369, "num_token_query": 31.3064, "num_token_union": 65.1016, "num_word_context": 202.047, "num_word_doc": 49.8513, "num_word_query": 23.2478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21255.347, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0869, "query_norm": 1.8926, "queue_k_norm": 4.4425, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3064, "sent_len_1": 66.7837, "sent_len_max_0": 128.0, "sent_len_max_1": 206.9175, "stdk": 0.0661, "stdq": 0.0563, "stdqueue_k": 0.0659, "stdqueue_q": 0.0, "step": 2700 }, { "accuracy": 19.0918, "active_queue_size": 16384.0, "cl_loss": 7.6469, "doc_norm": 4.2336, "encoder_q-embeddings": 2181.1882, "encoder_q-layer.0": 2058.9729, "encoder_q-layer.1": 3107.3892, "encoder_q-layer.10": 14214.8486, "encoder_q-layer.11": 13716.6924, "encoder_q-layer.2": 4900.9868, "encoder_q-layer.3": 6052.4146, "encoder_q-layer.4": 9295.4727, "encoder_q-layer.5": 14217.4648, "encoder_q-layer.6": 17462.5293, "encoder_q-layer.7": 18779.2988, "encoder_q-layer.8": 17886.3555, "encoder_q-layer.9": 13687.1787, "epoch": 0.01, "inbatch_neg_score": 0.4795, "inbatch_pos_score": 0.8926, "learning_rate": 1.4000000000000001e-05, "loss": 7.6469, "norm_diff": 2.3511, "norm_loss": 0.0, "num_token_doc": 66.7978, "num_token_overlap": 11.7025, "num_token_query": 31.4174, "num_token_union": 65.1289, "num_word_context": 201.9597, "num_word_doc": 49.8323, "num_word_query": 23.3322, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16538.0908, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.4795, "query_norm": 1.8825, "queue_k_norm": 4.2302, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4174, "sent_len_1": 66.7978, "sent_len_max_0": 127.97, "sent_len_max_1": 211.0325, "stdk": 0.0642, "stdq": 0.0564, "stdqueue_k": 0.0642, "stdqueue_q": 0.0, "step": 2800 }, { "accuracy": 19.4336, "active_queue_size": 16384.0, "cl_loss": 7.5275, "doc_norm": 4.0133, "encoder_q-embeddings": 2663.4236, "encoder_q-layer.0": 2360.301, "encoder_q-layer.1": 3618.5332, "encoder_q-layer.10": 15111.5254, "encoder_q-layer.11": 14976.5039, "encoder_q-layer.2": 5765.3628, "encoder_q-layer.3": 7024.8799, "encoder_q-layer.4": 10680.7979, "encoder_q-layer.5": 16509.8672, "encoder_q-layer.6": 20434.0605, "encoder_q-layer.7": 22358.002, "encoder_q-layer.8": 21079.9121, "encoder_q-layer.9": 15437.499, "epoch": 0.01, "inbatch_neg_score": 0.4958, "inbatch_pos_score": 0.916, "learning_rate": 1.45e-05, "loss": 7.5275, "norm_diff": 2.1442, "norm_loss": 0.0, "num_token_doc": 66.6932, "num_token_overlap": 11.6622, "num_token_query": 31.3057, "num_token_union": 65.0309, "num_word_context": 202.0296, "num_word_doc": 49.7797, "num_word_query": 23.2557, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19171.3941, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.4961, "query_norm": 1.8691, "queue_k_norm": 4.0395, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3057, "sent_len_1": 66.6932, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.2738, "stdk": 0.0631, "stdq": 0.0558, "stdqueue_k": 0.0625, "stdqueue_q": 0.0, "step": 2900 }, { "accuracy": 18.3594, "active_queue_size": 16384.0, "cl_loss": 7.3868, "doc_norm": 3.8494, "encoder_q-embeddings": 3136.8445, "encoder_q-layer.0": 2832.1504, "encoder_q-layer.1": 3991.4397, "encoder_q-layer.10": 17058.6035, "encoder_q-layer.11": 17449.3535, "encoder_q-layer.2": 6372.332, "encoder_q-layer.3": 7738.9385, "encoder_q-layer.4": 11528.8389, "encoder_q-layer.5": 17973.7363, "encoder_q-layer.6": 21949.6172, "encoder_q-layer.7": 23868.1523, "encoder_q-layer.8": 22637.7969, "encoder_q-layer.9": 17169.5605, "epoch": 0.01, "inbatch_neg_score": 1.0733, "inbatch_pos_score": 1.4932, "learning_rate": 1.5e-05, "loss": 7.3868, "norm_diff": 1.9509, "norm_loss": 0.0, "num_token_doc": 66.8107, "num_token_overlap": 11.6613, "num_token_query": 31.3897, "num_token_union": 65.1577, "num_word_context": 202.2743, "num_word_doc": 49.8435, "num_word_query": 23.3158, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20826.913, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0693, "query_norm": 1.8985, "queue_k_norm": 3.8562, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3897, "sent_len_1": 66.8107, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.3075, "stdk": 0.0608, "stdq": 0.0566, "stdqueue_k": 0.0613, "stdqueue_q": 0.0, "step": 3000 }, { "accuracy": 23.4863, "active_queue_size": 16384.0, "cl_loss": 7.2279, "doc_norm": 3.6735, "encoder_q-embeddings": 1506.7075, "encoder_q-layer.0": 1320.8055, "encoder_q-layer.1": 1777.8762, "encoder_q-layer.10": 5455.0801, "encoder_q-layer.11": 5990.5059, "encoder_q-layer.2": 2473.1721, "encoder_q-layer.3": 2902.8108, "encoder_q-layer.4": 4443.4736, "encoder_q-layer.5": 6746.7217, "encoder_q-layer.6": 7974.5601, "encoder_q-layer.7": 8344.5732, "encoder_q-layer.8": 8113.8398, "encoder_q-layer.9": 5846.6201, "epoch": 0.01, "inbatch_neg_score": 0.6681, "inbatch_pos_score": 1.0947, "learning_rate": 1.55e-05, "loss": 7.2279, "norm_diff": 1.7928, "norm_loss": 0.0, "num_token_doc": 66.7821, "num_token_overlap": 11.6813, "num_token_query": 31.3055, "num_token_union": 65.0837, "num_word_context": 202.2853, "num_word_doc": 49.8186, "num_word_query": 23.2419, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7499.7398, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.666, "query_norm": 1.8807, "queue_k_norm": 3.685, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3055, "sent_len_1": 66.7821, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.0525, "stdk": 0.0601, "stdq": 0.0548, "stdqueue_k": 0.0601, "stdqueue_q": 0.0, "step": 3100 }, { "accuracy": 20.9961, "active_queue_size": 16384.0, "cl_loss": 7.049, "doc_norm": 3.5156, "encoder_q-embeddings": 1908.4315, "encoder_q-layer.0": 1826.8334, "encoder_q-layer.1": 2603.4734, "encoder_q-layer.10": 12099.6758, "encoder_q-layer.11": 12036.1836, "encoder_q-layer.2": 3879.7759, "encoder_q-layer.3": 4815.7583, "encoder_q-layer.4": 7397.9263, "encoder_q-layer.5": 11663.1328, "encoder_q-layer.6": 14289.5293, "encoder_q-layer.7": 15766.873, "encoder_q-layer.8": 15013.8633, "encoder_q-layer.9": 11408.2822, "epoch": 0.01, "inbatch_neg_score": 0.5427, "inbatch_pos_score": 0.9624, "learning_rate": 1.6000000000000003e-05, "loss": 7.049, "norm_diff": 1.6641, "norm_loss": 0.0, "num_token_doc": 66.8063, "num_token_overlap": 11.6785, "num_token_query": 31.402, "num_token_union": 65.1468, "num_word_context": 202.1379, "num_word_doc": 49.8191, "num_word_query": 23.3157, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13858.8266, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.543, "query_norm": 1.8515, "queue_k_norm": 3.5288, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.402, "sent_len_1": 66.8063, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.6662, "stdk": 0.0588, "stdq": 0.0549, "stdqueue_k": 0.0588, "stdqueue_q": 0.0, "step": 3200 }, { "accuracy": 22.4609, "active_queue_size": 16384.0, "cl_loss": 6.8656, "doc_norm": 3.3625, "encoder_q-embeddings": 1653.4043, "encoder_q-layer.0": 1473.549, "encoder_q-layer.1": 2200.2488, "encoder_q-layer.10": 10576.6826, "encoder_q-layer.11": 11819.1055, "encoder_q-layer.2": 3331.9346, "encoder_q-layer.3": 4018.6499, "encoder_q-layer.4": 6135.376, "encoder_q-layer.5": 9350.585, "encoder_q-layer.6": 11687.9727, "encoder_q-layer.7": 13043.0762, "encoder_q-layer.8": 12536.6582, "encoder_q-layer.9": 9740.918, "epoch": 0.01, "inbatch_neg_score": 0.9193, "inbatch_pos_score": 1.3691, "learning_rate": 1.65e-05, "loss": 6.8656, "norm_diff": 1.5062, "norm_loss": 0.0, "num_token_doc": 66.7564, "num_token_overlap": 11.7009, "num_token_query": 31.4535, "num_token_union": 65.1244, "num_word_context": 202.0431, "num_word_doc": 49.8319, "num_word_query": 23.3676, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11737.0682, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9175, "query_norm": 1.8563, "queue_k_norm": 3.371, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4535, "sent_len_1": 66.7564, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.5175, "stdk": 0.0583, "stdq": 0.0546, "stdqueue_k": 0.0581, "stdqueue_q": 0.0, "step": 3300 }, { "accuracy": 24.1699, "active_queue_size": 16384.0, "cl_loss": 6.6842, "doc_norm": 3.2258, "encoder_q-embeddings": 1246.9934, "encoder_q-layer.0": 1174.1868, "encoder_q-layer.1": 1655.3883, "encoder_q-layer.10": 7603.4551, "encoder_q-layer.11": 8164.0327, "encoder_q-layer.2": 2508.2595, "encoder_q-layer.3": 2942.2087, "encoder_q-layer.4": 4354.2554, "encoder_q-layer.5": 6942.2104, "encoder_q-layer.6": 8487.9883, "encoder_q-layer.7": 9422.4355, "encoder_q-layer.8": 8978.0391, "encoder_q-layer.9": 7149.1592, "epoch": 0.01, "inbatch_neg_score": 0.5774, "inbatch_pos_score": 1.0137, "learning_rate": 1.7000000000000003e-05, "loss": 6.6842, "norm_diff": 1.432, "norm_loss": 0.0, "num_token_doc": 66.7589, "num_token_overlap": 11.6447, "num_token_query": 31.2465, "num_token_union": 65.0329, "num_word_context": 202.0298, "num_word_doc": 49.8083, "num_word_query": 23.1959, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8520.8038, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.5771, "query_norm": 1.7938, "queue_k_norm": 3.2265, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2465, "sent_len_1": 66.7589, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.5263, "stdk": 0.0575, "stdq": 0.0533, "stdqueue_k": 0.0572, "stdqueue_q": 0.0, "step": 3400 }, { "accuracy": 24.1211, "active_queue_size": 16384.0, "cl_loss": 6.4788, "doc_norm": 3.085, "encoder_q-embeddings": 1331.2938, "encoder_q-layer.0": 1188.0468, "encoder_q-layer.1": 1680.4355, "encoder_q-layer.10": 7620.4907, "encoder_q-layer.11": 8151.0068, "encoder_q-layer.2": 2454.3201, "encoder_q-layer.3": 2857.0796, "encoder_q-layer.4": 4307.0542, "encoder_q-layer.5": 6868.8794, "encoder_q-layer.6": 8719.3828, "encoder_q-layer.7": 10080.293, "encoder_q-layer.8": 9521.4082, "encoder_q-layer.9": 7022.1553, "epoch": 0.02, "inbatch_neg_score": 0.6468, "inbatch_pos_score": 1.0654, "learning_rate": 1.75e-05, "loss": 6.4788, "norm_diff": 1.3289, "norm_loss": 0.0, "num_token_doc": 66.8683, "num_token_overlap": 11.7106, "num_token_query": 31.4207, "num_token_union": 65.2023, "num_word_context": 202.5108, "num_word_doc": 49.92, "num_word_query": 23.3426, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8715.1178, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.645, "query_norm": 1.7561, "queue_k_norm": 3.0947, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4207, "sent_len_1": 66.8683, "sent_len_max_0": 127.9688, "sent_len_max_1": 207.9137, "stdk": 0.0561, "stdq": 0.0523, "stdqueue_k": 0.0562, "stdqueue_q": 0.0, "step": 3500 }, { "accuracy": 27.1484, "active_queue_size": 16384.0, "cl_loss": 6.3639, "doc_norm": 2.9715, "encoder_q-embeddings": 1424.3727, "encoder_q-layer.0": 1349.0222, "encoder_q-layer.1": 1831.6906, "encoder_q-layer.10": 8741.9805, "encoder_q-layer.11": 9753.4229, "encoder_q-layer.2": 2647.385, "encoder_q-layer.3": 3232.8088, "encoder_q-layer.4": 4767.2944, "encoder_q-layer.5": 7483.7119, "encoder_q-layer.6": 9084.5771, "encoder_q-layer.7": 10022.9863, "encoder_q-layer.8": 9568.7402, "encoder_q-layer.9": 7717.5459, "epoch": 0.02, "inbatch_neg_score": 0.8245, "inbatch_pos_score": 1.2617, "learning_rate": 1.8e-05, "loss": 6.3639, "norm_diff": 1.2291, "norm_loss": 0.0, "num_token_doc": 66.7153, "num_token_overlap": 11.6588, "num_token_query": 31.3026, "num_token_union": 65.0549, "num_word_context": 202.2833, "num_word_doc": 49.8025, "num_word_query": 23.2275, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9268.7489, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8228, "query_norm": 1.7423, "queue_k_norm": 2.9655, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3026, "sent_len_1": 66.7153, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.9913, "stdk": 0.0557, "stdq": 0.0519, "stdqueue_k": 0.0556, "stdqueue_q": 0.0, "step": 3600 }, { "accuracy": 26.0742, "active_queue_size": 16384.0, "cl_loss": 6.203, "doc_norm": 2.8296, "encoder_q-embeddings": 1183.5391, "encoder_q-layer.0": 1117.5503, "encoder_q-layer.1": 1349.6782, "encoder_q-layer.10": 4776.8213, "encoder_q-layer.11": 5685.1753, "encoder_q-layer.2": 1759.0076, "encoder_q-layer.3": 2051.8521, "encoder_q-layer.4": 3079.2031, "encoder_q-layer.5": 4572.6821, "encoder_q-layer.6": 5631.1587, "encoder_q-layer.7": 6266.1772, "encoder_q-layer.8": 6024.2539, "encoder_q-layer.9": 4606.6689, "epoch": 0.02, "inbatch_neg_score": 0.5897, "inbatch_pos_score": 1.0234, "learning_rate": 1.85e-05, "loss": 6.203, "norm_diff": 1.1086, "norm_loss": 0.0, "num_token_doc": 66.8997, "num_token_overlap": 11.71, "num_token_query": 31.3831, "num_token_union": 65.1682, "num_word_context": 202.2895, "num_word_doc": 49.9131, "num_word_query": 23.2945, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5728.8428, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.5889, "query_norm": 1.7209, "queue_k_norm": 2.8467, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3831, "sent_len_1": 66.8997, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.7038, "stdk": 0.0551, "stdq": 0.0508, "stdqueue_k": 0.0549, "stdqueue_q": 0.0, "step": 3700 }, { "accuracy": 25.0, "active_queue_size": 16384.0, "cl_loss": 6.0907, "doc_norm": 2.7223, "encoder_q-embeddings": 1114.33, "encoder_q-layer.0": 971.4221, "encoder_q-layer.1": 1299.3229, "encoder_q-layer.10": 6549.9985, "encoder_q-layer.11": 8220.7432, "encoder_q-layer.2": 1956.0759, "encoder_q-layer.3": 2221.7585, "encoder_q-layer.4": 3233.074, "encoder_q-layer.5": 5089.7349, "encoder_q-layer.6": 6287.145, "encoder_q-layer.7": 7176.8433, "encoder_q-layer.8": 6960.1431, "encoder_q-layer.9": 5766.8291, "epoch": 0.02, "inbatch_neg_score": 0.5777, "inbatch_pos_score": 1.0029, "learning_rate": 1.9e-05, "loss": 6.0907, "norm_diff": 1.0161, "norm_loss": 0.0, "num_token_doc": 66.7772, "num_token_overlap": 11.7144, "num_token_query": 31.4285, "num_token_union": 65.1229, "num_word_context": 202.3142, "num_word_doc": 49.8412, "num_word_query": 23.3429, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6936.5014, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.5752, "query_norm": 1.7062, "queue_k_norm": 2.7244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4285, "sent_len_1": 66.7772, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.1075, "stdk": 0.0539, "stdq": 0.0512, "stdqueue_k": 0.0542, "stdqueue_q": 0.0, "step": 3800 }, { "accuracy": 27.5879, "active_queue_size": 16384.0, "cl_loss": 5.9911, "doc_norm": 2.6214, "encoder_q-embeddings": 1073.4116, "encoder_q-layer.0": 950.4114, "encoder_q-layer.1": 1277.681, "encoder_q-layer.10": 5566.373, "encoder_q-layer.11": 6551.5806, "encoder_q-layer.2": 1927.3972, "encoder_q-layer.3": 2190.5732, "encoder_q-layer.4": 3229.6719, "encoder_q-layer.5": 5153.0513, "encoder_q-layer.6": 6265.5229, "encoder_q-layer.7": 6814.2271, "encoder_q-layer.8": 6571.0195, "encoder_q-layer.9": 5201.7197, "epoch": 0.02, "inbatch_neg_score": 0.7289, "inbatch_pos_score": 1.1758, "learning_rate": 1.9500000000000003e-05, "loss": 5.9911, "norm_diff": 0.9321, "norm_loss": 0.0, "num_token_doc": 66.6983, "num_token_overlap": 11.6673, "num_token_query": 31.3703, "num_token_union": 65.0669, "num_word_context": 202.1342, "num_word_doc": 49.783, "num_word_query": 23.3019, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6292.5694, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.7251, "query_norm": 1.6893, "queue_k_norm": 2.6156, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3703, "sent_len_1": 66.6983, "sent_len_max_0": 127.98, "sent_len_max_1": 207.8475, "stdk": 0.0533, "stdq": 0.05, "stdqueue_k": 0.0537, "stdqueue_q": 0.0, "step": 3900 }, { "accuracy": 27.2949, "active_queue_size": 16384.0, "cl_loss": 5.8769, "doc_norm": 2.5009, "encoder_q-embeddings": 809.189, "encoder_q-layer.0": 705.4789, "encoder_q-layer.1": 810.275, "encoder_q-layer.10": 2805.6375, "encoder_q-layer.11": 3611.4824, "encoder_q-layer.2": 1115.7587, "encoder_q-layer.3": 1168.9426, "encoder_q-layer.4": 1403.676, "encoder_q-layer.5": 2055.6064, "encoder_q-layer.6": 2525.6406, "encoder_q-layer.7": 2998.1699, "encoder_q-layer.8": 3076.0906, "encoder_q-layer.9": 2497.3289, "epoch": 0.02, "inbatch_neg_score": 0.5773, "inbatch_pos_score": 1.0098, "learning_rate": 2e-05, "loss": 5.8769, "norm_diff": 0.8124, "norm_loss": 0.0, "num_token_doc": 66.6661, "num_token_overlap": 11.6378, "num_token_query": 31.3579, "num_token_union": 65.0799, "num_word_context": 202.1364, "num_word_doc": 49.6968, "num_word_query": 23.2737, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2985.4258, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5752, "query_norm": 1.6885, "queue_k_norm": 2.5082, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3579, "sent_len_1": 66.6661, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.9638, "stdk": 0.053, "stdq": 0.0502, "stdqueue_k": 0.0531, "stdqueue_q": 0.0, "step": 4000 }, { "accuracy": 30.4199, "active_queue_size": 16384.0, "cl_loss": 5.7437, "doc_norm": 2.4019, "encoder_q-embeddings": 649.8817, "encoder_q-layer.0": 477.9779, "encoder_q-layer.1": 493.7279, "encoder_q-layer.10": 1519.0074, "encoder_q-layer.11": 3043.5286, "encoder_q-layer.2": 545.2213, "encoder_q-layer.3": 603.8825, "encoder_q-layer.4": 632.644, "encoder_q-layer.5": 601.6427, "encoder_q-layer.6": 642.5562, "encoder_q-layer.7": 684.4003, "encoder_q-layer.8": 886.7988, "encoder_q-layer.9": 906.8062, "epoch": 0.02, "inbatch_neg_score": 0.5731, "inbatch_pos_score": 1.0254, "learning_rate": 2.05e-05, "loss": 5.7437, "norm_diff": 0.7266, "norm_loss": 0.0, "num_token_doc": 66.596, "num_token_overlap": 11.6833, "num_token_query": 31.4132, "num_token_union": 65.0308, "num_word_context": 202.3734, "num_word_doc": 49.7037, "num_word_query": 23.3294, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1655.4925, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5723, "query_norm": 1.6753, "queue_k_norm": 2.4083, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4132, "sent_len_1": 66.596, "sent_len_max_0": 127.995, "sent_len_max_1": 208.1687, "stdk": 0.0525, "stdq": 0.0488, "stdqueue_k": 0.0526, "stdqueue_q": 0.0, "step": 4100 }, { "accuracy": 28.8574, "active_queue_size": 16384.0, "cl_loss": 5.6149, "doc_norm": 2.3081, "encoder_q-embeddings": 663.5416, "encoder_q-layer.0": 567.3252, "encoder_q-layer.1": 557.2503, "encoder_q-layer.10": 1836.6592, "encoder_q-layer.11": 2695.0879, "encoder_q-layer.2": 662.3602, "encoder_q-layer.3": 664.9254, "encoder_q-layer.4": 749.9355, "encoder_q-layer.5": 1018.6776, "encoder_q-layer.6": 1134.3307, "encoder_q-layer.7": 1312.5547, "encoder_q-layer.8": 1426.8489, "encoder_q-layer.9": 1297.7581, "epoch": 0.02, "inbatch_neg_score": 0.6182, "inbatch_pos_score": 1.0703, "learning_rate": 2.1e-05, "loss": 5.6149, "norm_diff": 0.6408, "norm_loss": 0.0, "num_token_doc": 66.7697, "num_token_overlap": 11.6766, "num_token_query": 31.3168, "num_token_union": 65.1077, "num_word_context": 202.4602, "num_word_doc": 49.8332, "num_word_query": 23.2467, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1766.443, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6172, "query_norm": 1.6673, "queue_k_norm": 2.3123, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3168, "sent_len_1": 66.7697, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.1087, "stdk": 0.0519, "stdq": 0.0488, "stdqueue_k": 0.0521, "stdqueue_q": 0.0, "step": 4200 }, { "accuracy": 29.3945, "active_queue_size": 16384.0, "cl_loss": 5.5016, "doc_norm": 2.2234, "encoder_q-embeddings": 754.5923, "encoder_q-layer.0": 660.0334, "encoder_q-layer.1": 714.7299, "encoder_q-layer.10": 1855.4226, "encoder_q-layer.11": 4013.1841, "encoder_q-layer.2": 786.4949, "encoder_q-layer.3": 724.0337, "encoder_q-layer.4": 711.0495, "encoder_q-layer.5": 671.108, "encoder_q-layer.6": 695.0261, "encoder_q-layer.7": 734.3057, "encoder_q-layer.8": 897.4301, "encoder_q-layer.9": 937.3898, "epoch": 0.02, "inbatch_neg_score": 0.5557, "inbatch_pos_score": 1.0029, "learning_rate": 2.15e-05, "loss": 5.5016, "norm_diff": 0.5779, "norm_loss": 0.0, "num_token_doc": 66.994, "num_token_overlap": 11.6811, "num_token_query": 31.4284, "num_token_union": 65.2998, "num_word_context": 202.5168, "num_word_doc": 49.9687, "num_word_query": 23.3447, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2173.0916, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5547, "query_norm": 1.6455, "queue_k_norm": 2.226, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4284, "sent_len_1": 66.994, "sent_len_max_0": 127.995, "sent_len_max_1": 210.05, "stdk": 0.0515, "stdq": 0.048, "stdqueue_k": 0.0516, "stdqueue_q": 0.0, "step": 4300 }, { "accuracy": 29.3457, "active_queue_size": 16384.0, "cl_loss": 5.42, "doc_norm": 2.1335, "encoder_q-embeddings": 604.2689, "encoder_q-layer.0": 479.8289, "encoder_q-layer.1": 495.1383, "encoder_q-layer.10": 1129.8418, "encoder_q-layer.11": 2174.4854, "encoder_q-layer.2": 566.0504, "encoder_q-layer.3": 558.6837, "encoder_q-layer.4": 522.772, "encoder_q-layer.5": 514.0818, "encoder_q-layer.6": 567.394, "encoder_q-layer.7": 576.955, "encoder_q-layer.8": 724.5721, "encoder_q-layer.9": 726.5837, "epoch": 0.02, "inbatch_neg_score": 0.5529, "inbatch_pos_score": 0.9927, "learning_rate": 2.2000000000000003e-05, "loss": 5.42, "norm_diff": 0.4925, "norm_loss": 0.0, "num_token_doc": 66.7282, "num_token_overlap": 11.6777, "num_token_query": 31.2834, "num_token_union": 65.0398, "num_word_context": 201.9383, "num_word_doc": 49.7977, "num_word_query": 23.2167, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1291.3679, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5537, "query_norm": 1.641, "queue_k_norm": 2.1467, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2834, "sent_len_1": 66.7282, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.025, "stdk": 0.051, "stdq": 0.048, "stdqueue_k": 0.0511, "stdqueue_q": 0.0, "step": 4400 }, { "accuracy": 31.1523, "active_queue_size": 16384.0, "cl_loss": 5.374, "doc_norm": 2.0687, "encoder_q-embeddings": 537.1549, "encoder_q-layer.0": 411.3494, "encoder_q-layer.1": 448.891, "encoder_q-layer.10": 952.0759, "encoder_q-layer.11": 1829.9973, "encoder_q-layer.2": 493.1574, "encoder_q-layer.3": 510.6607, "encoder_q-layer.4": 490.0735, "encoder_q-layer.5": 462.3017, "encoder_q-layer.6": 500.3927, "encoder_q-layer.7": 526.9885, "encoder_q-layer.8": 644.1806, "encoder_q-layer.9": 694.9036, "epoch": 0.02, "inbatch_neg_score": 0.5611, "inbatch_pos_score": 1.0107, "learning_rate": 2.25e-05, "loss": 5.374, "norm_diff": 0.4312, "norm_loss": 0.0, "num_token_doc": 66.8031, "num_token_overlap": 11.6941, "num_token_query": 31.4574, "num_token_union": 65.1394, "num_word_context": 202.316, "num_word_doc": 49.8271, "num_word_query": 23.3836, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1106.7501, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5605, "query_norm": 1.6375, "queue_k_norm": 2.0729, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4574, "sent_len_1": 66.8031, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.3837, "stdk": 0.0504, "stdq": 0.048, "stdqueue_k": 0.0505, "stdqueue_q": 0.0, "step": 4500 }, { "accuracy": 32.8125, "active_queue_size": 16384.0, "cl_loss": 5.2703, "doc_norm": 2.0205, "encoder_q-embeddings": 1047.0615, "encoder_q-layer.0": 771.9317, "encoder_q-layer.1": 811.345, "encoder_q-layer.10": 1866.8712, "encoder_q-layer.11": 3664.1289, "encoder_q-layer.2": 952.2792, "encoder_q-layer.3": 964.8569, "encoder_q-layer.4": 1133.6477, "encoder_q-layer.5": 1498.278, "encoder_q-layer.6": 1889.9347, "encoder_q-layer.7": 2000.7842, "encoder_q-layer.8": 1951.9336, "encoder_q-layer.9": 1661.8811, "epoch": 0.02, "inbatch_neg_score": 0.5714, "inbatch_pos_score": 1.0332, "learning_rate": 2.3000000000000003e-05, "loss": 5.2703, "norm_diff": 0.376, "norm_loss": 0.0, "num_token_doc": 66.8031, "num_token_overlap": 11.6869, "num_token_query": 31.3868, "num_token_union": 65.1375, "num_word_context": 202.3521, "num_word_doc": 49.8344, "num_word_query": 23.3154, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2509.4215, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5698, "query_norm": 1.6444, "queue_k_norm": 2.0149, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3868, "sent_len_1": 66.8031, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.96, "stdk": 0.0504, "stdq": 0.0482, "stdqueue_k": 0.0502, "stdqueue_q": 0.0, "step": 4600 }, { "accuracy": 31.543, "active_queue_size": 16384.0, "cl_loss": 5.2251, "doc_norm": 1.9603, "encoder_q-embeddings": 1639.7864, "encoder_q-layer.0": 1340.9985, "encoder_q-layer.1": 1663.853, "encoder_q-layer.10": 2653.9917, "encoder_q-layer.11": 4229.4385, "encoder_q-layer.2": 2134.791, "encoder_q-layer.3": 2089.7737, "encoder_q-layer.4": 2318.5789, "encoder_q-layer.5": 2788.2092, "encoder_q-layer.6": 3033.4761, "encoder_q-layer.7": 3405.3337, "encoder_q-layer.8": 3306.5234, "encoder_q-layer.9": 2525.2471, "epoch": 0.02, "inbatch_neg_score": 0.5608, "inbatch_pos_score": 1.0127, "learning_rate": 2.35e-05, "loss": 5.2251, "norm_diff": 0.3307, "norm_loss": 0.0, "num_token_doc": 66.5444, "num_token_overlap": 11.6457, "num_token_query": 31.3543, "num_token_union": 64.996, "num_word_context": 202.0793, "num_word_doc": 49.6848, "num_word_query": 23.2798, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3806.8823, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5591, "query_norm": 1.6295, "queue_k_norm": 1.9595, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3543, "sent_len_1": 66.5444, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.2675, "stdk": 0.0497, "stdq": 0.0473, "stdqueue_k": 0.0498, "stdqueue_q": 0.0, "step": 4700 }, { "accuracy": 33.3984, "active_queue_size": 16384.0, "cl_loss": 5.178, "doc_norm": 1.9203, "encoder_q-embeddings": 1577.528, "encoder_q-layer.0": 1367.8518, "encoder_q-layer.1": 1428.4614, "encoder_q-layer.10": 2928.4644, "encoder_q-layer.11": 4226.7021, "encoder_q-layer.2": 1481.6465, "encoder_q-layer.3": 1335.7765, "encoder_q-layer.4": 1538.3483, "encoder_q-layer.5": 2259.3115, "encoder_q-layer.6": 2931.0205, "encoder_q-layer.7": 3759.021, "encoder_q-layer.8": 3853.7495, "encoder_q-layer.9": 3031.6648, "epoch": 0.02, "inbatch_neg_score": 0.5605, "inbatch_pos_score": 1.0146, "learning_rate": 2.4e-05, "loss": 5.178, "norm_diff": 0.2829, "norm_loss": 0.0, "num_token_doc": 66.7775, "num_token_overlap": 11.6792, "num_token_query": 31.3345, "num_token_union": 65.0978, "num_word_context": 202.4883, "num_word_doc": 49.8377, "num_word_query": 23.274, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3709.046, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5596, "query_norm": 1.6373, "queue_k_norm": 1.9164, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3345, "sent_len_1": 66.7775, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.7763, "stdk": 0.0497, "stdq": 0.0471, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 4800 }, { "accuracy": 33.5449, "active_queue_size": 16384.0, "cl_loss": 5.1461, "doc_norm": 1.8807, "encoder_q-embeddings": 1011.7842, "encoder_q-layer.0": 742.978, "encoder_q-layer.1": 801.0015, "encoder_q-layer.10": 2058.1694, "encoder_q-layer.11": 3787.73, "encoder_q-layer.2": 852.873, "encoder_q-layer.3": 849.3287, "encoder_q-layer.4": 884.5811, "encoder_q-layer.5": 963.8892, "encoder_q-layer.6": 1151.8647, "encoder_q-layer.7": 1437.5908, "encoder_q-layer.8": 1707.2212, "encoder_q-layer.9": 1481.5818, "epoch": 0.02, "inbatch_neg_score": 0.6454, "inbatch_pos_score": 1.1152, "learning_rate": 2.45e-05, "loss": 5.1461, "norm_diff": 0.2148, "norm_loss": 0.0, "num_token_doc": 67.0675, "num_token_overlap": 11.6745, "num_token_query": 31.3167, "num_token_union": 65.2675, "num_word_context": 202.466, "num_word_doc": 50.0184, "num_word_query": 23.2633, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2329.8652, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6426, "query_norm": 1.6659, "queue_k_norm": 1.8784, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3167, "sent_len_1": 67.0675, "sent_len_max_0": 127.99, "sent_len_max_1": 210.675, "stdk": 0.0493, "stdq": 0.0475, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 4900 }, { "accuracy": 34.0332, "active_queue_size": 16384.0, "cl_loss": 5.0903, "doc_norm": 1.8457, "encoder_q-embeddings": 1145.3085, "encoder_q-layer.0": 900.7869, "encoder_q-layer.1": 935.8586, "encoder_q-layer.10": 1651.3423, "encoder_q-layer.11": 3328.1814, "encoder_q-layer.2": 1029.3593, "encoder_q-layer.3": 1059.5642, "encoder_q-layer.4": 1002.8712, "encoder_q-layer.5": 920.5853, "encoder_q-layer.6": 1018.0141, "encoder_q-layer.7": 1081.2396, "encoder_q-layer.8": 1278.2943, "encoder_q-layer.9": 1175.4905, "epoch": 0.02, "inbatch_neg_score": 0.6289, "inbatch_pos_score": 1.085, "learning_rate": 2.5e-05, "loss": 5.0903, "norm_diff": 0.1837, "norm_loss": 0.0, "num_token_doc": 66.6719, "num_token_overlap": 11.7083, "num_token_query": 31.4139, "num_token_union": 65.067, "num_word_context": 202.0665, "num_word_doc": 49.7742, "num_word_query": 23.3393, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2181.6387, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6279, "query_norm": 1.662, "queue_k_norm": 1.8494, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4139, "sent_len_1": 66.6719, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.6312, "stdk": 0.0488, "stdq": 0.0474, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 5000 }, { "accuracy": 34.4727, "active_queue_size": 16384.0, "cl_loss": 5.0536, "doc_norm": 1.819, "encoder_q-embeddings": 976.6014, "encoder_q-layer.0": 748.1111, "encoder_q-layer.1": 727.3822, "encoder_q-layer.10": 1393.5745, "encoder_q-layer.11": 2927.6414, "encoder_q-layer.2": 776.9757, "encoder_q-layer.3": 778.5284, "encoder_q-layer.4": 815.6364, "encoder_q-layer.5": 845.7886, "encoder_q-layer.6": 900.255, "encoder_q-layer.7": 961.2374, "encoder_q-layer.8": 1090.4202, "encoder_q-layer.9": 1007.1848, "epoch": 0.02, "inbatch_neg_score": 0.6288, "inbatch_pos_score": 1.1006, "learning_rate": 2.5500000000000003e-05, "loss": 5.0536, "norm_diff": 0.1488, "norm_loss": 0.0, "num_token_doc": 66.7119, "num_token_overlap": 11.6551, "num_token_query": 31.3028, "num_token_union": 65.0222, "num_word_context": 202.115, "num_word_doc": 49.7513, "num_word_query": 23.2465, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1839.0663, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.626, "query_norm": 1.6702, "queue_k_norm": 1.8253, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3028, "sent_len_1": 66.7119, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.9363, "stdk": 0.0483, "stdq": 0.0478, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 5100 }, { "accuracy": 33.9844, "active_queue_size": 16384.0, "cl_loss": 5.0084, "doc_norm": 1.8014, "encoder_q-embeddings": 1215.6844, "encoder_q-layer.0": 972.0933, "encoder_q-layer.1": 983.2487, "encoder_q-layer.10": 1570.2852, "encoder_q-layer.11": 3141.8213, "encoder_q-layer.2": 1058.8987, "encoder_q-layer.3": 931.4338, "encoder_q-layer.4": 872.788, "encoder_q-layer.5": 796.7152, "encoder_q-layer.6": 926.8129, "encoder_q-layer.7": 953.6401, "encoder_q-layer.8": 1147.8396, "encoder_q-layer.9": 1160.0519, "epoch": 0.02, "inbatch_neg_score": 0.6113, "inbatch_pos_score": 1.0732, "learning_rate": 2.6000000000000002e-05, "loss": 5.0084, "norm_diff": 0.1438, "norm_loss": 0.0, "num_token_doc": 66.6216, "num_token_overlap": 11.6755, "num_token_query": 31.3705, "num_token_union": 65.0291, "num_word_context": 201.8646, "num_word_doc": 49.6885, "num_word_query": 23.2909, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1992.4567, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6099, "query_norm": 1.6576, "queue_k_norm": 1.8018, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3705, "sent_len_1": 66.6216, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.0563, "stdk": 0.0484, "stdq": 0.0466, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 5200 }, { "accuracy": 34.5215, "active_queue_size": 16384.0, "cl_loss": 4.9856, "doc_norm": 1.7787, "encoder_q-embeddings": 1295.697, "encoder_q-layer.0": 1378.2101, "encoder_q-layer.1": 1655.0717, "encoder_q-layer.10": 1668.9473, "encoder_q-layer.11": 3427.0046, "encoder_q-layer.2": 2257.0361, "encoder_q-layer.3": 2228.9543, "encoder_q-layer.4": 2039.5537, "encoder_q-layer.5": 1295.5533, "encoder_q-layer.6": 984.0105, "encoder_q-layer.7": 923.1531, "encoder_q-layer.8": 1099.6327, "encoder_q-layer.9": 1100.6943, "epoch": 0.02, "inbatch_neg_score": 0.6366, "inbatch_pos_score": 1.1113, "learning_rate": 2.6500000000000004e-05, "loss": 4.9856, "norm_diff": 0.0883, "norm_loss": 0.0, "num_token_doc": 66.8547, "num_token_overlap": 11.6904, "num_token_query": 31.3575, "num_token_union": 65.1287, "num_word_context": 202.3071, "num_word_doc": 49.9016, "num_word_query": 23.2845, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2750.4709, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6333, "query_norm": 1.6904, "queue_k_norm": 1.7818, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3575, "sent_len_1": 66.8547, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.5075, "stdk": 0.048, "stdq": 0.0476, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 5300 }, { "accuracy": 34.668, "active_queue_size": 16384.0, "cl_loss": 4.9563, "doc_norm": 1.7711, "encoder_q-embeddings": 858.461, "encoder_q-layer.0": 649.0377, "encoder_q-layer.1": 723.4353, "encoder_q-layer.10": 1721.925, "encoder_q-layer.11": 3148.927, "encoder_q-layer.2": 836.2954, "encoder_q-layer.3": 857.5436, "encoder_q-layer.4": 1038.8845, "encoder_q-layer.5": 1248.3546, "encoder_q-layer.6": 1499.3594, "encoder_q-layer.7": 1697.1246, "encoder_q-layer.8": 1685.3368, "encoder_q-layer.9": 1367.8298, "epoch": 0.02, "inbatch_neg_score": 0.5971, "inbatch_pos_score": 1.0762, "learning_rate": 2.7000000000000002e-05, "loss": 4.9563, "norm_diff": 0.0859, "norm_loss": 0.0, "num_token_doc": 66.686, "num_token_overlap": 11.6436, "num_token_query": 31.2633, "num_token_union": 65.0263, "num_word_context": 202.0403, "num_word_doc": 49.7716, "num_word_query": 23.2107, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2154.5032, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5972, "query_norm": 1.6852, "queue_k_norm": 1.7664, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2633, "sent_len_1": 66.686, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.435, "stdk": 0.048, "stdq": 0.0475, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 5400 }, { "accuracy": 34.082, "active_queue_size": 16384.0, "cl_loss": 4.9068, "doc_norm": 1.7539, "encoder_q-embeddings": 3753.9209, "encoder_q-layer.0": 2972.656, "encoder_q-layer.1": 3143.5918, "encoder_q-layer.10": 1662.0764, "encoder_q-layer.11": 2812.686, "encoder_q-layer.2": 3070.2759, "encoder_q-layer.3": 2731.8062, "encoder_q-layer.4": 2135.7102, "encoder_q-layer.5": 1652.7076, "encoder_q-layer.6": 1782.2729, "encoder_q-layer.7": 1956.1172, "encoder_q-layer.8": 1791.9818, "encoder_q-layer.9": 1422.1912, "epoch": 0.02, "inbatch_neg_score": 0.5843, "inbatch_pos_score": 1.0625, "learning_rate": 2.7500000000000004e-05, "loss": 4.9068, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.7757, "num_token_overlap": 11.7168, "num_token_query": 31.5231, "num_token_union": 65.2008, "num_word_context": 202.3859, "num_word_doc": 49.8237, "num_word_query": 23.4114, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3826.0324, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.583, "query_norm": 1.6835, "queue_k_norm": 1.7542, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5231, "sent_len_1": 66.7757, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.6225, "stdk": 0.0477, "stdq": 0.0472, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 5500 }, { "accuracy": 35.8398, "active_queue_size": 16384.0, "cl_loss": 4.8828, "doc_norm": 1.7387, "encoder_q-embeddings": 1108.3365, "encoder_q-layer.0": 840.7993, "encoder_q-layer.1": 846.2749, "encoder_q-layer.10": 1367.4611, "encoder_q-layer.11": 2777.0977, "encoder_q-layer.2": 920.1428, "encoder_q-layer.3": 825.295, "encoder_q-layer.4": 752.9602, "encoder_q-layer.5": 716.0859, "encoder_q-layer.6": 836.8988, "encoder_q-layer.7": 898.6588, "encoder_q-layer.8": 964.082, "encoder_q-layer.9": 954.7189, "epoch": 0.02, "inbatch_neg_score": 0.599, "inbatch_pos_score": 1.085, "learning_rate": 2.8000000000000003e-05, "loss": 4.8828, "norm_diff": 0.0239, "norm_loss": 0.0, "num_token_doc": 66.73, "num_token_overlap": 11.7035, "num_token_query": 31.4243, "num_token_union": 65.0799, "num_word_context": 202.1966, "num_word_doc": 49.8163, "num_word_query": 23.3465, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1794.2912, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5967, "query_norm": 1.7173, "queue_k_norm": 1.7442, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4243, "sent_len_1": 66.73, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.8688, "stdk": 0.0474, "stdq": 0.0482, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 5600 }, { "accuracy": 35.0098, "active_queue_size": 16384.0, "cl_loss": 4.8622, "doc_norm": 1.726, "encoder_q-embeddings": 789.3044, "encoder_q-layer.0": 604.6611, "encoder_q-layer.1": 671.7847, "encoder_q-layer.10": 1281.5105, "encoder_q-layer.11": 2621.6292, "encoder_q-layer.2": 769.0263, "encoder_q-layer.3": 733.3525, "encoder_q-layer.4": 708.9321, "encoder_q-layer.5": 689.5634, "encoder_q-layer.6": 748.8452, "encoder_q-layer.7": 879.782, "encoder_q-layer.8": 1025.5521, "encoder_q-layer.9": 925.7203, "epoch": 0.02, "inbatch_neg_score": 0.6002, "inbatch_pos_score": 1.0791, "learning_rate": 2.8499999999999998e-05, "loss": 4.8622, "norm_diff": 0.0294, "norm_loss": 0.0, "num_token_doc": 66.8872, "num_token_overlap": 11.6905, "num_token_query": 31.4059, "num_token_union": 65.1989, "num_word_context": 202.4271, "num_word_doc": 49.9064, "num_word_query": 23.3297, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1627.0354, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5996, "query_norm": 1.6966, "queue_k_norm": 1.7311, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4059, "sent_len_1": 66.8872, "sent_len_max_0": 128.0, "sent_len_max_1": 208.0475, "stdk": 0.0473, "stdq": 0.0473, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 5700 }, { "accuracy": 34.9609, "active_queue_size": 16384.0, "cl_loss": 4.8589, "doc_norm": 1.7173, "encoder_q-embeddings": 1037.5959, "encoder_q-layer.0": 841.0414, "encoder_q-layer.1": 866.548, "encoder_q-layer.10": 1677.4087, "encoder_q-layer.11": 3364.6636, "encoder_q-layer.2": 963.5671, "encoder_q-layer.3": 913.8025, "encoder_q-layer.4": 1007.3892, "encoder_q-layer.5": 1266.3517, "encoder_q-layer.6": 1589.0292, "encoder_q-layer.7": 1947.7532, "encoder_q-layer.8": 1909.3602, "encoder_q-layer.9": 1450.2855, "epoch": 0.03, "inbatch_neg_score": 0.567, "inbatch_pos_score": 1.0439, "learning_rate": 2.9e-05, "loss": 4.8589, "norm_diff": 0.0182, "norm_loss": 0.0, "num_token_doc": 66.7489, "num_token_overlap": 11.6695, "num_token_query": 31.4013, "num_token_union": 65.1138, "num_word_context": 202.3507, "num_word_doc": 49.7765, "num_word_query": 23.3211, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2349.4336, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5654, "query_norm": 1.7007, "queue_k_norm": 1.7192, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4013, "sent_len_1": 66.7489, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.5437, "stdk": 0.0472, "stdq": 0.0468, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 5800 }, { "accuracy": 35.9375, "active_queue_size": 16384.0, "cl_loss": 4.858, "doc_norm": 1.7093, "encoder_q-embeddings": 857.3249, "encoder_q-layer.0": 632.9384, "encoder_q-layer.1": 696.1045, "encoder_q-layer.10": 1858.8297, "encoder_q-layer.11": 3760.8572, "encoder_q-layer.2": 811.9901, "encoder_q-layer.3": 833.4288, "encoder_q-layer.4": 842.9941, "encoder_q-layer.5": 812.3936, "encoder_q-layer.6": 901.17, "encoder_q-layer.7": 972.6239, "encoder_q-layer.8": 1139.7736, "encoder_q-layer.9": 1109.0708, "epoch": 0.03, "inbatch_neg_score": 0.5703, "inbatch_pos_score": 1.0566, "learning_rate": 2.95e-05, "loss": 4.858, "norm_diff": 0.0163, "norm_loss": 0.0, "num_token_doc": 66.6646, "num_token_overlap": 11.65, "num_token_query": 31.2764, "num_token_union": 64.9766, "num_word_context": 202.1212, "num_word_doc": 49.7221, "num_word_query": 23.2207, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2162.6264, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5688, "query_norm": 1.7227, "queue_k_norm": 1.709, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2764, "sent_len_1": 66.6646, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.4563, "stdk": 0.0473, "stdq": 0.0473, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 5900 }, { "accuracy": 34.4727, "active_queue_size": 16384.0, "cl_loss": 4.818, "doc_norm": 1.6991, "encoder_q-embeddings": 713.3121, "encoder_q-layer.0": 509.4849, "encoder_q-layer.1": 550.0421, "encoder_q-layer.10": 1366.6261, "encoder_q-layer.11": 2758.6775, "encoder_q-layer.2": 626.6872, "encoder_q-layer.3": 659.5823, "encoder_q-layer.4": 635.346, "encoder_q-layer.5": 634.7166, "encoder_q-layer.6": 715.7445, "encoder_q-layer.7": 800.6738, "encoder_q-layer.8": 995.5681, "encoder_q-layer.9": 934.1769, "epoch": 0.03, "inbatch_neg_score": 0.5897, "inbatch_pos_score": 1.0547, "learning_rate": 3e-05, "loss": 4.818, "norm_diff": 0.0096, "norm_loss": 0.0, "num_token_doc": 66.775, "num_token_overlap": 11.6911, "num_token_query": 31.3587, "num_token_union": 65.0984, "num_word_context": 201.9251, "num_word_doc": 49.7955, "num_word_query": 23.2743, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1623.946, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5879, "query_norm": 1.6956, "queue_k_norm": 1.6998, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3587, "sent_len_1": 66.775, "sent_len_max_0": 127.9425, "sent_len_max_1": 209.7988, "stdk": 0.0471, "stdq": 0.0461, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 6000 }, { "accuracy": 35.8887, "active_queue_size": 16384.0, "cl_loss": 4.7902, "doc_norm": 1.6885, "encoder_q-embeddings": 2418.021, "encoder_q-layer.0": 2000.7834, "encoder_q-layer.1": 1820.6736, "encoder_q-layer.10": 1605.4647, "encoder_q-layer.11": 3055.2539, "encoder_q-layer.2": 1831.2643, "encoder_q-layer.3": 1516.2241, "encoder_q-layer.4": 1252.7178, "encoder_q-layer.5": 1173.235, "encoder_q-layer.6": 1187.121, "encoder_q-layer.7": 1267.9954, "encoder_q-layer.8": 1331.6224, "encoder_q-layer.9": 1194.1886, "epoch": 0.03, "inbatch_neg_score": 0.5863, "inbatch_pos_score": 1.0527, "learning_rate": 3.05e-05, "loss": 4.7902, "norm_diff": 0.0117, "norm_loss": 0.0, "num_token_doc": 66.7903, "num_token_overlap": 11.675, "num_token_query": 31.4148, "num_token_union": 65.1571, "num_word_context": 202.3976, "num_word_doc": 49.856, "num_word_query": 23.3306, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2698.1903, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.584, "query_norm": 1.685, "queue_k_norm": 1.6912, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4148, "sent_len_1": 66.7903, "sent_len_max_0": 128.0, "sent_len_max_1": 207.8975, "stdk": 0.047, "stdq": 0.0453, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 6100 }, { "accuracy": 36.377, "active_queue_size": 16384.0, "cl_loss": 4.7659, "doc_norm": 1.6872, "encoder_q-embeddings": 2324.2158, "encoder_q-layer.0": 1826.9913, "encoder_q-layer.1": 1823.2664, "encoder_q-layer.10": 1605.29, "encoder_q-layer.11": 2844.0349, "encoder_q-layer.2": 2006.8235, "encoder_q-layer.3": 1721.6158, "encoder_q-layer.4": 1361.9127, "encoder_q-layer.5": 997.2897, "encoder_q-layer.6": 977.2451, "encoder_q-layer.7": 1052.934, "encoder_q-layer.8": 1207.5094, "encoder_q-layer.9": 1104.3573, "epoch": 0.03, "inbatch_neg_score": 0.5871, "inbatch_pos_score": 1.0801, "learning_rate": 3.1e-05, "loss": 4.7659, "norm_diff": 0.0457, "norm_loss": 0.0, "num_token_doc": 66.7643, "num_token_overlap": 11.6926, "num_token_query": 31.4191, "num_token_union": 65.1415, "num_word_context": 202.2036, "num_word_doc": 49.7905, "num_word_query": 23.3346, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2584.768, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5859, "query_norm": 1.7329, "queue_k_norm": 1.6802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4191, "sent_len_1": 66.7643, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.7163, "stdk": 0.0472, "stdq": 0.0476, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 6200 }, { "accuracy": 35.3516, "active_queue_size": 16384.0, "cl_loss": 4.7574, "doc_norm": 1.6666, "encoder_q-embeddings": 2044.4021, "encoder_q-layer.0": 1681.8972, "encoder_q-layer.1": 1661.9188, "encoder_q-layer.10": 1308.1144, "encoder_q-layer.11": 2594.5703, "encoder_q-layer.2": 1870.5438, "encoder_q-layer.3": 1698.8999, "encoder_q-layer.4": 1355.744, "encoder_q-layer.5": 1236.3313, "encoder_q-layer.6": 1383.7056, "encoder_q-layer.7": 1422.2136, "encoder_q-layer.8": 1327.1565, "encoder_q-layer.9": 1094.101, "epoch": 0.03, "inbatch_neg_score": 0.6021, "inbatch_pos_score": 1.085, "learning_rate": 3.15e-05, "loss": 4.7574, "norm_diff": 0.056, "norm_loss": 0.0, "num_token_doc": 66.8426, "num_token_overlap": 11.6631, "num_token_query": 31.3678, "num_token_union": 65.14, "num_word_context": 202.5387, "num_word_doc": 49.8807, "num_word_query": 23.2994, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2474.63, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6006, "query_norm": 1.7226, "queue_k_norm": 1.6723, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3678, "sent_len_1": 66.8426, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.8525, "stdk": 0.0468, "stdq": 0.0477, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 6300 }, { "accuracy": 36.1816, "active_queue_size": 16384.0, "cl_loss": 4.7558, "doc_norm": 1.6589, "encoder_q-embeddings": 905.7628, "encoder_q-layer.0": 762.8909, "encoder_q-layer.1": 671.2789, "encoder_q-layer.10": 1961.38, "encoder_q-layer.11": 3797.9893, "encoder_q-layer.2": 755.2224, "encoder_q-layer.3": 789.1887, "encoder_q-layer.4": 776.4489, "encoder_q-layer.5": 774.1599, "encoder_q-layer.6": 826.0903, "encoder_q-layer.7": 870.3393, "encoder_q-layer.8": 1115.566, "encoder_q-layer.9": 1097.8977, "epoch": 0.03, "inbatch_neg_score": 0.5869, "inbatch_pos_score": 1.0762, "learning_rate": 3.2000000000000005e-05, "loss": 4.7558, "norm_diff": 0.0346, "norm_loss": 0.0, "num_token_doc": 66.7509, "num_token_overlap": 11.6429, "num_token_query": 31.3044, "num_token_union": 65.1115, "num_word_context": 202.3521, "num_word_doc": 49.8031, "num_word_query": 23.2471, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2083.9571, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5845, "query_norm": 1.6936, "queue_k_norm": 1.6621, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3044, "sent_len_1": 66.7509, "sent_len_max_0": 127.97, "sent_len_max_1": 209.5913, "stdk": 0.0468, "stdq": 0.0473, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 6400 }, { "accuracy": 40.4785, "active_queue_size": 16384.0, "cl_loss": 4.725, "doc_norm": 1.6492, "encoder_q-embeddings": 1624.875, "encoder_q-layer.0": 1449.783, "encoder_q-layer.1": 1473.5768, "encoder_q-layer.10": 1833.4164, "encoder_q-layer.11": 3722.7515, "encoder_q-layer.2": 1580.0638, "encoder_q-layer.3": 1417.3755, "encoder_q-layer.4": 1313.9141, "encoder_q-layer.5": 1064.1383, "encoder_q-layer.6": 1024.3508, "encoder_q-layer.7": 1131.5055, "encoder_q-layer.8": 1320.0945, "encoder_q-layer.9": 1134.3644, "epoch": 0.03, "inbatch_neg_score": 0.588, "inbatch_pos_score": 1.0986, "learning_rate": 3.2500000000000004e-05, "loss": 4.725, "norm_diff": 0.0331, "norm_loss": 0.0, "num_token_doc": 66.6191, "num_token_overlap": 11.6637, "num_token_query": 31.3637, "num_token_union": 65.0108, "num_word_context": 201.9335, "num_word_doc": 49.7057, "num_word_query": 23.2979, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2463.0971, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5869, "query_norm": 1.6823, "queue_k_norm": 1.653, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3637, "sent_len_1": 66.6191, "sent_len_max_0": 127.99, "sent_len_max_1": 208.1687, "stdk": 0.0467, "stdq": 0.0473, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 6500 }, { "accuracy": 38.0371, "active_queue_size": 16384.0, "cl_loss": 4.7193, "doc_norm": 1.6482, "encoder_q-embeddings": 1782.4866, "encoder_q-layer.0": 1338.1643, "encoder_q-layer.1": 1323.8236, "encoder_q-layer.10": 4719.562, "encoder_q-layer.11": 8951.0625, "encoder_q-layer.2": 1440.5293, "encoder_q-layer.3": 1461.3789, "encoder_q-layer.4": 1409.536, "encoder_q-layer.5": 1381.5906, "encoder_q-layer.6": 1684.4177, "encoder_q-layer.7": 1919.7256, "encoder_q-layer.8": 2215.1001, "encoder_q-layer.9": 2195.8416, "epoch": 0.03, "inbatch_neg_score": 0.5776, "inbatch_pos_score": 1.0635, "learning_rate": 3.3e-05, "loss": 4.7193, "norm_diff": 0.0279, "norm_loss": 0.0, "num_token_doc": 66.8461, "num_token_overlap": 11.6728, "num_token_query": 31.3206, "num_token_union": 65.1182, "num_word_context": 202.3055, "num_word_doc": 49.9039, "num_word_query": 23.2637, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4733.8502, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5767, "query_norm": 1.6761, "queue_k_norm": 1.6438, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3206, "sent_len_1": 66.8461, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.6, "stdk": 0.0469, "stdq": 0.0468, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 6600 }, { "accuracy": 36.9141, "active_queue_size": 16384.0, "cl_loss": 4.6867, "doc_norm": 1.6373, "encoder_q-embeddings": 1493.6355, "encoder_q-layer.0": 1041.8296, "encoder_q-layer.1": 1135.564, "encoder_q-layer.10": 3003.1848, "encoder_q-layer.11": 5955.3159, "encoder_q-layer.2": 1279.5825, "encoder_q-layer.3": 1287.9376, "encoder_q-layer.4": 1381.705, "encoder_q-layer.5": 1508.1066, "encoder_q-layer.6": 1663.6163, "encoder_q-layer.7": 1848.4573, "encoder_q-layer.8": 2143.7141, "encoder_q-layer.9": 2011.0485, "epoch": 0.03, "inbatch_neg_score": 0.5532, "inbatch_pos_score": 1.0381, "learning_rate": 3.35e-05, "loss": 4.6867, "norm_diff": 0.0299, "norm_loss": 0.0, "num_token_doc": 66.6665, "num_token_overlap": 11.6886, "num_token_query": 31.3755, "num_token_union": 65.0407, "num_word_context": 202.3413, "num_word_doc": 49.7439, "num_word_query": 23.3033, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3413.7468, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5508, "query_norm": 1.6672, "queue_k_norm": 1.6349, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3755, "sent_len_1": 66.6665, "sent_len_max_0": 128.0, "sent_len_max_1": 209.9588, "stdk": 0.0468, "stdq": 0.0469, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 6700 }, { "accuracy": 37.4023, "active_queue_size": 16384.0, "cl_loss": 4.6649, "doc_norm": 1.6273, "encoder_q-embeddings": 1908.5497, "encoder_q-layer.0": 1409.9355, "encoder_q-layer.1": 1479.7839, "encoder_q-layer.10": 2314.3228, "encoder_q-layer.11": 4885.4512, "encoder_q-layer.2": 1441.9767, "encoder_q-layer.3": 1432.9795, "encoder_q-layer.4": 1352.5082, "encoder_q-layer.5": 1208.3575, "encoder_q-layer.6": 1308.3873, "encoder_q-layer.7": 1389.8529, "encoder_q-layer.8": 1628.5978, "encoder_q-layer.9": 1557.0275, "epoch": 0.03, "inbatch_neg_score": 0.5194, "inbatch_pos_score": 1.0068, "learning_rate": 3.4000000000000007e-05, "loss": 4.6649, "norm_diff": 0.025, "norm_loss": 0.0, "num_token_doc": 66.7975, "num_token_overlap": 11.7121, "num_token_query": 31.3973, "num_token_union": 65.1656, "num_word_context": 202.5439, "num_word_doc": 49.8738, "num_word_query": 23.3324, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3091.8292, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5195, "query_norm": 1.6523, "queue_k_norm": 1.6244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3973, "sent_len_1": 66.7975, "sent_len_max_0": 127.995, "sent_len_max_1": 206.6175, "stdk": 0.0468, "stdq": 0.0465, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 6800 }, { "accuracy": 37.6953, "active_queue_size": 16384.0, "cl_loss": 4.6611, "doc_norm": 1.6179, "encoder_q-embeddings": 1958.5785, "encoder_q-layer.0": 1465.9098, "encoder_q-layer.1": 1612.7125, "encoder_q-layer.10": 2896.8557, "encoder_q-layer.11": 5748.6025, "encoder_q-layer.2": 1845.8746, "encoder_q-layer.3": 1875.5427, "encoder_q-layer.4": 1842.0, "encoder_q-layer.5": 1515.2258, "encoder_q-layer.6": 1561.5323, "encoder_q-layer.7": 1576.2275, "encoder_q-layer.8": 1918.9257, "encoder_q-layer.9": 1844.4048, "epoch": 0.03, "inbatch_neg_score": 0.5302, "inbatch_pos_score": 1.0273, "learning_rate": 3.45e-05, "loss": 4.6611, "norm_diff": 0.0279, "norm_loss": 0.0, "num_token_doc": 66.892, "num_token_overlap": 11.7361, "num_token_query": 31.4659, "num_token_union": 65.2287, "num_word_context": 202.5509, "num_word_doc": 49.9461, "num_word_query": 23.379, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3558.8225, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5293, "query_norm": 1.6459, "queue_k_norm": 1.6199, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4659, "sent_len_1": 66.892, "sent_len_max_0": 127.9963, "sent_len_max_1": 206.4638, "stdk": 0.0467, "stdq": 0.0463, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 6900 }, { "accuracy": 36.6699, "active_queue_size": 16384.0, "cl_loss": 4.6591, "doc_norm": 1.6081, "encoder_q-embeddings": 1330.3507, "encoder_q-layer.0": 967.2167, "encoder_q-layer.1": 1018.5017, "encoder_q-layer.10": 4293.7485, "encoder_q-layer.11": 7473.6372, "encoder_q-layer.2": 1145.5767, "encoder_q-layer.3": 1194.116, "encoder_q-layer.4": 1306.5785, "encoder_q-layer.5": 1473.8271, "encoder_q-layer.6": 1731.4756, "encoder_q-layer.7": 2085.5171, "encoder_q-layer.8": 2263.8914, "encoder_q-layer.9": 2116.3025, "epoch": 0.03, "inbatch_neg_score": 0.538, "inbatch_pos_score": 1.0322, "learning_rate": 3.5e-05, "loss": 4.6591, "norm_diff": 0.0317, "norm_loss": 0.0, "num_token_doc": 66.7505, "num_token_overlap": 11.7068, "num_token_query": 31.4088, "num_token_union": 65.1177, "num_word_context": 202.4058, "num_word_doc": 49.8224, "num_word_query": 23.3422, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3939.0523, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5371, "query_norm": 1.6398, "queue_k_norm": 1.6116, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4088, "sent_len_1": 66.7505, "sent_len_max_0": 127.995, "sent_len_max_1": 206.5137, "stdk": 0.0467, "stdq": 0.0464, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 7000 }, { "accuracy": 37.5977, "active_queue_size": 16384.0, "cl_loss": 4.6521, "doc_norm": 1.6025, "encoder_q-embeddings": 1800.567, "encoder_q-layer.0": 1398.3317, "encoder_q-layer.1": 1441.1466, "encoder_q-layer.10": 2360.009, "encoder_q-layer.11": 5076.5776, "encoder_q-layer.2": 1600.8944, "encoder_q-layer.3": 1762.9258, "encoder_q-layer.4": 1616.3271, "encoder_q-layer.5": 1531.0049, "encoder_q-layer.6": 1447.2349, "encoder_q-layer.7": 1441.0452, "encoder_q-layer.8": 1752.5996, "encoder_q-layer.9": 1669.1765, "epoch": 0.03, "inbatch_neg_score": 0.5049, "inbatch_pos_score": 1.0, "learning_rate": 3.55e-05, "loss": 4.6521, "norm_diff": 0.022, "norm_loss": 0.0, "num_token_doc": 66.7517, "num_token_overlap": 11.6408, "num_token_query": 31.3139, "num_token_union": 65.0906, "num_word_context": 202.0796, "num_word_doc": 49.8071, "num_word_query": 23.2475, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3275.9866, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5049, "query_norm": 1.6206, "queue_k_norm": 1.6045, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3139, "sent_len_1": 66.7517, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.945, "stdk": 0.0467, "stdq": 0.0468, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 7100 }, { "accuracy": 38.1836, "active_queue_size": 16384.0, "cl_loss": 4.6207, "doc_norm": 1.5996, "encoder_q-embeddings": 2668.7764, "encoder_q-layer.0": 1887.2638, "encoder_q-layer.1": 1664.809, "encoder_q-layer.10": 2486.8538, "encoder_q-layer.11": 5357.8564, "encoder_q-layer.2": 1673.4086, "encoder_q-layer.3": 1443.1125, "encoder_q-layer.4": 1375.5262, "encoder_q-layer.5": 1344.3805, "encoder_q-layer.6": 1473.7943, "encoder_q-layer.7": 1526.5774, "encoder_q-layer.8": 1853.2085, "encoder_q-layer.9": 1703.0254, "epoch": 0.03, "inbatch_neg_score": 0.5017, "inbatch_pos_score": 0.9971, "learning_rate": 3.6e-05, "loss": 4.6207, "norm_diff": 0.0154, "norm_loss": 0.0, "num_token_doc": 66.7118, "num_token_overlap": 11.7022, "num_token_query": 31.3725, "num_token_union": 65.0525, "num_word_context": 202.2078, "num_word_doc": 49.7833, "num_word_query": 23.2969, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3452.0911, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5005, "query_norm": 1.6016, "queue_k_norm": 1.5961, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3725, "sent_len_1": 66.7118, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.18, "stdk": 0.0469, "stdq": 0.0466, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 7200 }, { "accuracy": 38.1836, "active_queue_size": 16384.0, "cl_loss": 4.6201, "doc_norm": 1.5887, "encoder_q-embeddings": 1585.2111, "encoder_q-layer.0": 1112.6096, "encoder_q-layer.1": 1203.8118, "encoder_q-layer.10": 2217.6179, "encoder_q-layer.11": 4591.9004, "encoder_q-layer.2": 1405.0786, "encoder_q-layer.3": 1507.0388, "encoder_q-layer.4": 1420.8135, "encoder_q-layer.5": 1304.6361, "encoder_q-layer.6": 1411.5903, "encoder_q-layer.7": 1447.5723, "encoder_q-layer.8": 1604.4457, "encoder_q-layer.9": 1477.0634, "epoch": 0.03, "inbatch_neg_score": 0.5022, "inbatch_pos_score": 0.9946, "learning_rate": 3.65e-05, "loss": 4.6201, "norm_diff": 0.0156, "norm_loss": 0.0, "num_token_doc": 66.775, "num_token_overlap": 11.6641, "num_token_query": 31.3319, "num_token_union": 65.111, "num_word_context": 202.2556, "num_word_doc": 49.8071, "num_word_query": 23.2642, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2907.9235, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5005, "query_norm": 1.5881, "queue_k_norm": 1.5902, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3319, "sent_len_1": 66.775, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.1875, "stdk": 0.0467, "stdq": 0.0461, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7300 }, { "accuracy": 40.4785, "active_queue_size": 16384.0, "cl_loss": 4.5835, "doc_norm": 1.5873, "encoder_q-embeddings": 1380.5935, "encoder_q-layer.0": 939.8845, "encoder_q-layer.1": 959.6462, "encoder_q-layer.10": 2594.376, "encoder_q-layer.11": 5067.2715, "encoder_q-layer.2": 1053.7295, "encoder_q-layer.3": 1121.6798, "encoder_q-layer.4": 1148.4, "encoder_q-layer.5": 1114.5968, "encoder_q-layer.6": 1254.0287, "encoder_q-layer.7": 1394.1182, "encoder_q-layer.8": 1749.5836, "encoder_q-layer.9": 1569.0665, "epoch": 0.03, "inbatch_neg_score": 0.4968, "inbatch_pos_score": 1.001, "learning_rate": 3.7e-05, "loss": 4.5835, "norm_diff": 0.02, "norm_loss": 0.0, "num_token_doc": 66.8816, "num_token_overlap": 11.7576, "num_token_query": 31.5112, "num_token_union": 65.1814, "num_word_context": 202.4358, "num_word_doc": 49.92, "num_word_query": 23.4063, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2927.2863, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4968, "query_norm": 1.6061, "queue_k_norm": 1.5829, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5112, "sent_len_1": 66.8816, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.755, "stdk": 0.047, "stdq": 0.0468, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7400 }, { "accuracy": 38.5742, "active_queue_size": 16384.0, "cl_loss": 4.5596, "doc_norm": 1.5786, "encoder_q-embeddings": 1458.5645, "encoder_q-layer.0": 1151.4998, "encoder_q-layer.1": 1311.8805, "encoder_q-layer.10": 1964.7145, "encoder_q-layer.11": 4170.9272, "encoder_q-layer.2": 1589.549, "encoder_q-layer.3": 1560.6637, "encoder_q-layer.4": 1492.4467, "encoder_q-layer.5": 1262.174, "encoder_q-layer.6": 1339.9019, "encoder_q-layer.7": 1347.8948, "encoder_q-layer.8": 1467.6338, "encoder_q-layer.9": 1428.3328, "epoch": 0.03, "inbatch_neg_score": 0.5026, "inbatch_pos_score": 1.0049, "learning_rate": 3.7500000000000003e-05, "loss": 4.5596, "norm_diff": 0.016, "norm_loss": 0.0, "num_token_doc": 66.82, "num_token_overlap": 11.7097, "num_token_query": 31.4845, "num_token_union": 65.2013, "num_word_context": 202.5209, "num_word_doc": 49.8786, "num_word_query": 23.3971, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2751.3188, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5015, "query_norm": 1.5929, "queue_k_norm": 1.577, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4845, "sent_len_1": 66.82, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.6238, "stdk": 0.0469, "stdq": 0.0462, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7500 }, { "accuracy": 40.7715, "active_queue_size": 16384.0, "cl_loss": 4.5798, "doc_norm": 1.5672, "encoder_q-embeddings": 7337.938, "encoder_q-layer.0": 5528.5356, "encoder_q-layer.1": 5127.8647, "encoder_q-layer.10": 2080.8481, "encoder_q-layer.11": 4615.0156, "encoder_q-layer.2": 4913.2773, "encoder_q-layer.3": 4767.3008, "encoder_q-layer.4": 3411.0173, "encoder_q-layer.5": 2095.9045, "encoder_q-layer.6": 1577.9635, "encoder_q-layer.7": 1291.2581, "encoder_q-layer.8": 1505.2477, "encoder_q-layer.9": 1490.6167, "epoch": 0.03, "inbatch_neg_score": 0.498, "inbatch_pos_score": 1.0, "learning_rate": 3.8e-05, "loss": 4.5798, "norm_diff": 0.0269, "norm_loss": 0.0, "num_token_doc": 66.6422, "num_token_overlap": 11.6645, "num_token_query": 31.417, "num_token_union": 65.0661, "num_word_context": 202.2037, "num_word_doc": 49.7335, "num_word_query": 23.3415, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6408.4488, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4978, "query_norm": 1.5939, "queue_k_norm": 1.5712, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.417, "sent_len_1": 66.6422, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.8925, "stdk": 0.0468, "stdq": 0.0465, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7600 }, { "accuracy": 38.3301, "active_queue_size": 16384.0, "cl_loss": 4.5514, "doc_norm": 1.567, "encoder_q-embeddings": 3213.2654, "encoder_q-layer.0": 2599.2456, "encoder_q-layer.1": 2627.8784, "encoder_q-layer.10": 2227.4895, "encoder_q-layer.11": 5012.8599, "encoder_q-layer.2": 2783.002, "encoder_q-layer.3": 2608.021, "encoder_q-layer.4": 2321.6353, "encoder_q-layer.5": 1931.5294, "encoder_q-layer.6": 1662.1508, "encoder_q-layer.7": 1418.2886, "encoder_q-layer.8": 1603.8228, "encoder_q-layer.9": 1475.394, "epoch": 0.03, "inbatch_neg_score": 0.517, "inbatch_pos_score": 1.0059, "learning_rate": 3.85e-05, "loss": 4.5514, "norm_diff": 0.0109, "norm_loss": 0.0, "num_token_doc": 66.5908, "num_token_overlap": 11.6773, "num_token_query": 31.3745, "num_token_union": 65.014, "num_word_context": 201.9414, "num_word_doc": 49.7005, "num_word_query": 23.3105, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3955.8251, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5151, "query_norm": 1.5672, "queue_k_norm": 1.5681, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3745, "sent_len_1": 66.5908, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.2012, "stdk": 0.0469, "stdq": 0.0458, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 7700 }, { "accuracy": 38.6719, "active_queue_size": 16384.0, "cl_loss": 4.5475, "doc_norm": 1.5594, "encoder_q-embeddings": 1792.053, "encoder_q-layer.0": 1269.3037, "encoder_q-layer.1": 1286.4934, "encoder_q-layer.10": 2306.8843, "encoder_q-layer.11": 4700.5571, "encoder_q-layer.2": 1449.6154, "encoder_q-layer.3": 1549.5007, "encoder_q-layer.4": 1389.8885, "encoder_q-layer.5": 1297.5524, "encoder_q-layer.6": 1461.0413, "encoder_q-layer.7": 1671.2385, "encoder_q-layer.8": 1928.0879, "encoder_q-layer.9": 1652.0486, "epoch": 0.03, "inbatch_neg_score": 0.5035, "inbatch_pos_score": 1.0049, "learning_rate": 3.9000000000000006e-05, "loss": 4.5475, "norm_diff": 0.0142, "norm_loss": 0.0, "num_token_doc": 66.6709, "num_token_overlap": 11.654, "num_token_query": 31.3199, "num_token_union": 65.0372, "num_word_context": 202.3047, "num_word_doc": 49.7285, "num_word_query": 23.2524, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3009.0437, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5015, "query_norm": 1.5668, "queue_k_norm": 1.5598, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3199, "sent_len_1": 66.6709, "sent_len_max_0": 127.9575, "sent_len_max_1": 209.1287, "stdk": 0.0468, "stdq": 0.0462, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7800 }, { "accuracy": 38.1836, "active_queue_size": 16384.0, "cl_loss": 4.5483, "doc_norm": 1.554, "encoder_q-embeddings": 1705.0282, "encoder_q-layer.0": 1185.1824, "encoder_q-layer.1": 1256.3975, "encoder_q-layer.10": 2998.8228, "encoder_q-layer.11": 7053.9097, "encoder_q-layer.2": 1424.6692, "encoder_q-layer.3": 1452.95, "encoder_q-layer.4": 1415.5172, "encoder_q-layer.5": 1276.9049, "encoder_q-layer.6": 1347.5913, "encoder_q-layer.7": 1331.0059, "encoder_q-layer.8": 1616.087, "encoder_q-layer.9": 1603.0947, "epoch": 0.03, "inbatch_neg_score": 0.4777, "inbatch_pos_score": 0.9849, "learning_rate": 3.9500000000000005e-05, "loss": 4.5483, "norm_diff": 0.0101, "norm_loss": 0.0, "num_token_doc": 66.7578, "num_token_overlap": 11.6328, "num_token_query": 31.2855, "num_token_union": 65.1094, "num_word_context": 202.2792, "num_word_doc": 49.8241, "num_word_query": 23.2242, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3838.7788, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4775, "query_norm": 1.5542, "queue_k_norm": 1.5566, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2855, "sent_len_1": 66.7578, "sent_len_max_0": 127.985, "sent_len_max_1": 210.3738, "stdk": 0.0468, "stdq": 0.0456, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 7900 }, { "accuracy": 40.4297, "active_queue_size": 16384.0, "cl_loss": 4.4952, "doc_norm": 1.5545, "encoder_q-embeddings": 1260.8591, "encoder_q-layer.0": 874.1161, "encoder_q-layer.1": 940.633, "encoder_q-layer.10": 2030.6091, "encoder_q-layer.11": 4560.4653, "encoder_q-layer.2": 1058.3842, "encoder_q-layer.3": 1116.483, "encoder_q-layer.4": 1134.3309, "encoder_q-layer.5": 1140.1519, "encoder_q-layer.6": 1254.3452, "encoder_q-layer.7": 1324.4967, "encoder_q-layer.8": 1567.6489, "encoder_q-layer.9": 1424.6936, "epoch": 0.03, "inbatch_neg_score": 0.4903, "inbatch_pos_score": 0.9878, "learning_rate": 4e-05, "loss": 4.4952, "norm_diff": 0.0143, "norm_loss": 0.0, "num_token_doc": 66.8725, "num_token_overlap": 11.6851, "num_token_query": 31.3843, "num_token_union": 65.1763, "num_word_context": 202.4621, "num_word_doc": 49.899, "num_word_query": 23.3051, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2752.5989, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4893, "query_norm": 1.552, "queue_k_norm": 1.5521, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3843, "sent_len_1": 66.8725, "sent_len_max_0": 127.975, "sent_len_max_1": 209.7713, "stdk": 0.0469, "stdq": 0.0458, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 8000 }, { "accuracy": 39.7949, "active_queue_size": 16384.0, "cl_loss": 4.5043, "doc_norm": 1.5516, "encoder_q-embeddings": 1753.8336, "encoder_q-layer.0": 1365.1223, "encoder_q-layer.1": 1460.306, "encoder_q-layer.10": 1876.4689, "encoder_q-layer.11": 4274.5293, "encoder_q-layer.2": 1510.2603, "encoder_q-layer.3": 1424.7938, "encoder_q-layer.4": 1307.1434, "encoder_q-layer.5": 1343.7214, "encoder_q-layer.6": 1655.7561, "encoder_q-layer.7": 1826.7002, "encoder_q-layer.8": 1862.7339, "encoder_q-layer.9": 1578.5868, "epoch": 0.04, "inbatch_neg_score": 0.4886, "inbatch_pos_score": 0.9946, "learning_rate": 4.05e-05, "loss": 4.5043, "norm_diff": 0.0133, "norm_loss": 0.0, "num_token_doc": 66.8447, "num_token_overlap": 11.6727, "num_token_query": 31.3713, "num_token_union": 65.1791, "num_word_context": 202.5076, "num_word_doc": 49.8615, "num_word_query": 23.2881, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2971.8016, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4888, "query_norm": 1.5473, "queue_k_norm": 1.5502, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3713, "sent_len_1": 66.8447, "sent_len_max_0": 128.0, "sent_len_max_1": 208.3925, "stdk": 0.047, "stdq": 0.0453, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 8100 }, { "accuracy": 41.3574, "active_queue_size": 16384.0, "cl_loss": 4.4875, "doc_norm": 1.5449, "encoder_q-embeddings": 1499.7028, "encoder_q-layer.0": 1123.3239, "encoder_q-layer.1": 1262.2075, "encoder_q-layer.10": 2340.197, "encoder_q-layer.11": 4893.5723, "encoder_q-layer.2": 1437.9041, "encoder_q-layer.3": 1536.6056, "encoder_q-layer.4": 1474.4785, "encoder_q-layer.5": 1286.0021, "encoder_q-layer.6": 1266.0112, "encoder_q-layer.7": 1281.0522, "encoder_q-layer.8": 1594.1111, "encoder_q-layer.9": 1611.3431, "epoch": 0.04, "inbatch_neg_score": 0.4921, "inbatch_pos_score": 0.9941, "learning_rate": 4.1e-05, "loss": 4.4875, "norm_diff": 0.0162, "norm_loss": 0.0, "num_token_doc": 66.7578, "num_token_overlap": 11.6981, "num_token_query": 31.3345, "num_token_union": 65.0731, "num_word_context": 202.146, "num_word_doc": 49.8304, "num_word_query": 23.2793, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2948.7767, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4917, "query_norm": 1.5312, "queue_k_norm": 1.5451, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3345, "sent_len_1": 66.7578, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.225, "stdk": 0.0469, "stdq": 0.0451, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 8200 }, { "accuracy": 40.2344, "active_queue_size": 16384.0, "cl_loss": 4.4828, "doc_norm": 1.5444, "encoder_q-embeddings": 1618.6686, "encoder_q-layer.0": 1236.5934, "encoder_q-layer.1": 1348.1326, "encoder_q-layer.10": 2124.6218, "encoder_q-layer.11": 4920.519, "encoder_q-layer.2": 1576.5073, "encoder_q-layer.3": 1546.8347, "encoder_q-layer.4": 1485.8226, "encoder_q-layer.5": 1354.0626, "encoder_q-layer.6": 1454.7577, "encoder_q-layer.7": 1305.496, "encoder_q-layer.8": 1524.2581, "encoder_q-layer.9": 1449.1658, "epoch": 0.04, "inbatch_neg_score": 0.5022, "inbatch_pos_score": 0.9951, "learning_rate": 4.15e-05, "loss": 4.4828, "norm_diff": 0.0445, "norm_loss": 0.0, "num_token_doc": 66.7647, "num_token_overlap": 11.6397, "num_token_query": 31.2647, "num_token_union": 65.0816, "num_word_context": 202.028, "num_word_doc": 49.806, "num_word_query": 23.196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3058.6303, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5015, "query_norm": 1.4999, "queue_k_norm": 1.5441, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2647, "sent_len_1": 66.7647, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.6887, "stdk": 0.047, "stdq": 0.0439, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 8300 }, { "accuracy": 40.0391, "active_queue_size": 16384.0, "cl_loss": 4.4554, "doc_norm": 1.5407, "encoder_q-embeddings": 1971.0461, "encoder_q-layer.0": 1576.1191, "encoder_q-layer.1": 1581.4746, "encoder_q-layer.10": 1899.7335, "encoder_q-layer.11": 4228.8853, "encoder_q-layer.2": 1691.1696, "encoder_q-layer.3": 1645.2513, "encoder_q-layer.4": 1631.136, "encoder_q-layer.5": 1236.9579, "encoder_q-layer.6": 1177.6604, "encoder_q-layer.7": 1109.9238, "encoder_q-layer.8": 1281.397, "encoder_q-layer.9": 1237.8855, "epoch": 0.04, "inbatch_neg_score": 0.489, "inbatch_pos_score": 1.001, "learning_rate": 4.2e-05, "loss": 4.4554, "norm_diff": 0.0154, "norm_loss": 0.0, "num_token_doc": 66.8183, "num_token_overlap": 11.7036, "num_token_query": 31.512, "num_token_union": 65.205, "num_word_context": 202.3003, "num_word_doc": 49.8719, "num_word_query": 23.422, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2921.5127, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4878, "query_norm": 1.5254, "queue_k_norm": 1.5389, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.512, "sent_len_1": 66.8183, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.2725, "stdk": 0.047, "stdq": 0.0452, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 8400 }, { "accuracy": 41.0156, "active_queue_size": 16384.0, "cl_loss": 4.4605, "doc_norm": 1.541, "encoder_q-embeddings": 1357.9706, "encoder_q-layer.0": 940.7751, "encoder_q-layer.1": 989.7165, "encoder_q-layer.10": 2323.2898, "encoder_q-layer.11": 5145.7593, "encoder_q-layer.2": 1095.636, "encoder_q-layer.3": 1126.5504, "encoder_q-layer.4": 1125.7659, "encoder_q-layer.5": 1064.0261, "encoder_q-layer.6": 1154.5547, "encoder_q-layer.7": 1233.2485, "encoder_q-layer.8": 1497.8325, "encoder_q-layer.9": 1473.7861, "epoch": 0.04, "inbatch_neg_score": 0.4885, "inbatch_pos_score": 0.9937, "learning_rate": 4.25e-05, "loss": 4.4605, "norm_diff": 0.056, "norm_loss": 0.0, "num_token_doc": 66.6483, "num_token_overlap": 11.6827, "num_token_query": 31.4913, "num_token_union": 65.1232, "num_word_context": 202.4118, "num_word_doc": 49.7389, "num_word_query": 23.4012, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2820.3325, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.488, "query_norm": 1.4849, "queue_k_norm": 1.537, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4913, "sent_len_1": 66.6483, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.0613, "stdk": 0.0472, "stdq": 0.0436, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 8500 }, { "accuracy": 42.8711, "active_queue_size": 16384.0, "cl_loss": 4.4372, "doc_norm": 1.5329, "encoder_q-embeddings": 2733.5215, "encoder_q-layer.0": 1891.1476, "encoder_q-layer.1": 2078.5288, "encoder_q-layer.10": 4599.2686, "encoder_q-layer.11": 9926.9883, "encoder_q-layer.2": 2332.1348, "encoder_q-layer.3": 2404.564, "encoder_q-layer.4": 2431.8083, "encoder_q-layer.5": 2366.8054, "encoder_q-layer.6": 2636.6516, "encoder_q-layer.7": 2777.0461, "encoder_q-layer.8": 3164.2559, "encoder_q-layer.9": 2920.292, "epoch": 0.04, "inbatch_neg_score": 0.4903, "inbatch_pos_score": 1.0156, "learning_rate": 4.3e-05, "loss": 4.4372, "norm_diff": 0.0206, "norm_loss": 0.0, "num_token_doc": 66.707, "num_token_overlap": 11.6659, "num_token_query": 31.3531, "num_token_union": 65.0939, "num_word_context": 202.3943, "num_word_doc": 49.789, "num_word_query": 23.2919, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5745.9306, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4907, "query_norm": 1.5155, "queue_k_norm": 1.5333, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3531, "sent_len_1": 66.707, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.55, "stdk": 0.047, "stdq": 0.0448, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 8600 }, { "accuracy": 41.1133, "active_queue_size": 16384.0, "cl_loss": 4.4375, "doc_norm": 1.5321, "encoder_q-embeddings": 2623.9443, "encoder_q-layer.0": 1704.6067, "encoder_q-layer.1": 1832.5895, "encoder_q-layer.10": 5350.6123, "encoder_q-layer.11": 13530.6191, "encoder_q-layer.2": 2073.3933, "encoder_q-layer.3": 2169.7795, "encoder_q-layer.4": 2174.7898, "encoder_q-layer.5": 2157.0969, "encoder_q-layer.6": 2383.6187, "encoder_q-layer.7": 2563.9773, "encoder_q-layer.8": 3037.6487, "encoder_q-layer.9": 2905.0693, "epoch": 0.04, "inbatch_neg_score": 0.4906, "inbatch_pos_score": 1.0166, "learning_rate": 4.35e-05, "loss": 4.4375, "norm_diff": 0.0073, "norm_loss": 0.0, "num_token_doc": 66.5739, "num_token_overlap": 11.6943, "num_token_query": 31.4978, "num_token_union": 65.0601, "num_word_context": 202.0709, "num_word_doc": 49.6638, "num_word_query": 23.4004, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6992.1992, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.491, "query_norm": 1.5339, "queue_k_norm": 1.5299, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4978, "sent_len_1": 66.5739, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.055, "stdk": 0.0471, "stdq": 0.0454, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 8700 }, { "accuracy": 40.1855, "active_queue_size": 16384.0, "cl_loss": 4.4294, "doc_norm": 1.5271, "encoder_q-embeddings": 2797.8318, "encoder_q-layer.0": 1968.0609, "encoder_q-layer.1": 2026.9514, "encoder_q-layer.10": 4164.0669, "encoder_q-layer.11": 9570.9863, "encoder_q-layer.2": 2275.0801, "encoder_q-layer.3": 2241.5166, "encoder_q-layer.4": 2233.9514, "encoder_q-layer.5": 2181.3477, "encoder_q-layer.6": 2192.1797, "encoder_q-layer.7": 2163.8923, "encoder_q-layer.8": 2826.6597, "encoder_q-layer.9": 2664.3936, "epoch": 0.04, "inbatch_neg_score": 0.4872, "inbatch_pos_score": 0.9844, "learning_rate": 4.4000000000000006e-05, "loss": 4.4294, "norm_diff": 0.0239, "norm_loss": 0.0, "num_token_doc": 66.524, "num_token_overlap": 11.6785, "num_token_query": 31.451, "num_token_union": 65.0264, "num_word_context": 202.0189, "num_word_doc": 49.6316, "num_word_query": 23.365, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5427.9917, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4868, "query_norm": 1.505, "queue_k_norm": 1.529, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.451, "sent_len_1": 66.524, "sent_len_max_0": 127.99, "sent_len_max_1": 208.2125, "stdk": 0.047, "stdq": 0.0445, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 8800 }, { "accuracy": 41.3086, "active_queue_size": 16384.0, "cl_loss": 4.4217, "doc_norm": 1.5281, "encoder_q-embeddings": 2614.4011, "encoder_q-layer.0": 1801.3556, "encoder_q-layer.1": 1924.95, "encoder_q-layer.10": 3834.7136, "encoder_q-layer.11": 8901.8975, "encoder_q-layer.2": 2092.4885, "encoder_q-layer.3": 2272.623, "encoder_q-layer.4": 2375.3508, "encoder_q-layer.5": 2488.3796, "encoder_q-layer.6": 2798.5544, "encoder_q-layer.7": 2952.7605, "encoder_q-layer.8": 3308.6189, "encoder_q-layer.9": 2930.8865, "epoch": 0.04, "inbatch_neg_score": 0.4866, "inbatch_pos_score": 1.0049, "learning_rate": 4.4500000000000004e-05, "loss": 4.4217, "norm_diff": 0.0164, "norm_loss": 0.0, "num_token_doc": 66.878, "num_token_overlap": 11.6677, "num_token_query": 31.3393, "num_token_union": 65.1841, "num_word_context": 202.5396, "num_word_doc": 49.9012, "num_word_query": 23.2867, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5407.6316, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4866, "query_norm": 1.5116, "queue_k_norm": 1.5273, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3393, "sent_len_1": 66.878, "sent_len_max_0": 127.9475, "sent_len_max_1": 207.7, "stdk": 0.0471, "stdq": 0.0449, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 8900 }, { "accuracy": 41.6504, "active_queue_size": 16384.0, "cl_loss": 4.3831, "doc_norm": 1.5211, "encoder_q-embeddings": 2531.386, "encoder_q-layer.0": 1667.415, "encoder_q-layer.1": 1777.0181, "encoder_q-layer.10": 3458.613, "encoder_q-layer.11": 8459.4297, "encoder_q-layer.2": 1944.6088, "encoder_q-layer.3": 2130.8735, "encoder_q-layer.4": 2219.927, "encoder_q-layer.5": 2158.5344, "encoder_q-layer.6": 2309.9858, "encoder_q-layer.7": 2392.5891, "encoder_q-layer.8": 2587.769, "encoder_q-layer.9": 2528.9534, "epoch": 0.04, "inbatch_neg_score": 0.4789, "inbatch_pos_score": 1.0039, "learning_rate": 4.5e-05, "loss": 4.3831, "norm_diff": 0.0228, "norm_loss": 0.0, "num_token_doc": 66.9443, "num_token_overlap": 11.7079, "num_token_query": 31.5404, "num_token_union": 65.2974, "num_word_context": 202.433, "num_word_doc": 49.9402, "num_word_query": 23.4388, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4972.7192, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4795, "query_norm": 1.5037, "queue_k_norm": 1.5225, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5404, "sent_len_1": 66.9443, "sent_len_max_0": 128.0, "sent_len_max_1": 211.0488, "stdk": 0.0469, "stdq": 0.0447, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9000 }, { "accuracy": 41.4062, "active_queue_size": 16384.0, "cl_loss": 4.3964, "doc_norm": 1.5182, "encoder_q-embeddings": 2213.1396, "encoder_q-layer.0": 1434.1195, "encoder_q-layer.1": 1582.8683, "encoder_q-layer.10": 4276.6885, "encoder_q-layer.11": 9572.0537, "encoder_q-layer.2": 1840.2734, "encoder_q-layer.3": 1884.8799, "encoder_q-layer.4": 1937.7517, "encoder_q-layer.5": 1895.6888, "encoder_q-layer.6": 2051.1772, "encoder_q-layer.7": 2110.3779, "encoder_q-layer.8": 2602.3708, "encoder_q-layer.9": 2650.1658, "epoch": 0.04, "inbatch_neg_score": 0.4699, "inbatch_pos_score": 0.9839, "learning_rate": 4.55e-05, "loss": 4.3964, "norm_diff": 0.0332, "norm_loss": 0.0, "num_token_doc": 66.8313, "num_token_overlap": 11.6343, "num_token_query": 31.3086, "num_token_union": 65.1502, "num_word_context": 202.1636, "num_word_doc": 49.8499, "num_word_query": 23.2452, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5260.2016, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4697, "query_norm": 1.485, "queue_k_norm": 1.5197, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3086, "sent_len_1": 66.8313, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.1475, "stdk": 0.0469, "stdq": 0.0446, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9100 }, { "accuracy": 41.7969, "active_queue_size": 16384.0, "cl_loss": 4.3778, "doc_norm": 1.517, "encoder_q-embeddings": 5577.7178, "encoder_q-layer.0": 4467.4062, "encoder_q-layer.1": 4258.2544, "encoder_q-layer.10": 3476.3816, "encoder_q-layer.11": 7855.7681, "encoder_q-layer.2": 4619.5508, "encoder_q-layer.3": 4515.6904, "encoder_q-layer.4": 3335.0059, "encoder_q-layer.5": 2737.126, "encoder_q-layer.6": 2490.6077, "encoder_q-layer.7": 2204.4221, "encoder_q-layer.8": 2715.8572, "encoder_q-layer.9": 2397.3374, "epoch": 0.04, "inbatch_neg_score": 0.4783, "inbatch_pos_score": 0.9922, "learning_rate": 4.600000000000001e-05, "loss": 4.3778, "norm_diff": 0.0246, "norm_loss": 0.0, "num_token_doc": 66.6215, "num_token_overlap": 11.6689, "num_token_query": 31.3978, "num_token_union": 65.0546, "num_word_context": 202.4173, "num_word_doc": 49.7331, "num_word_query": 23.3237, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6504.3737, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4771, "query_norm": 1.4924, "queue_k_norm": 1.5173, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3978, "sent_len_1": 66.6215, "sent_len_max_0": 127.9613, "sent_len_max_1": 206.8363, "stdk": 0.047, "stdq": 0.0443, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 9200 }, { "accuracy": 41.9434, "active_queue_size": 16384.0, "cl_loss": 4.384, "doc_norm": 1.5124, "encoder_q-embeddings": 2600.3174, "encoder_q-layer.0": 1753.323, "encoder_q-layer.1": 1800.7324, "encoder_q-layer.10": 3629.0056, "encoder_q-layer.11": 8740.6006, "encoder_q-layer.2": 1990.7451, "encoder_q-layer.3": 2015.4406, "encoder_q-layer.4": 2035.7688, "encoder_q-layer.5": 2008.316, "encoder_q-layer.6": 2135.4048, "encoder_q-layer.7": 2276.1277, "encoder_q-layer.8": 2895.3784, "encoder_q-layer.9": 2727.4302, "epoch": 0.04, "inbatch_neg_score": 0.4686, "inbatch_pos_score": 0.978, "learning_rate": 4.6500000000000005e-05, "loss": 4.384, "norm_diff": 0.0267, "norm_loss": 0.0, "num_token_doc": 66.6963, "num_token_overlap": 11.6756, "num_token_query": 31.4116, "num_token_union": 65.0827, "num_word_context": 202.149, "num_word_doc": 49.7599, "num_word_query": 23.3274, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4973.655, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4668, "query_norm": 1.4857, "queue_k_norm": 1.5133, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4116, "sent_len_1": 66.6963, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.0875, "stdk": 0.0469, "stdq": 0.0443, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9300 }, { "accuracy": 40.4785, "active_queue_size": 16384.0, "cl_loss": 4.354, "doc_norm": 1.5079, "encoder_q-embeddings": 2907.1108, "encoder_q-layer.0": 2032.8888, "encoder_q-layer.1": 2235.1345, "encoder_q-layer.10": 3570.8826, "encoder_q-layer.11": 8588.6045, "encoder_q-layer.2": 2677.396, "encoder_q-layer.3": 2589.9082, "encoder_q-layer.4": 2953.863, "encoder_q-layer.5": 2706.6843, "encoder_q-layer.6": 2465.4155, "encoder_q-layer.7": 2283.4773, "encoder_q-layer.8": 2590.5671, "encoder_q-layer.9": 2503.1975, "epoch": 0.04, "inbatch_neg_score": 0.4603, "inbatch_pos_score": 0.9712, "learning_rate": 4.7e-05, "loss": 4.354, "norm_diff": 0.0242, "norm_loss": 0.0, "num_token_doc": 66.8518, "num_token_overlap": 11.7101, "num_token_query": 31.432, "num_token_union": 65.1827, "num_word_context": 202.3527, "num_word_doc": 49.8508, "num_word_query": 23.3391, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5394.4564, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4597, "query_norm": 1.4836, "queue_k_norm": 1.5113, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.432, "sent_len_1": 66.8518, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.3137, "stdk": 0.0469, "stdq": 0.0445, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9400 }, { "accuracy": 41.6016, "active_queue_size": 16384.0, "cl_loss": 4.3496, "doc_norm": 1.5047, "encoder_q-embeddings": 2309.6797, "encoder_q-layer.0": 1471.6729, "encoder_q-layer.1": 1609.9385, "encoder_q-layer.10": 4546.2314, "encoder_q-layer.11": 9802.9775, "encoder_q-layer.2": 1824.0403, "encoder_q-layer.3": 1939.7924, "encoder_q-layer.4": 2003.6044, "encoder_q-layer.5": 2050.4019, "encoder_q-layer.6": 2245.2434, "encoder_q-layer.7": 2281.5598, "encoder_q-layer.8": 3078.6523, "encoder_q-layer.9": 2879.4585, "epoch": 0.04, "inbatch_neg_score": 0.4399, "inbatch_pos_score": 0.9365, "learning_rate": 4.75e-05, "loss": 4.3496, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.9304, "num_token_overlap": 11.6786, "num_token_query": 31.3864, "num_token_union": 65.2181, "num_word_context": 202.6518, "num_word_doc": 49.909, "num_word_query": 23.2985, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5326.1249, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4395, "query_norm": 1.4312, "queue_k_norm": 1.5065, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3864, "sent_len_1": 66.9304, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.4675, "stdk": 0.0468, "stdq": 0.0428, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9500 }, { "accuracy": 41.7969, "active_queue_size": 16384.0, "cl_loss": 4.3296, "doc_norm": 1.5054, "encoder_q-embeddings": 2080.4353, "encoder_q-layer.0": 1377.3827, "encoder_q-layer.1": 1513.3669, "encoder_q-layer.10": 4181.6748, "encoder_q-layer.11": 9494.6299, "encoder_q-layer.2": 1795.2852, "encoder_q-layer.3": 1890.4661, "encoder_q-layer.4": 2013.3645, "encoder_q-layer.5": 1998.0541, "encoder_q-layer.6": 2357.7705, "encoder_q-layer.7": 2660.7695, "encoder_q-layer.8": 3121.1313, "encoder_q-layer.9": 2847.2913, "epoch": 0.04, "inbatch_neg_score": 0.4452, "inbatch_pos_score": 0.9678, "learning_rate": 4.8e-05, "loss": 4.3296, "norm_diff": 0.0429, "norm_loss": 0.0, "num_token_doc": 66.8722, "num_token_overlap": 11.6966, "num_token_query": 31.4088, "num_token_union": 65.1617, "num_word_context": 202.3833, "num_word_doc": 49.8743, "num_word_query": 23.3231, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5317.6953, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4451, "query_norm": 1.4625, "queue_k_norm": 1.5067, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4088, "sent_len_1": 66.8722, "sent_len_max_0": 127.99, "sent_len_max_1": 209.835, "stdk": 0.0469, "stdq": 0.0438, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 9600 }, { "accuracy": 42.5293, "active_queue_size": 16384.0, "cl_loss": 4.3261, "doc_norm": 1.5043, "encoder_q-embeddings": 2726.1504, "encoder_q-layer.0": 2017.2838, "encoder_q-layer.1": 2020.6016, "encoder_q-layer.10": 3980.7637, "encoder_q-layer.11": 9194.8711, "encoder_q-layer.2": 2220.0112, "encoder_q-layer.3": 2293.4941, "encoder_q-layer.4": 2423.8135, "encoder_q-layer.5": 2208.2295, "encoder_q-layer.6": 2176.6506, "encoder_q-layer.7": 2289.3289, "encoder_q-layer.8": 2760.7361, "encoder_q-layer.9": 2534.2771, "epoch": 0.04, "inbatch_neg_score": 0.4414, "inbatch_pos_score": 0.9624, "learning_rate": 4.85e-05, "loss": 4.3261, "norm_diff": 0.021, "norm_loss": 0.0, "num_token_doc": 66.8855, "num_token_overlap": 11.6949, "num_token_query": 31.3658, "num_token_union": 65.1552, "num_word_context": 202.3894, "num_word_doc": 49.9483, "num_word_query": 23.2996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5292.3137, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4404, "query_norm": 1.4848, "queue_k_norm": 1.5011, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3658, "sent_len_1": 66.8855, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.4575, "stdk": 0.0471, "stdq": 0.0448, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9700 }, { "accuracy": 41.6016, "active_queue_size": 16384.0, "cl_loss": 4.2975, "doc_norm": 1.5005, "encoder_q-embeddings": 2333.0879, "encoder_q-layer.0": 1541.3053, "encoder_q-layer.1": 1645.0211, "encoder_q-layer.10": 3239.1968, "encoder_q-layer.11": 7713.9971, "encoder_q-layer.2": 1861.4274, "encoder_q-layer.3": 2000.1199, "encoder_q-layer.4": 2021.2444, "encoder_q-layer.5": 1960.5509, "encoder_q-layer.6": 2146.7891, "encoder_q-layer.7": 2288.2952, "encoder_q-layer.8": 2542.876, "encoder_q-layer.9": 2349.7766, "epoch": 0.04, "inbatch_neg_score": 0.4357, "inbatch_pos_score": 0.9624, "learning_rate": 4.9e-05, "loss": 4.2975, "norm_diff": 0.0415, "norm_loss": 0.0, "num_token_doc": 66.9034, "num_token_overlap": 11.7121, "num_token_query": 31.412, "num_token_union": 65.1748, "num_word_context": 202.438, "num_word_doc": 49.9165, "num_word_query": 23.3226, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4682.1893, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4355, "query_norm": 1.459, "queue_k_norm": 1.4995, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.412, "sent_len_1": 66.9034, "sent_len_max_0": 127.965, "sent_len_max_1": 209.655, "stdk": 0.047, "stdq": 0.044, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9800 }, { "accuracy": 39.8926, "active_queue_size": 16384.0, "cl_loss": 4.2916, "doc_norm": 1.4971, "encoder_q-embeddings": 2675.5562, "encoder_q-layer.0": 1889.4662, "encoder_q-layer.1": 2023.6193, "encoder_q-layer.10": 3104.6685, "encoder_q-layer.11": 8230.9316, "encoder_q-layer.2": 2276.1978, "encoder_q-layer.3": 2333.9937, "encoder_q-layer.4": 2388.8062, "encoder_q-layer.5": 2316.7139, "encoder_q-layer.6": 2465.8416, "encoder_q-layer.7": 2549.6055, "encoder_q-layer.8": 2600.2402, "encoder_q-layer.9": 2377.3584, "epoch": 0.04, "inbatch_neg_score": 0.4452, "inbatch_pos_score": 0.9526, "learning_rate": 4.9500000000000004e-05, "loss": 4.2916, "norm_diff": 0.0442, "norm_loss": 0.0, "num_token_doc": 66.8151, "num_token_overlap": 11.6884, "num_token_query": 31.3778, "num_token_union": 65.1519, "num_word_context": 202.1915, "num_word_doc": 49.8769, "num_word_query": 23.29, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5066.2353, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4438, "query_norm": 1.4528, "queue_k_norm": 1.4963, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3778, "sent_len_1": 66.8151, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.125, "stdk": 0.0469, "stdq": 0.0435, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 9900 }, { "accuracy": 42.1875, "active_queue_size": 16384.0, "cl_loss": 4.2856, "doc_norm": 1.4952, "encoder_q-embeddings": 2234.0645, "encoder_q-layer.0": 1562.6958, "encoder_q-layer.1": 1637.9476, "encoder_q-layer.10": 3597.1416, "encoder_q-layer.11": 8815.3193, "encoder_q-layer.2": 1909.8351, "encoder_q-layer.3": 2051.9202, "encoder_q-layer.4": 2143.9954, "encoder_q-layer.5": 2325.5117, "encoder_q-layer.6": 2569.7402, "encoder_q-layer.7": 2673.6125, "encoder_q-layer.8": 3048.6228, "encoder_q-layer.9": 2693.2749, "epoch": 0.04, "inbatch_neg_score": 0.4233, "inbatch_pos_score": 0.9326, "learning_rate": 5e-05, "loss": 4.2856, "norm_diff": 0.0865, "norm_loss": 0.0, "num_token_doc": 66.8736, "num_token_overlap": 11.7086, "num_token_query": 31.449, "num_token_union": 65.2093, "num_word_context": 202.6157, "num_word_doc": 49.9093, "num_word_query": 23.3528, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5140.5269, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4238, "query_norm": 1.4088, "queue_k_norm": 1.4947, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.449, "sent_len_1": 66.8736, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.8413, "stdk": 0.047, "stdq": 0.0422, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 10000 }, { "dev_runtime": 28.7853, "dev_samples_per_second": 1.112, "dev_steps_per_second": 0.035, "epoch": 0.04, "step": 10000, "test_accuracy": 91.3330078125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.5359745621681213, "test_doc_norm": 1.4143710136413574, "test_inbatch_neg_score": 0.6434903740882874, "test_inbatch_pos_score": 1.422333836555481, "test_loss": 0.5359745621681213, "test_loss_align": 1.1280993223190308, "test_loss_unif": 3.647287607192993, "test_loss_unif_q@queue": 3.647287607192993, "test_norm_diff": 0.016806721687316895, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.41647595167160034, "test_query_norm": 1.3975763320922852, "test_queue_k_norm": 1.495121955871582, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.039758309721946716, "test_stdq": 0.037743665277957916, "test_stdqueue_k": 0.04701870679855347, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.7853, "dev_samples_per_second": 1.112, "dev_steps_per_second": 0.035, "epoch": 0.04, "eval_beir-arguana_ndcg@10": 0.3248, "eval_beir-arguana_recall@10": 0.56686, "eval_beir-arguana_recall@100": 0.8862, "eval_beir-arguana_recall@20": 0.70697, "eval_beir-avg_ndcg@10": 0.33298125, "eval_beir-avg_recall@10": 0.4003310833333334, "eval_beir-avg_recall@100": 0.5832823333333333, "eval_beir-avg_recall@20": 0.4631643333333334, "eval_beir-cqadupstack_ndcg@10": 0.21387250000000002, "eval_beir-cqadupstack_recall@10": 0.30021083333333326, "eval_beir-cqadupstack_recall@100": 0.5226733333333333, "eval_beir-cqadupstack_recall@20": 0.36239333333333335, "eval_beir-fiqa_ndcg@10": 0.17907, "eval_beir-fiqa_recall@10": 0.23154, "eval_beir-fiqa_recall@100": 0.4875, "eval_beir-fiqa_recall@20": 0.30216, "eval_beir-nfcorpus_ndcg@10": 0.25181, "eval_beir-nfcorpus_recall@10": 0.11875, "eval_beir-nfcorpus_recall@100": 0.24192, "eval_beir-nfcorpus_recall@20": 0.15134, "eval_beir-nq_ndcg@10": 0.2342, "eval_beir-nq_recall@10": 0.39045, "eval_beir-nq_recall@100": 0.73059, "eval_beir-nq_recall@20": 0.50232, "eval_beir-quora_ndcg@10": 0.7098, "eval_beir-quora_recall@10": 0.8321, "eval_beir-quora_recall@100": 0.96112, "eval_beir-quora_recall@20": 0.8868, "eval_beir-scidocs_ndcg@10": 0.12546, "eval_beir-scidocs_recall@10": 0.13303, "eval_beir-scidocs_recall@100": 0.32198, "eval_beir-scidocs_recall@20": 0.18568, "eval_beir-scifact_ndcg@10": 0.5731, "eval_beir-scifact_recall@10": 0.7285, "eval_beir-scifact_recall@100": 0.87689, "eval_beir-scifact_recall@20": 0.80344, "eval_beir-trec-covid_ndcg@10": 0.5158, "eval_beir-trec-covid_recall@10": 0.568, "eval_beir-trec-covid_recall@100": 0.3956, "eval_beir-trec-covid_recall@20": 0.535, "eval_beir-webis-touche2020_ndcg@10": 0.2019, "eval_beir-webis-touche2020_recall@10": 0.13387, "eval_beir-webis-touche2020_recall@100": 0.40835, "eval_beir-webis-touche2020_recall@20": 0.19554, "eval_senteval-avg_sts": 0.7430058898520857, "eval_senteval-sickr_spearman": 0.7159969874273663, "eval_senteval-stsb_spearman": 0.7700147922768052, "step": 10000, "test_accuracy": 91.3330078125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.5359745621681213, "test_doc_norm": 1.4143710136413574, "test_inbatch_neg_score": 0.6434903740882874, "test_inbatch_pos_score": 1.422333836555481, "test_loss": 0.5359745621681213, "test_loss_align": 1.1280993223190308, "test_loss_unif": 3.647287607192993, "test_loss_unif_q@queue": 3.647287607192993, "test_norm_diff": 0.016806721687316895, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.41647595167160034, "test_query_norm": 1.3975763320922852, "test_queue_k_norm": 1.495121955871582, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.039758309721946716, "test_stdq": 0.037743665277957916, "test_stdqueue_k": 0.04701870679855347, "test_stdqueue_q": 0.0 }, { "accuracy": 41.4551, "active_queue_size": 16384.0, "cl_loss": 4.2859, "doc_norm": 1.4875, "encoder_q-embeddings": 2485.9861, "encoder_q-layer.0": 1714.8248, "encoder_q-layer.1": 1817.2434, "encoder_q-layer.10": 3184.895, "encoder_q-layer.11": 8482.7852, "encoder_q-layer.2": 1989.465, "encoder_q-layer.3": 2160.1216, "encoder_q-layer.4": 2306.3242, "encoder_q-layer.5": 2364.9873, "encoder_q-layer.6": 2414.7334, "encoder_q-layer.7": 2313.7927, "encoder_q-layer.8": 2802.5022, "encoder_q-layer.9": 2431.8428, "epoch": 0.04, "inbatch_neg_score": 0.4175, "inbatch_pos_score": 0.9429, "learning_rate": 4.9973684210526314e-05, "loss": 4.2859, "norm_diff": 0.0529, "norm_loss": 0.0, "num_token_doc": 66.802, "num_token_overlap": 11.6364, "num_token_query": 31.2722, "num_token_union": 65.1102, "num_word_context": 202.2383, "num_word_doc": 49.8481, "num_word_query": 23.2273, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4935.5829, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4167, "query_norm": 1.4346, "queue_k_norm": 1.4904, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2722, "sent_len_1": 66.802, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.8025, "stdk": 0.0468, "stdq": 0.0434, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10100 }, { "accuracy": 41.5527, "active_queue_size": 16384.0, "cl_loss": 4.2637, "doc_norm": 1.4955, "encoder_q-embeddings": 2490.6028, "encoder_q-layer.0": 1794.1995, "encoder_q-layer.1": 1909.7621, "encoder_q-layer.10": 3005.7053, "encoder_q-layer.11": 7449.6084, "encoder_q-layer.2": 2001.9727, "encoder_q-layer.3": 1963.9944, "encoder_q-layer.4": 2056.9033, "encoder_q-layer.5": 2049.9358, "encoder_q-layer.6": 2331.3079, "encoder_q-layer.7": 2150.7661, "encoder_q-layer.8": 2379.5312, "encoder_q-layer.9": 2214.5884, "epoch": 0.04, "inbatch_neg_score": 0.4324, "inbatch_pos_score": 0.9551, "learning_rate": 4.994736842105263e-05, "loss": 4.2637, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.7717, "num_token_overlap": 11.6926, "num_token_query": 31.3914, "num_token_union": 65.0825, "num_word_context": 202.2029, "num_word_doc": 49.8177, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4565.3812, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4312, "query_norm": 1.4202, "queue_k_norm": 1.4888, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3914, "sent_len_1": 66.7717, "sent_len_max_0": 127.98, "sent_len_max_1": 209.5875, "stdk": 0.0472, "stdq": 0.043, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 10200 }, { "accuracy": 40.3809, "active_queue_size": 16384.0, "cl_loss": 4.2428, "doc_norm": 1.478, "encoder_q-embeddings": 2261.9412, "encoder_q-layer.0": 1577.5626, "encoder_q-layer.1": 1623.693, "encoder_q-layer.10": 2741.0627, "encoder_q-layer.11": 7703.2148, "encoder_q-layer.2": 1837.2318, "encoder_q-layer.3": 1989.687, "encoder_q-layer.4": 2067.2944, "encoder_q-layer.5": 2010.5411, "encoder_q-layer.6": 2077.9629, "encoder_q-layer.7": 2081.9368, "encoder_q-layer.8": 2427.5247, "encoder_q-layer.9": 2240.1177, "epoch": 0.04, "inbatch_neg_score": 0.4204, "inbatch_pos_score": 0.9189, "learning_rate": 4.992105263157895e-05, "loss": 4.2428, "norm_diff": 0.0887, "norm_loss": 0.0, "num_token_doc": 66.801, "num_token_overlap": 11.6686, "num_token_query": 31.3398, "num_token_union": 65.1214, "num_word_context": 202.3614, "num_word_doc": 49.8382, "num_word_query": 23.2679, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4407.2279, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4189, "query_norm": 1.3893, "queue_k_norm": 1.4842, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3398, "sent_len_1": 66.801, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.5575, "stdk": 0.0466, "stdq": 0.0416, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10300 }, { "accuracy": 39.0137, "active_queue_size": 16384.0, "cl_loss": 4.2374, "doc_norm": 1.4791, "encoder_q-embeddings": 1654.3175, "encoder_q-layer.0": 1265.5026, "encoder_q-layer.1": 1575.2933, "encoder_q-layer.10": 1539.4871, "encoder_q-layer.11": 4547.8203, "encoder_q-layer.2": 1714.6047, "encoder_q-layer.3": 1729.5673, "encoder_q-layer.4": 1581.5338, "encoder_q-layer.5": 1347.1013, "encoder_q-layer.6": 1320.9487, "encoder_q-layer.7": 1250.5837, "encoder_q-layer.8": 1283.9219, "encoder_q-layer.9": 1148.925, "epoch": 0.05, "inbatch_neg_score": 0.3911, "inbatch_pos_score": 0.8936, "learning_rate": 4.989473684210527e-05, "loss": 4.2374, "norm_diff": 0.1118, "norm_loss": 0.0, "num_token_doc": 66.7347, "num_token_overlap": 11.6821, "num_token_query": 31.4368, "num_token_union": 65.156, "num_word_context": 202.4994, "num_word_doc": 49.8157, "num_word_query": 23.3586, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2827.3384, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3909, "query_norm": 1.3673, "queue_k_norm": 1.4818, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4368, "sent_len_1": 66.7347, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.8512, "stdk": 0.0468, "stdq": 0.0416, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10400 }, { "accuracy": 41.4062, "active_queue_size": 16384.0, "cl_loss": 4.2206, "doc_norm": 1.4809, "encoder_q-embeddings": 1758.1155, "encoder_q-layer.0": 1333.6688, "encoder_q-layer.1": 1643.5521, "encoder_q-layer.10": 1548.0603, "encoder_q-layer.11": 3683.5771, "encoder_q-layer.2": 1919.6592, "encoder_q-layer.3": 2125.998, "encoder_q-layer.4": 2568.385, "encoder_q-layer.5": 2678.3901, "encoder_q-layer.6": 3626.6125, "encoder_q-layer.7": 2797.2202, "encoder_q-layer.8": 1743.7764, "encoder_q-layer.9": 1255.2942, "epoch": 0.05, "inbatch_neg_score": 0.3997, "inbatch_pos_score": 0.9307, "learning_rate": 4.986842105263158e-05, "loss": 4.2206, "norm_diff": 0.068, "norm_loss": 0.0, "num_token_doc": 66.7367, "num_token_overlap": 11.7025, "num_token_query": 31.4131, "num_token_union": 65.0895, "num_word_context": 202.1447, "num_word_doc": 49.7997, "num_word_query": 23.3395, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3418.848, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3982, "query_norm": 1.4129, "queue_k_norm": 1.4798, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4131, "sent_len_1": 66.7367, "sent_len_max_0": 127.9525, "sent_len_max_1": 209.005, "stdk": 0.0469, "stdq": 0.043, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10500 }, { "accuracy": 42.334, "active_queue_size": 16384.0, "cl_loss": 4.25, "doc_norm": 1.4767, "encoder_q-embeddings": 1328.6001, "encoder_q-layer.0": 947.2817, "encoder_q-layer.1": 1091.6189, "encoder_q-layer.10": 1850.3174, "encoder_q-layer.11": 4875.2715, "encoder_q-layer.2": 1159.7839, "encoder_q-layer.3": 1212.7598, "encoder_q-layer.4": 1277.1453, "encoder_q-layer.5": 1325.885, "encoder_q-layer.6": 1296.0312, "encoder_q-layer.7": 1426.9554, "encoder_q-layer.8": 1628.8636, "encoder_q-layer.9": 1436.1433, "epoch": 0.05, "inbatch_neg_score": 0.3803, "inbatch_pos_score": 0.8906, "learning_rate": 4.984210526315789e-05, "loss": 4.25, "norm_diff": 0.1084, "norm_loss": 0.0, "num_token_doc": 66.8517, "num_token_overlap": 11.6381, "num_token_query": 31.3102, "num_token_union": 65.1425, "num_word_context": 202.2525, "num_word_doc": 49.8845, "num_word_query": 23.2558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2689.0748, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3809, "query_norm": 1.3683, "queue_k_norm": 1.4758, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3102, "sent_len_1": 66.8517, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.8562, "stdk": 0.0469, "stdq": 0.0415, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10600 }, { "accuracy": 42.9688, "active_queue_size": 16384.0, "cl_loss": 4.232, "doc_norm": 1.4758, "encoder_q-embeddings": 2628.7063, "encoder_q-layer.0": 1991.6401, "encoder_q-layer.1": 2357.9985, "encoder_q-layer.10": 1489.5986, "encoder_q-layer.11": 3995.8604, "encoder_q-layer.2": 2609.615, "encoder_q-layer.3": 2554.2651, "encoder_q-layer.4": 2552.5083, "encoder_q-layer.5": 2580.281, "encoder_q-layer.6": 2512.8547, "encoder_q-layer.7": 1760.9874, "encoder_q-layer.8": 1623.3141, "encoder_q-layer.9": 1310.2548, "epoch": 0.05, "inbatch_neg_score": 0.3808, "inbatch_pos_score": 0.9072, "learning_rate": 4.981578947368422e-05, "loss": 4.232, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.8906, "num_token_overlap": 11.6607, "num_token_query": 31.3268, "num_token_union": 65.1651, "num_word_context": 202.2595, "num_word_doc": 49.8928, "num_word_query": 23.2691, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3573.6616, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3811, "query_norm": 1.4029, "queue_k_norm": 1.4757, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3268, "sent_len_1": 66.8906, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.02, "stdk": 0.0469, "stdq": 0.043, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10700 }, { "accuracy": 42.9199, "active_queue_size": 16384.0, "cl_loss": 4.2121, "doc_norm": 1.4773, "encoder_q-embeddings": 2442.9829, "encoder_q-layer.0": 1872.9718, "encoder_q-layer.1": 1987.7446, "encoder_q-layer.10": 1497.8059, "encoder_q-layer.11": 4006.3486, "encoder_q-layer.2": 2395.1663, "encoder_q-layer.3": 2206.8101, "encoder_q-layer.4": 2383.3284, "encoder_q-layer.5": 2254.2617, "encoder_q-layer.6": 2311.8311, "encoder_q-layer.7": 1697.417, "encoder_q-layer.8": 1563.5547, "encoder_q-layer.9": 1211.9922, "epoch": 0.05, "inbatch_neg_score": 0.3429, "inbatch_pos_score": 0.8584, "learning_rate": 4.978947368421053e-05, "loss": 4.2121, "norm_diff": 0.1114, "norm_loss": 0.0, "num_token_doc": 66.8004, "num_token_overlap": 11.6648, "num_token_query": 31.3544, "num_token_union": 65.1253, "num_word_context": 202.5485, "num_word_doc": 49.8611, "num_word_query": 23.2887, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3338.1047, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.343, "query_norm": 1.3659, "queue_k_norm": 1.4727, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3544, "sent_len_1": 66.8004, "sent_len_max_0": 127.96, "sent_len_max_1": 208.6163, "stdk": 0.047, "stdq": 0.0423, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10800 }, { "accuracy": 42.9199, "active_queue_size": 16384.0, "cl_loss": 4.1951, "doc_norm": 1.4784, "encoder_q-embeddings": 3013.7996, "encoder_q-layer.0": 2232.7124, "encoder_q-layer.1": 2608.4463, "encoder_q-layer.10": 1428.3328, "encoder_q-layer.11": 3684.0798, "encoder_q-layer.2": 3282.1194, "encoder_q-layer.3": 3176.4956, "encoder_q-layer.4": 3192.5356, "encoder_q-layer.5": 3140.3582, "encoder_q-layer.6": 2693.614, "encoder_q-layer.7": 1882.1776, "encoder_q-layer.8": 1685.1381, "encoder_q-layer.9": 1214.2723, "epoch": 0.05, "inbatch_neg_score": 0.324, "inbatch_pos_score": 0.856, "learning_rate": 4.976315789473685e-05, "loss": 4.1951, "norm_diff": 0.1078, "norm_loss": 0.0, "num_token_doc": 67.051, "num_token_overlap": 11.74, "num_token_query": 31.516, "num_token_union": 65.3461, "num_word_context": 202.7112, "num_word_doc": 50.0275, "num_word_query": 23.4175, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3952.4363, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3245, "query_norm": 1.3706, "queue_k_norm": 1.475, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.516, "sent_len_1": 67.051, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.2188, "stdk": 0.047, "stdq": 0.0422, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 10900 }, { "accuracy": 40.8203, "active_queue_size": 16384.0, "cl_loss": 4.1986, "doc_norm": 1.4769, "encoder_q-embeddings": 6742.3335, "encoder_q-layer.0": 5066.3354, "encoder_q-layer.1": 5888.1509, "encoder_q-layer.10": 1419.9009, "encoder_q-layer.11": 3446.291, "encoder_q-layer.2": 6710.4956, "encoder_q-layer.3": 8214.3301, "encoder_q-layer.4": 8313.877, "encoder_q-layer.5": 10044.1777, "encoder_q-layer.6": 11618.6152, "encoder_q-layer.7": 6475.3994, "encoder_q-layer.8": 2419.7534, "encoder_q-layer.9": 1210.8186, "epoch": 0.05, "inbatch_neg_score": 0.3512, "inbatch_pos_score": 0.8555, "learning_rate": 4.973684210526316e-05, "loss": 4.1986, "norm_diff": 0.108, "norm_loss": 0.0, "num_token_doc": 66.8138, "num_token_overlap": 11.6954, "num_token_query": 31.4507, "num_token_union": 65.2106, "num_word_context": 202.3247, "num_word_doc": 49.8528, "num_word_query": 23.3651, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10010.624, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3516, "query_norm": 1.3689, "queue_k_norm": 1.4831, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4507, "sent_len_1": 66.8138, "sent_len_max_0": 127.9562, "sent_len_max_1": 206.5462, "stdk": 0.0468, "stdq": 0.0411, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 11000 }, { "accuracy": 43.3594, "active_queue_size": 16384.0, "cl_loss": 4.2168, "doc_norm": 1.4883, "encoder_q-embeddings": 3679.1753, "encoder_q-layer.0": 2802.1487, "encoder_q-layer.1": 3116.666, "encoder_q-layer.10": 1295.6655, "encoder_q-layer.11": 3285.304, "encoder_q-layer.2": 3479.041, "encoder_q-layer.3": 3827.8989, "encoder_q-layer.4": 4059.0208, "encoder_q-layer.5": 3943.4307, "encoder_q-layer.6": 4222.9712, "encoder_q-layer.7": 3285.9971, "encoder_q-layer.8": 2120.7698, "encoder_q-layer.9": 1143.9285, "epoch": 0.05, "inbatch_neg_score": 0.3458, "inbatch_pos_score": 0.8662, "learning_rate": 4.971052631578948e-05, "loss": 4.2168, "norm_diff": 0.0952, "norm_loss": 0.0, "num_token_doc": 66.6731, "num_token_overlap": 11.6406, "num_token_query": 31.2793, "num_token_union": 65.0312, "num_word_context": 202.0655, "num_word_doc": 49.7643, "num_word_query": 23.224, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4800.793, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3457, "query_norm": 1.3931, "queue_k_norm": 1.4884, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2793, "sent_len_1": 66.6731, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.7163, "stdk": 0.0471, "stdq": 0.0419, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 11100 }, { "accuracy": 42.7734, "active_queue_size": 16384.0, "cl_loss": 4.2125, "doc_norm": 1.492, "encoder_q-embeddings": 1557.5969, "encoder_q-layer.0": 1037.1083, "encoder_q-layer.1": 1216.4387, "encoder_q-layer.10": 1393.6013, "encoder_q-layer.11": 3955.679, "encoder_q-layer.2": 1423.0818, "encoder_q-layer.3": 1461.1857, "encoder_q-layer.4": 1465.4431, "encoder_q-layer.5": 1325.8899, "encoder_q-layer.6": 1467.8352, "encoder_q-layer.7": 1425.3727, "encoder_q-layer.8": 1529.7035, "encoder_q-layer.9": 1170.4138, "epoch": 0.05, "inbatch_neg_score": 0.3077, "inbatch_pos_score": 0.8384, "learning_rate": 4.9684210526315796e-05, "loss": 4.2125, "norm_diff": 0.0851, "norm_loss": 0.0, "num_token_doc": 66.8308, "num_token_overlap": 11.6438, "num_token_query": 31.2852, "num_token_union": 65.1446, "num_word_context": 202.3767, "num_word_doc": 49.8968, "num_word_query": 23.2487, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2478.4057, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3071, "query_norm": 1.4069, "queue_k_norm": 1.4943, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2852, "sent_len_1": 66.8308, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.2025, "stdk": 0.0471, "stdq": 0.0426, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 11200 }, { "accuracy": 43.4082, "active_queue_size": 16384.0, "cl_loss": 4.1795, "doc_norm": 1.5063, "encoder_q-embeddings": 3445.6345, "encoder_q-layer.0": 2666.8079, "encoder_q-layer.1": 3030.3235, "encoder_q-layer.10": 1389.7041, "encoder_q-layer.11": 3493.0745, "encoder_q-layer.2": 3657.269, "encoder_q-layer.3": 4068.613, "encoder_q-layer.4": 4406.5981, "encoder_q-layer.5": 4387.7729, "encoder_q-layer.6": 4611.9194, "encoder_q-layer.7": 2682.3623, "encoder_q-layer.8": 1750.6661, "encoder_q-layer.9": 1174.916, "epoch": 0.05, "inbatch_neg_score": 0.3001, "inbatch_pos_score": 0.8311, "learning_rate": 4.965789473684211e-05, "loss": 4.1795, "norm_diff": 0.1068, "norm_loss": 0.0, "num_token_doc": 66.8778, "num_token_overlap": 11.6941, "num_token_query": 31.5316, "num_token_union": 65.2865, "num_word_context": 202.7033, "num_word_doc": 49.8977, "num_word_query": 23.4366, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4933.8448, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3003, "query_norm": 1.3994, "queue_k_norm": 1.5043, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5316, "sent_len_1": 66.8778, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.315, "stdk": 0.0474, "stdq": 0.0426, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 11300 }, { "accuracy": 43.2617, "active_queue_size": 16384.0, "cl_loss": 4.1728, "doc_norm": 1.5161, "encoder_q-embeddings": 1297.6826, "encoder_q-layer.0": 917.7459, "encoder_q-layer.1": 1029.04, "encoder_q-layer.10": 1442.3202, "encoder_q-layer.11": 3650.083, "encoder_q-layer.2": 1215.645, "encoder_q-layer.3": 1310.9966, "encoder_q-layer.4": 1320.6509, "encoder_q-layer.5": 1258.7323, "encoder_q-layer.6": 1234.6663, "encoder_q-layer.7": 1171.7841, "encoder_q-layer.8": 1375.1113, "encoder_q-layer.9": 1175.0757, "epoch": 0.05, "inbatch_neg_score": 0.2738, "inbatch_pos_score": 0.813, "learning_rate": 4.9631578947368426e-05, "loss": 4.1728, "norm_diff": 0.1217, "norm_loss": 0.0, "num_token_doc": 66.957, "num_token_overlap": 11.7418, "num_token_query": 31.4988, "num_token_union": 65.2748, "num_word_context": 202.3731, "num_word_doc": 49.9721, "num_word_query": 23.3979, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2244.6448, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2761, "query_norm": 1.3945, "queue_k_norm": 1.5187, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4988, "sent_len_1": 66.957, "sent_len_max_0": 127.98, "sent_len_max_1": 206.8663, "stdk": 0.0475, "stdq": 0.042, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 11400 }, { "accuracy": 42.334, "active_queue_size": 16384.0, "cl_loss": 4.1811, "doc_norm": 1.5287, "encoder_q-embeddings": 2972.2202, "encoder_q-layer.0": 2237.5061, "encoder_q-layer.1": 2723.4678, "encoder_q-layer.10": 1626.715, "encoder_q-layer.11": 3713.8333, "encoder_q-layer.2": 3354.6162, "encoder_q-layer.3": 3380.1235, "encoder_q-layer.4": 3451.333, "encoder_q-layer.5": 3284.6086, "encoder_q-layer.6": 2962.8108, "encoder_q-layer.7": 2250.7595, "encoder_q-layer.8": 1814.1656, "encoder_q-layer.9": 1305.0256, "epoch": 0.05, "inbatch_neg_score": 0.3013, "inbatch_pos_score": 0.8076, "learning_rate": 4.960526315789474e-05, "loss": 4.1811, "norm_diff": 0.1423, "norm_loss": 0.0, "num_token_doc": 66.7953, "num_token_overlap": 11.6874, "num_token_query": 31.4245, "num_token_union": 65.1794, "num_word_context": 202.2085, "num_word_doc": 49.8144, "num_word_query": 23.3531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4177.1033, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3008, "query_norm": 1.3863, "queue_k_norm": 1.5286, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4245, "sent_len_1": 66.7953, "sent_len_max_0": 127.9613, "sent_len_max_1": 210.8113, "stdk": 0.0477, "stdq": 0.0405, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 11500 }, { "accuracy": 44.4336, "active_queue_size": 16384.0, "cl_loss": 4.143, "doc_norm": 1.5344, "encoder_q-embeddings": 3525.7898, "encoder_q-layer.0": 2475.8076, "encoder_q-layer.1": 2954.0952, "encoder_q-layer.10": 1296.7483, "encoder_q-layer.11": 2909.8896, "encoder_q-layer.2": 3687.1372, "encoder_q-layer.3": 4213.4307, "encoder_q-layer.4": 4237.1646, "encoder_q-layer.5": 4350.6118, "encoder_q-layer.6": 3883.8342, "encoder_q-layer.7": 2389.2285, "encoder_q-layer.8": 1741.995, "encoder_q-layer.9": 1186.0435, "epoch": 0.05, "inbatch_neg_score": 0.3218, "inbatch_pos_score": 0.8569, "learning_rate": 4.9578947368421055e-05, "loss": 4.143, "norm_diff": 0.1379, "norm_loss": 0.0, "num_token_doc": 66.828, "num_token_overlap": 11.6557, "num_token_query": 31.2997, "num_token_union": 65.1492, "num_word_context": 202.3207, "num_word_doc": 49.8673, "num_word_query": 23.2524, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4719.6144, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3206, "query_norm": 1.3966, "queue_k_norm": 1.5323, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2997, "sent_len_1": 66.828, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.3762, "stdk": 0.0478, "stdq": 0.041, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 11600 }, { "accuracy": 42.627, "active_queue_size": 16384.0, "cl_loss": 4.1543, "doc_norm": 1.5335, "encoder_q-embeddings": 1060.6772, "encoder_q-layer.0": 726.5411, "encoder_q-layer.1": 840.6171, "encoder_q-layer.10": 1293.623, "encoder_q-layer.11": 3116.5029, "encoder_q-layer.2": 1003.2411, "encoder_q-layer.3": 1082.744, "encoder_q-layer.4": 1079.3351, "encoder_q-layer.5": 1001.1329, "encoder_q-layer.6": 1098.7537, "encoder_q-layer.7": 1037.4342, "encoder_q-layer.8": 1140.9604, "encoder_q-layer.9": 1056.1162, "epoch": 0.05, "inbatch_neg_score": 0.3105, "inbatch_pos_score": 0.8384, "learning_rate": 4.9552631578947374e-05, "loss": 4.1543, "norm_diff": 0.1037, "norm_loss": 0.0, "num_token_doc": 66.877, "num_token_overlap": 11.6779, "num_token_query": 31.4823, "num_token_union": 65.2309, "num_word_context": 202.4189, "num_word_doc": 49.9226, "num_word_query": 23.3885, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1926.974, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3091, "query_norm": 1.4298, "queue_k_norm": 1.5379, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4823, "sent_len_1": 66.877, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.4538, "stdk": 0.0475, "stdq": 0.0418, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 11700 }, { "accuracy": 43.1641, "active_queue_size": 16384.0, "cl_loss": 4.1463, "doc_norm": 1.5429, "encoder_q-embeddings": 5847.0767, "encoder_q-layer.0": 4555.7339, "encoder_q-layer.1": 5562.9639, "encoder_q-layer.10": 1250.2197, "encoder_q-layer.11": 3195.5593, "encoder_q-layer.2": 6930.3452, "encoder_q-layer.3": 7328.9473, "encoder_q-layer.4": 6383.8203, "encoder_q-layer.5": 4842.9043, "encoder_q-layer.6": 3567.001, "encoder_q-layer.7": 2157.1824, "encoder_q-layer.8": 1659.0005, "encoder_q-layer.9": 1115.0519, "epoch": 0.05, "inbatch_neg_score": 0.3326, "inbatch_pos_score": 0.8555, "learning_rate": 4.9526315789473685e-05, "loss": 4.1463, "norm_diff": 0.1359, "norm_loss": 0.0, "num_token_doc": 66.9054, "num_token_overlap": 11.6762, "num_token_query": 31.4277, "num_token_union": 65.2649, "num_word_context": 202.4068, "num_word_doc": 49.9779, "num_word_query": 23.334, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7152.1407, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3318, "query_norm": 1.4071, "queue_k_norm": 1.5405, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4277, "sent_len_1": 66.9054, "sent_len_max_0": 127.9925, "sent_len_max_1": 206.4988, "stdk": 0.0477, "stdq": 0.0411, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 11800 }, { "accuracy": 42.3828, "active_queue_size": 16384.0, "cl_loss": 4.1466, "doc_norm": 1.5358, "encoder_q-embeddings": 1708.5524, "encoder_q-layer.0": 1140.7579, "encoder_q-layer.1": 1303.754, "encoder_q-layer.10": 1342.3525, "encoder_q-layer.11": 3579.1377, "encoder_q-layer.2": 1543.1163, "encoder_q-layer.3": 1639.2682, "encoder_q-layer.4": 1631.3123, "encoder_q-layer.5": 1421.8385, "encoder_q-layer.6": 1472.9144, "encoder_q-layer.7": 1405.3647, "encoder_q-layer.8": 1300.1698, "encoder_q-layer.9": 1055.8564, "epoch": 0.05, "inbatch_neg_score": 0.3384, "inbatch_pos_score": 0.8545, "learning_rate": 4.9500000000000004e-05, "loss": 4.1466, "norm_diff": 0.1158, "norm_loss": 0.0, "num_token_doc": 66.7852, "num_token_overlap": 11.6735, "num_token_query": 31.2768, "num_token_union": 65.0636, "num_word_context": 202.047, "num_word_doc": 49.8383, "num_word_query": 23.2279, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2502.2947, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3369, "query_norm": 1.42, "queue_k_norm": 1.5389, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2768, "sent_len_1": 66.7852, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.6413, "stdk": 0.0475, "stdq": 0.0417, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 11900 }, { "accuracy": 42.3828, "active_queue_size": 16384.0, "cl_loss": 4.1333, "doc_norm": 1.53, "encoder_q-embeddings": 4451.2642, "encoder_q-layer.0": 3362.0764, "encoder_q-layer.1": 3599.166, "encoder_q-layer.10": 1385.3374, "encoder_q-layer.11": 4097.5874, "encoder_q-layer.2": 4142.8677, "encoder_q-layer.3": 4600.4429, "encoder_q-layer.4": 4966.8696, "encoder_q-layer.5": 4044.2104, "encoder_q-layer.6": 3817.4426, "encoder_q-layer.7": 2715.8318, "encoder_q-layer.8": 2058.5537, "encoder_q-layer.9": 1163.3964, "epoch": 0.05, "inbatch_neg_score": 0.358, "inbatch_pos_score": 0.8779, "learning_rate": 4.9473684210526315e-05, "loss": 4.1333, "norm_diff": 0.0966, "norm_loss": 0.0, "num_token_doc": 66.88, "num_token_overlap": 11.7406, "num_token_query": 31.4161, "num_token_union": 65.1421, "num_word_context": 202.5614, "num_word_doc": 49.9218, "num_word_query": 23.332, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5450.3611, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3574, "query_norm": 1.4333, "queue_k_norm": 1.5334, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4161, "sent_len_1": 66.88, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.4638, "stdk": 0.0473, "stdq": 0.0419, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 12000 }, { "accuracy": 40.7227, "active_queue_size": 16384.0, "cl_loss": 4.158, "doc_norm": 1.5242, "encoder_q-embeddings": 2218.7561, "encoder_q-layer.0": 1656.3802, "encoder_q-layer.1": 1927.3011, "encoder_q-layer.10": 1255.467, "encoder_q-layer.11": 3051.5325, "encoder_q-layer.2": 2115.6646, "encoder_q-layer.3": 2031.3082, "encoder_q-layer.4": 1979.8347, "encoder_q-layer.5": 1579.0339, "encoder_q-layer.6": 1393.9805, "encoder_q-layer.7": 1154.3485, "encoder_q-layer.8": 1160.7031, "encoder_q-layer.9": 1001.8219, "epoch": 0.05, "inbatch_neg_score": 0.3188, "inbatch_pos_score": 0.834, "learning_rate": 4.9447368421052634e-05, "loss": 4.158, "norm_diff": 0.1144, "norm_loss": 0.0, "num_token_doc": 66.963, "num_token_overlap": 11.7351, "num_token_query": 31.4772, "num_token_union": 65.265, "num_word_context": 202.5863, "num_word_doc": 49.961, "num_word_query": 23.3889, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2809.8221, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3174, "query_norm": 1.4098, "queue_k_norm": 1.53, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4772, "sent_len_1": 66.963, "sent_len_max_0": 127.99, "sent_len_max_1": 209.4512, "stdk": 0.0472, "stdq": 0.0415, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 12100 }, { "accuracy": 42.7246, "active_queue_size": 16384.0, "cl_loss": 4.1549, "doc_norm": 1.5238, "encoder_q-embeddings": 3519.4514, "encoder_q-layer.0": 2660.7524, "encoder_q-layer.1": 2752.0295, "encoder_q-layer.10": 1392.3621, "encoder_q-layer.11": 3594.1091, "encoder_q-layer.2": 3070.6921, "encoder_q-layer.3": 3004.8718, "encoder_q-layer.4": 2698.7498, "encoder_q-layer.5": 1854.4354, "encoder_q-layer.6": 1735.0447, "encoder_q-layer.7": 1338.1543, "encoder_q-layer.8": 1347.5698, "encoder_q-layer.9": 1110.6194, "epoch": 0.05, "inbatch_neg_score": 0.3307, "inbatch_pos_score": 0.8418, "learning_rate": 4.942105263157895e-05, "loss": 4.1549, "norm_diff": 0.1252, "norm_loss": 0.0, "num_token_doc": 66.8204, "num_token_overlap": 11.6855, "num_token_query": 31.4309, "num_token_union": 65.174, "num_word_context": 202.2842, "num_word_doc": 49.8399, "num_word_query": 23.3383, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3795.617, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3298, "query_norm": 1.3985, "queue_k_norm": 1.526, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4309, "sent_len_1": 66.8204, "sent_len_max_0": 127.9587, "sent_len_max_1": 212.825, "stdk": 0.0473, "stdq": 0.041, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 12200 }, { "accuracy": 42.7246, "active_queue_size": 16384.0, "cl_loss": 4.1629, "doc_norm": 1.5153, "encoder_q-embeddings": 11299.1348, "encoder_q-layer.0": 8776.3369, "encoder_q-layer.1": 8126.189, "encoder_q-layer.10": 734.452, "encoder_q-layer.11": 1828.6072, "encoder_q-layer.2": 9296.6865, "encoder_q-layer.3": 9144.2217, "encoder_q-layer.4": 9630.0244, "encoder_q-layer.5": 8271.6553, "encoder_q-layer.6": 7180.2437, "encoder_q-layer.7": 5150.0645, "encoder_q-layer.8": 2424.1389, "encoder_q-layer.9": 729.9243, "epoch": 0.05, "inbatch_neg_score": 0.314, "inbatch_pos_score": 0.8398, "learning_rate": 4.9394736842105264e-05, "loss": 4.1629, "norm_diff": 0.0982, "norm_loss": 0.0, "num_token_doc": 66.7502, "num_token_overlap": 11.6757, "num_token_query": 31.3726, "num_token_union": 65.0698, "num_word_context": 202.2071, "num_word_doc": 49.7926, "num_word_query": 23.2911, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11200.5587, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.314, "query_norm": 1.4171, "queue_k_norm": 1.5183, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3726, "sent_len_1": 66.7502, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.5687, "stdk": 0.0472, "stdq": 0.0423, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 12300 }, { "accuracy": 44.1406, "active_queue_size": 16384.0, "cl_loss": 4.1672, "doc_norm": 1.5125, "encoder_q-embeddings": 1297.8113, "encoder_q-layer.0": 936.7819, "encoder_q-layer.1": 1073.1489, "encoder_q-layer.10": 616.0825, "encoder_q-layer.11": 1493.8533, "encoder_q-layer.2": 1212.0455, "encoder_q-layer.3": 1378.7141, "encoder_q-layer.4": 1493.2842, "encoder_q-layer.5": 1460.1016, "encoder_q-layer.6": 1593.4067, "encoder_q-layer.7": 1598.6946, "encoder_q-layer.8": 1259.1569, "encoder_q-layer.9": 537.6279, "epoch": 0.05, "inbatch_neg_score": 0.2752, "inbatch_pos_score": 0.8086, "learning_rate": 4.936842105263158e-05, "loss": 4.1672, "norm_diff": 0.1269, "norm_loss": 0.0, "num_token_doc": 66.7129, "num_token_overlap": 11.6334, "num_token_query": 31.3383, "num_token_union": 65.0873, "num_word_context": 202.2544, "num_word_doc": 49.7943, "num_word_query": 23.2662, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1855.0477, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2744, "query_norm": 1.3856, "queue_k_norm": 1.5071, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3383, "sent_len_1": 66.7129, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.2688, "stdk": 0.0473, "stdq": 0.042, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 12400 }, { "accuracy": 42.334, "active_queue_size": 16384.0, "cl_loss": 4.1484, "doc_norm": 1.4965, "encoder_q-embeddings": 1017.2812, "encoder_q-layer.0": 668.9263, "encoder_q-layer.1": 798.3719, "encoder_q-layer.10": 656.28, "encoder_q-layer.11": 1534.3582, "encoder_q-layer.2": 955.9298, "encoder_q-layer.3": 1041.7946, "encoder_q-layer.4": 1107.2318, "encoder_q-layer.5": 982.4708, "encoder_q-layer.6": 897.4889, "encoder_q-layer.7": 684.7886, "encoder_q-layer.8": 603.0354, "encoder_q-layer.9": 505.2349, "epoch": 0.05, "inbatch_neg_score": 0.2728, "inbatch_pos_score": 0.7915, "learning_rate": 4.9342105263157894e-05, "loss": 4.1484, "norm_diff": 0.1243, "norm_loss": 0.0, "num_token_doc": 66.8875, "num_token_overlap": 11.6779, "num_token_query": 31.3433, "num_token_union": 65.1435, "num_word_context": 202.454, "num_word_doc": 49.8849, "num_word_query": 23.2637, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1377.8595, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2722, "query_norm": 1.3722, "queue_k_norm": 1.5005, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3433, "sent_len_1": 66.8875, "sent_len_max_0": 128.0, "sent_len_max_1": 209.3787, "stdk": 0.0471, "stdq": 0.0415, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 12500 }, { "accuracy": 43.1641, "active_queue_size": 16384.0, "cl_loss": 4.1465, "doc_norm": 1.4874, "encoder_q-embeddings": 1694.4386, "encoder_q-layer.0": 1144.543, "encoder_q-layer.1": 1334.7228, "encoder_q-layer.10": 684.797, "encoder_q-layer.11": 1899.7997, "encoder_q-layer.2": 1685.2883, "encoder_q-layer.3": 1899.0433, "encoder_q-layer.4": 2130.5193, "encoder_q-layer.5": 2234.1548, "encoder_q-layer.6": 2023.7577, "encoder_q-layer.7": 1510.7997, "encoder_q-layer.8": 1120.2489, "encoder_q-layer.9": 571.4814, "epoch": 0.05, "inbatch_neg_score": 0.2464, "inbatch_pos_score": 0.7896, "learning_rate": 4.931578947368421e-05, "loss": 4.1465, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.7031, "num_token_overlap": 11.6691, "num_token_query": 31.3383, "num_token_union": 65.061, "num_word_context": 202.3702, "num_word_doc": 49.7783, "num_word_query": 23.2663, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2385.0802, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2465, "query_norm": 1.4152, "queue_k_norm": 1.4918, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3383, "sent_len_1": 66.7031, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.1075, "stdk": 0.047, "stdq": 0.043, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 12600 }, { "accuracy": 44.3359, "active_queue_size": 16384.0, "cl_loss": 4.1282, "doc_norm": 1.4834, "encoder_q-embeddings": 2082.219, "encoder_q-layer.0": 1567.3833, "encoder_q-layer.1": 1640.1774, "encoder_q-layer.10": 650.977, "encoder_q-layer.11": 1677.8828, "encoder_q-layer.2": 1832.7463, "encoder_q-layer.3": 2024.9751, "encoder_q-layer.4": 2119.8398, "encoder_q-layer.5": 2118.9597, "encoder_q-layer.6": 1806.9545, "encoder_q-layer.7": 1648.9277, "encoder_q-layer.8": 1200.3348, "encoder_q-layer.9": 591.5959, "epoch": 0.06, "inbatch_neg_score": 0.2461, "inbatch_pos_score": 0.7886, "learning_rate": 4.928947368421053e-05, "loss": 4.1282, "norm_diff": 0.0817, "norm_loss": 0.0, "num_token_doc": 66.9415, "num_token_overlap": 11.6696, "num_token_query": 31.3249, "num_token_union": 65.2021, "num_word_context": 202.3881, "num_word_doc": 49.9401, "num_word_query": 23.2664, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2549.4252, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2451, "query_norm": 1.4017, "queue_k_norm": 1.4863, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3249, "sent_len_1": 66.9415, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.635, "stdk": 0.047, "stdq": 0.0427, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 12700 }, { "accuracy": 44.3848, "active_queue_size": 16384.0, "cl_loss": 4.1469, "doc_norm": 1.4819, "encoder_q-embeddings": 2859.843, "encoder_q-layer.0": 2030.1531, "encoder_q-layer.1": 2187.2188, "encoder_q-layer.10": 623.8583, "encoder_q-layer.11": 1989.7593, "encoder_q-layer.2": 2501.6282, "encoder_q-layer.3": 2555.4329, "encoder_q-layer.4": 2655.5938, "encoder_q-layer.5": 2672.1865, "encoder_q-layer.6": 2661.9651, "encoder_q-layer.7": 2447.8486, "encoder_q-layer.8": 1785.8876, "encoder_q-layer.9": 643.6848, "epoch": 0.06, "inbatch_neg_score": 0.2706, "inbatch_pos_score": 0.8105, "learning_rate": 4.926315789473684e-05, "loss": 4.1469, "norm_diff": 0.0628, "norm_loss": 0.0, "num_token_doc": 66.8029, "num_token_overlap": 11.6757, "num_token_query": 31.378, "num_token_union": 65.1237, "num_word_context": 202.3817, "num_word_doc": 49.8267, "num_word_query": 23.2973, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3359.0466, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2698, "query_norm": 1.4192, "queue_k_norm": 1.4811, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.378, "sent_len_1": 66.8029, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.5213, "stdk": 0.0472, "stdq": 0.0434, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 12800 }, { "accuracy": 41.6504, "active_queue_size": 16384.0, "cl_loss": 4.1399, "doc_norm": 1.4737, "encoder_q-embeddings": 21983.2734, "encoder_q-layer.0": 17137.8379, "encoder_q-layer.1": 18036.6035, "encoder_q-layer.10": 645.962, "encoder_q-layer.11": 1844.4518, "encoder_q-layer.2": 20949.8828, "encoder_q-layer.3": 20913.127, "encoder_q-layer.4": 20981.9238, "encoder_q-layer.5": 18354.4434, "encoder_q-layer.6": 16998.6504, "encoder_q-layer.7": 13099.9648, "encoder_q-layer.8": 6204.2656, "encoder_q-layer.9": 1030.0553, "epoch": 0.06, "inbatch_neg_score": 0.2294, "inbatch_pos_score": 0.7554, "learning_rate": 4.923684210526316e-05, "loss": 4.1399, "norm_diff": 0.0638, "norm_loss": 0.0, "num_token_doc": 66.5769, "num_token_overlap": 11.6984, "num_token_query": 31.4048, "num_token_union": 64.9825, "num_word_context": 202.0383, "num_word_doc": 49.6804, "num_word_query": 23.3367, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24377.8078, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.2285, "query_norm": 1.4099, "queue_k_norm": 1.4733, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4048, "sent_len_1": 66.5769, "sent_len_max_0": 127.9875, "sent_len_max_1": 205.125, "stdk": 0.0471, "stdq": 0.0433, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 12900 }, { "accuracy": 45.9473, "active_queue_size": 16384.0, "cl_loss": 4.1298, "doc_norm": 1.4681, "encoder_q-embeddings": 5316.1372, "encoder_q-layer.0": 4173.7129, "encoder_q-layer.1": 4072.6152, "encoder_q-layer.10": 546.0249, "encoder_q-layer.11": 1349.9022, "encoder_q-layer.2": 3832.2456, "encoder_q-layer.3": 3521.739, "encoder_q-layer.4": 2808.5156, "encoder_q-layer.5": 2096.394, "encoder_q-layer.6": 1812.2134, "encoder_q-layer.7": 1560.6356, "encoder_q-layer.8": 1028.1343, "encoder_q-layer.9": 530.2501, "epoch": 0.06, "inbatch_neg_score": 0.2258, "inbatch_pos_score": 0.7637, "learning_rate": 4.921052631578947e-05, "loss": 4.1298, "norm_diff": 0.0688, "norm_loss": 0.0, "num_token_doc": 66.6459, "num_token_overlap": 11.6765, "num_token_query": 31.3932, "num_token_union": 65.0595, "num_word_context": 202.3596, "num_word_doc": 49.7458, "num_word_query": 23.327, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4670.2325, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2249, "query_norm": 1.3993, "queue_k_norm": 1.4692, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3932, "sent_len_1": 66.6459, "sent_len_max_0": 127.9513, "sent_len_max_1": 208.7663, "stdk": 0.047, "stdq": 0.0423, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 13000 }, { "accuracy": 43.0176, "active_queue_size": 16384.0, "cl_loss": 4.1573, "doc_norm": 1.4599, "encoder_q-embeddings": 6367.2109, "encoder_q-layer.0": 4333.7393, "encoder_q-layer.1": 4926.3877, "encoder_q-layer.10": 613.2057, "encoder_q-layer.11": 1559.181, "encoder_q-layer.2": 5632.2051, "encoder_q-layer.3": 5475.7163, "encoder_q-layer.4": 5777.5425, "encoder_q-layer.5": 5032.8037, "encoder_q-layer.6": 3987.2373, "encoder_q-layer.7": 2997.0073, "encoder_q-layer.8": 1661.2802, "encoder_q-layer.9": 660.3167, "epoch": 0.06, "inbatch_neg_score": 0.2494, "inbatch_pos_score": 0.771, "learning_rate": 4.91842105263158e-05, "loss": 4.1573, "norm_diff": 0.09, "norm_loss": 0.0, "num_token_doc": 66.607, "num_token_overlap": 11.6616, "num_token_query": 31.3548, "num_token_union": 64.975, "num_word_context": 202.0644, "num_word_doc": 49.703, "num_word_query": 23.281, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6567.6034, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2499, "query_norm": 1.3699, "queue_k_norm": 1.4633, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3548, "sent_len_1": 66.607, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.44, "stdk": 0.0469, "stdq": 0.0417, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 13100 }, { "accuracy": 43.9453, "active_queue_size": 16384.0, "cl_loss": 4.1157, "doc_norm": 1.4634, "encoder_q-embeddings": 6786.7104, "encoder_q-layer.0": 5286.7256, "encoder_q-layer.1": 5057.9858, "encoder_q-layer.10": 644.0671, "encoder_q-layer.11": 1715.6342, "encoder_q-layer.2": 5806.6357, "encoder_q-layer.3": 5974.5103, "encoder_q-layer.4": 5876.1226, "encoder_q-layer.5": 6410.6499, "encoder_q-layer.6": 6734.5923, "encoder_q-layer.7": 7645.105, "encoder_q-layer.8": 5784.9849, "encoder_q-layer.9": 1803.423, "epoch": 0.06, "inbatch_neg_score": 0.2486, "inbatch_pos_score": 0.7939, "learning_rate": 4.915789473684211e-05, "loss": 4.1157, "norm_diff": 0.0489, "norm_loss": 0.0, "num_token_doc": 66.9152, "num_token_overlap": 11.7242, "num_token_query": 31.4569, "num_token_union": 65.2102, "num_word_context": 202.4718, "num_word_doc": 49.9204, "num_word_query": 23.3548, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8443.0663, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2489, "query_norm": 1.4145, "queue_k_norm": 1.4573, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4569, "sent_len_1": 66.9152, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.5062, "stdk": 0.0473, "stdq": 0.0431, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 13200 }, { "accuracy": 43.0664, "active_queue_size": 16384.0, "cl_loss": 4.1393, "doc_norm": 1.4538, "encoder_q-embeddings": 922.9689, "encoder_q-layer.0": 667.3638, "encoder_q-layer.1": 795.1378, "encoder_q-layer.10": 321.3636, "encoder_q-layer.11": 760.245, "encoder_q-layer.2": 900.9216, "encoder_q-layer.3": 983.2035, "encoder_q-layer.4": 1003.5618, "encoder_q-layer.5": 896.5073, "encoder_q-layer.6": 884.4429, "encoder_q-layer.7": 858.9408, "encoder_q-layer.8": 858.7709, "encoder_q-layer.9": 318.5728, "epoch": 0.06, "inbatch_neg_score": 0.2569, "inbatch_pos_score": 0.79, "learning_rate": 4.913157894736842e-05, "loss": 4.1393, "norm_diff": 0.0452, "norm_loss": 0.0, "num_token_doc": 66.6577, "num_token_overlap": 11.6312, "num_token_query": 31.2916, "num_token_union": 64.9941, "num_word_context": 202.0231, "num_word_doc": 49.7212, "num_word_query": 23.2292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1209.1207, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2573, "query_norm": 1.4086, "queue_k_norm": 1.453, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2916, "sent_len_1": 66.6577, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.9187, "stdk": 0.0471, "stdq": 0.0424, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 13300 }, { "accuracy": 40.8691, "active_queue_size": 16384.0, "cl_loss": 4.1115, "doc_norm": 1.4415, "encoder_q-embeddings": 586.3137, "encoder_q-layer.0": 412.2009, "encoder_q-layer.1": 464.8342, "encoder_q-layer.10": 312.0807, "encoder_q-layer.11": 878.5349, "encoder_q-layer.2": 524.9707, "encoder_q-layer.3": 559.4534, "encoder_q-layer.4": 580.7861, "encoder_q-layer.5": 541.0971, "encoder_q-layer.6": 534.1012, "encoder_q-layer.7": 442.0788, "encoder_q-layer.8": 363.4222, "encoder_q-layer.9": 272.6577, "epoch": 0.06, "inbatch_neg_score": 0.2671, "inbatch_pos_score": 0.7822, "learning_rate": 4.910526315789474e-05, "loss": 4.1115, "norm_diff": 0.0227, "norm_loss": 0.0, "num_token_doc": 66.6458, "num_token_overlap": 11.6531, "num_token_query": 31.3713, "num_token_union": 65.093, "num_word_context": 202.1544, "num_word_doc": 49.7164, "num_word_query": 23.2876, "postclip_grad_norm": 1.0, "preclip_grad_norm": 778.9049, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2664, "query_norm": 1.4188, "queue_k_norm": 1.4473, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3713, "sent_len_1": 66.6458, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.9563, "stdk": 0.0468, "stdq": 0.0428, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 13400 }, { "accuracy": 42.8711, "active_queue_size": 16384.0, "cl_loss": 4.1292, "doc_norm": 1.4409, "encoder_q-embeddings": 646.7969, "encoder_q-layer.0": 464.1559, "encoder_q-layer.1": 549.8748, "encoder_q-layer.10": 304.4394, "encoder_q-layer.11": 803.1112, "encoder_q-layer.2": 638.3044, "encoder_q-layer.3": 673.8983, "encoder_q-layer.4": 663.6859, "encoder_q-layer.5": 660.2905, "encoder_q-layer.6": 644.2543, "encoder_q-layer.7": 474.5513, "encoder_q-layer.8": 335.8569, "encoder_q-layer.9": 249.2134, "epoch": 0.06, "inbatch_neg_score": 0.2594, "inbatch_pos_score": 0.7881, "learning_rate": 4.907894736842106e-05, "loss": 4.1292, "norm_diff": 0.0211, "norm_loss": 0.0, "num_token_doc": 66.6767, "num_token_overlap": 11.6475, "num_token_query": 31.3386, "num_token_union": 65.0686, "num_word_context": 202.2598, "num_word_doc": 49.7554, "num_word_query": 23.2674, "postclip_grad_norm": 1.0, "preclip_grad_norm": 864.1315, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2585, "query_norm": 1.4198, "queue_k_norm": 1.4397, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3386, "sent_len_1": 66.6767, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.2025, "stdk": 0.047, "stdq": 0.0423, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 13500 }, { "accuracy": 43.1152, "active_queue_size": 16384.0, "cl_loss": 4.1012, "doc_norm": 1.432, "encoder_q-embeddings": 6353.2104, "encoder_q-layer.0": 4683.2603, "encoder_q-layer.1": 4323.1138, "encoder_q-layer.10": 354.3848, "encoder_q-layer.11": 808.6419, "encoder_q-layer.2": 4724.2578, "encoder_q-layer.3": 5296.873, "encoder_q-layer.4": 5123.9717, "encoder_q-layer.5": 4156.3281, "encoder_q-layer.6": 3974.9844, "encoder_q-layer.7": 3382.7268, "encoder_q-layer.8": 2034.0114, "encoder_q-layer.9": 461.2447, "epoch": 0.06, "inbatch_neg_score": 0.247, "inbatch_pos_score": 0.7812, "learning_rate": 4.9052631578947375e-05, "loss": 4.1012, "norm_diff": 0.0093, "norm_loss": 0.0, "num_token_doc": 66.8748, "num_token_overlap": 11.7258, "num_token_query": 31.4293, "num_token_union": 65.1868, "num_word_context": 202.5605, "num_word_doc": 49.9269, "num_word_query": 23.3521, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6243.4118, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2462, "query_norm": 1.4266, "queue_k_norm": 1.436, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4293, "sent_len_1": 66.8748, "sent_len_max_0": 127.97, "sent_len_max_1": 208.7525, "stdk": 0.0468, "stdq": 0.0427, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 13600 }, { "accuracy": 42.2852, "active_queue_size": 16384.0, "cl_loss": 4.0976, "doc_norm": 1.4285, "encoder_q-embeddings": 560.3361, "encoder_q-layer.0": 406.5032, "encoder_q-layer.1": 462.673, "encoder_q-layer.10": 308.9745, "encoder_q-layer.11": 734.8492, "encoder_q-layer.2": 549.6495, "encoder_q-layer.3": 628.5969, "encoder_q-layer.4": 622.6882, "encoder_q-layer.5": 558.3777, "encoder_q-layer.6": 598.0753, "encoder_q-layer.7": 568.8099, "encoder_q-layer.8": 525.0167, "encoder_q-layer.9": 327.2018, "epoch": 0.06, "inbatch_neg_score": 0.2359, "inbatch_pos_score": 0.7578, "learning_rate": 4.902631578947369e-05, "loss": 4.0976, "norm_diff": 0.0116, "norm_loss": 0.0, "num_token_doc": 66.8949, "num_token_overlap": 11.7183, "num_token_query": 31.434, "num_token_union": 65.2343, "num_word_context": 202.6568, "num_word_doc": 49.957, "num_word_query": 23.3599, "postclip_grad_norm": 1.0, "preclip_grad_norm": 799.585, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2349, "query_norm": 1.4401, "queue_k_norm": 1.4318, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.434, "sent_len_1": 66.8949, "sent_len_max_0": 127.98, "sent_len_max_1": 209.4688, "stdk": 0.0468, "stdq": 0.0418, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 13700 }, { "accuracy": 41.4062, "active_queue_size": 16384.0, "cl_loss": 4.114, "doc_norm": 1.4244, "encoder_q-embeddings": 556.9622, "encoder_q-layer.0": 394.0476, "encoder_q-layer.1": 445.8925, "encoder_q-layer.10": 327.1558, "encoder_q-layer.11": 828.6865, "encoder_q-layer.2": 530.7934, "encoder_q-layer.3": 570.6962, "encoder_q-layer.4": 641.3592, "encoder_q-layer.5": 681.9541, "encoder_q-layer.6": 856.2503, "encoder_q-layer.7": 833.569, "encoder_q-layer.8": 625.8636, "encoder_q-layer.9": 486.009, "epoch": 0.06, "inbatch_neg_score": 0.2676, "inbatch_pos_score": 0.7891, "learning_rate": 4.9e-05, "loss": 4.114, "norm_diff": 0.0223, "norm_loss": 0.0, "num_token_doc": 66.7944, "num_token_overlap": 11.6965, "num_token_query": 31.3725, "num_token_union": 65.1134, "num_word_context": 202.2616, "num_word_doc": 49.84, "num_word_query": 23.2951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 905.9443, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2666, "query_norm": 1.4467, "queue_k_norm": 1.4261, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3725, "sent_len_1": 66.7944, "sent_len_max_0": 127.995, "sent_len_max_1": 206.9363, "stdk": 0.0467, "stdq": 0.0428, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 13800 }, { "accuracy": 44.6777, "active_queue_size": 16384.0, "cl_loss": 4.1171, "doc_norm": 1.4215, "encoder_q-embeddings": 3742.4033, "encoder_q-layer.0": 2904.1904, "encoder_q-layer.1": 2822.5972, "encoder_q-layer.10": 299.8621, "encoder_q-layer.11": 887.4989, "encoder_q-layer.2": 3201.0745, "encoder_q-layer.3": 3305.1873, "encoder_q-layer.4": 2905.001, "encoder_q-layer.5": 2902.6614, "encoder_q-layer.6": 2800.2515, "encoder_q-layer.7": 1973.8619, "encoder_q-layer.8": 1052.0579, "encoder_q-layer.9": 422.3742, "epoch": 0.06, "inbatch_neg_score": 0.2785, "inbatch_pos_score": 0.8188, "learning_rate": 4.897368421052632e-05, "loss": 4.1171, "norm_diff": 0.0302, "norm_loss": 0.0, "num_token_doc": 66.6568, "num_token_overlap": 11.6777, "num_token_query": 31.4114, "num_token_union": 65.078, "num_word_context": 201.9672, "num_word_doc": 49.7617, "num_word_query": 23.3291, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3863.2281, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2781, "query_norm": 1.4517, "queue_k_norm": 1.4236, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4114, "sent_len_1": 66.6568, "sent_len_max_0": 127.9712, "sent_len_max_1": 206.8762, "stdk": 0.0467, "stdq": 0.0436, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 13900 }, { "accuracy": 45.8496, "active_queue_size": 16384.0, "cl_loss": 4.1115, "doc_norm": 1.4173, "encoder_q-embeddings": 405.3995, "encoder_q-layer.0": 280.0427, "encoder_q-layer.1": 317.7815, "encoder_q-layer.10": 333.7409, "encoder_q-layer.11": 773.5087, "encoder_q-layer.2": 378.0012, "encoder_q-layer.3": 373.9242, "encoder_q-layer.4": 353.8253, "encoder_q-layer.5": 294.0417, "encoder_q-layer.6": 289.1362, "encoder_q-layer.7": 271.0944, "encoder_q-layer.8": 283.994, "encoder_q-layer.9": 266.5187, "epoch": 0.06, "inbatch_neg_score": 0.2602, "inbatch_pos_score": 0.7964, "learning_rate": 4.8947368421052635e-05, "loss": 4.1115, "norm_diff": 0.066, "norm_loss": 0.0, "num_token_doc": 66.7683, "num_token_overlap": 11.678, "num_token_query": 31.3116, "num_token_union": 65.0783, "num_word_context": 202.1902, "num_word_doc": 49.843, "num_word_query": 23.2538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 553.1505, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2612, "query_norm": 1.4832, "queue_k_norm": 1.4194, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3116, "sent_len_1": 66.7683, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.1113, "stdk": 0.0467, "stdq": 0.0425, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 14000 }, { "accuracy": 44.7266, "active_queue_size": 16384.0, "cl_loss": 4.1112, "doc_norm": 1.4183, "encoder_q-embeddings": 1005.873, "encoder_q-layer.0": 715.0676, "encoder_q-layer.1": 835.6938, "encoder_q-layer.10": 306.8505, "encoder_q-layer.11": 757.2634, "encoder_q-layer.2": 966.8318, "encoder_q-layer.3": 964.7645, "encoder_q-layer.4": 972.7808, "encoder_q-layer.5": 923.4133, "encoder_q-layer.6": 983.4688, "encoder_q-layer.7": 876.8048, "encoder_q-layer.8": 744.2866, "encoder_q-layer.9": 405.0046, "epoch": 0.06, "inbatch_neg_score": 0.2777, "inbatch_pos_score": 0.8125, "learning_rate": 4.8921052631578953e-05, "loss": 4.1112, "norm_diff": 0.0356, "norm_loss": 0.0, "num_token_doc": 66.7629, "num_token_overlap": 11.6395, "num_token_query": 31.2581, "num_token_union": 65.0744, "num_word_context": 202.1765, "num_word_doc": 49.8256, "num_word_query": 23.2174, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1239.102, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2769, "query_norm": 1.4539, "queue_k_norm": 1.4163, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2581, "sent_len_1": 66.7629, "sent_len_max_0": 128.0, "sent_len_max_1": 210.4625, "stdk": 0.0467, "stdq": 0.0428, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 14100 }, { "accuracy": 46.2402, "active_queue_size": 16384.0, "cl_loss": 4.1281, "doc_norm": 1.4074, "encoder_q-embeddings": 1081.6343, "encoder_q-layer.0": 752.3714, "encoder_q-layer.1": 834.3423, "encoder_q-layer.10": 353.4848, "encoder_q-layer.11": 770.293, "encoder_q-layer.2": 920.72, "encoder_q-layer.3": 970.0049, "encoder_q-layer.4": 968.0801, "encoder_q-layer.5": 937.4216, "encoder_q-layer.6": 928.1295, "encoder_q-layer.7": 868.1292, "encoder_q-layer.8": 1163.0945, "encoder_q-layer.9": 1057.6251, "epoch": 0.06, "inbatch_neg_score": 0.2755, "inbatch_pos_score": 0.8179, "learning_rate": 4.8894736842105265e-05, "loss": 4.1281, "norm_diff": 0.0975, "norm_loss": 0.0, "num_token_doc": 66.7195, "num_token_overlap": 11.671, "num_token_query": 31.3043, "num_token_union": 65.0356, "num_word_context": 202.325, "num_word_doc": 49.7604, "num_word_query": 23.2477, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1367.4986, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2739, "query_norm": 1.505, "queue_k_norm": 1.4135, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3043, "sent_len_1": 66.7195, "sent_len_max_0": 127.975, "sent_len_max_1": 209.3713, "stdk": 0.0464, "stdq": 0.043, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 14200 }, { "accuracy": 41.8945, "active_queue_size": 16384.0, "cl_loss": 4.1036, "doc_norm": 1.4106, "encoder_q-embeddings": 967.9284, "encoder_q-layer.0": 765.7979, "encoder_q-layer.1": 818.9968, "encoder_q-layer.10": 315.2484, "encoder_q-layer.11": 882.7838, "encoder_q-layer.2": 894.3557, "encoder_q-layer.3": 917.9337, "encoder_q-layer.4": 855.1702, "encoder_q-layer.5": 698.1514, "encoder_q-layer.6": 615.124, "encoder_q-layer.7": 507.1271, "encoder_q-layer.8": 507.1643, "encoder_q-layer.9": 333.2693, "epoch": 0.06, "inbatch_neg_score": 0.2856, "inbatch_pos_score": 0.8027, "learning_rate": 4.886842105263158e-05, "loss": 4.1036, "norm_diff": 0.0567, "norm_loss": 0.0, "num_token_doc": 66.7753, "num_token_overlap": 11.7105, "num_token_query": 31.4053, "num_token_union": 65.1037, "num_word_context": 202.168, "num_word_doc": 49.8106, "num_word_query": 23.3247, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1099.501, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2844, "query_norm": 1.4673, "queue_k_norm": 1.4104, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4053, "sent_len_1": 66.7753, "sent_len_max_0": 128.0, "sent_len_max_1": 209.2475, "stdk": 0.0465, "stdq": 0.0424, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 14300 }, { "accuracy": 43.8477, "active_queue_size": 16384.0, "cl_loss": 4.101, "doc_norm": 1.4037, "encoder_q-embeddings": 2962.0032, "encoder_q-layer.0": 2335.1428, "encoder_q-layer.1": 2441.6819, "encoder_q-layer.10": 321.3606, "encoder_q-layer.11": 873.1422, "encoder_q-layer.2": 2715.3003, "encoder_q-layer.3": 2820.0671, "encoder_q-layer.4": 2748.2754, "encoder_q-layer.5": 2597.5222, "encoder_q-layer.6": 2466.4248, "encoder_q-layer.7": 1854.1633, "encoder_q-layer.8": 1379.6816, "encoder_q-layer.9": 596.4918, "epoch": 0.06, "inbatch_neg_score": 0.2826, "inbatch_pos_score": 0.8071, "learning_rate": 4.8842105263157895e-05, "loss": 4.101, "norm_diff": 0.1179, "norm_loss": 0.0, "num_token_doc": 66.7299, "num_token_overlap": 11.6904, "num_token_query": 31.4151, "num_token_union": 65.13, "num_word_context": 202.2464, "num_word_doc": 49.8147, "num_word_query": 23.3551, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3334.9307, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2815, "query_norm": 1.5217, "queue_k_norm": 1.4091, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4151, "sent_len_1": 66.7299, "sent_len_max_0": 127.965, "sent_len_max_1": 210.7425, "stdk": 0.0463, "stdq": 0.0435, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 14400 }, { "accuracy": 41.3574, "active_queue_size": 16384.0, "cl_loss": 4.1186, "doc_norm": 1.4108, "encoder_q-embeddings": 369.4982, "encoder_q-layer.0": 248.3632, "encoder_q-layer.1": 275.3571, "encoder_q-layer.10": 348.2555, "encoder_q-layer.11": 846.0343, "encoder_q-layer.2": 305.7503, "encoder_q-layer.3": 338.106, "encoder_q-layer.4": 334.8008, "encoder_q-layer.5": 339.2592, "encoder_q-layer.6": 363.0708, "encoder_q-layer.7": 360.6147, "encoder_q-layer.8": 400.8007, "encoder_q-layer.9": 322.6238, "epoch": 0.06, "inbatch_neg_score": 0.2886, "inbatch_pos_score": 0.832, "learning_rate": 4.881578947368421e-05, "loss": 4.1186, "norm_diff": 0.1376, "norm_loss": 0.0, "num_token_doc": 66.9725, "num_token_overlap": 11.6794, "num_token_query": 31.3569, "num_token_union": 65.1924, "num_word_context": 202.4602, "num_word_doc": 49.932, "num_word_query": 23.2801, "postclip_grad_norm": 1.0, "preclip_grad_norm": 576.3327, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2876, "query_norm": 1.5484, "queue_k_norm": 1.4081, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3569, "sent_len_1": 66.9725, "sent_len_max_0": 127.975, "sent_len_max_1": 212.1175, "stdk": 0.0467, "stdq": 0.0441, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 14500 }, { "accuracy": 41.6504, "active_queue_size": 16384.0, "cl_loss": 4.1027, "doc_norm": 1.4054, "encoder_q-embeddings": 1501.7491, "encoder_q-layer.0": 1018.3691, "encoder_q-layer.1": 1112.9661, "encoder_q-layer.10": 331.9124, "encoder_q-layer.11": 844.6509, "encoder_q-layer.2": 1225.2614, "encoder_q-layer.3": 1365.5013, "encoder_q-layer.4": 1455.6514, "encoder_q-layer.5": 1595.7126, "encoder_q-layer.6": 1531.7118, "encoder_q-layer.7": 1109.0884, "encoder_q-layer.8": 703.0399, "encoder_q-layer.9": 308.2377, "epoch": 0.06, "inbatch_neg_score": 0.3043, "inbatch_pos_score": 0.8286, "learning_rate": 4.878947368421053e-05, "loss": 4.1027, "norm_diff": 0.1209, "norm_loss": 0.0, "num_token_doc": 66.7735, "num_token_overlap": 11.6766, "num_token_query": 31.2872, "num_token_union": 65.0692, "num_word_context": 202.1438, "num_word_doc": 49.8021, "num_word_query": 23.2459, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1752.3136, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3037, "query_norm": 1.5263, "queue_k_norm": 1.4057, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2872, "sent_len_1": 66.7735, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.4837, "stdk": 0.0465, "stdq": 0.0435, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 14600 }, { "accuracy": 42.8223, "active_queue_size": 16384.0, "cl_loss": 4.1133, "doc_norm": 1.4091, "encoder_q-embeddings": 1417.7109, "encoder_q-layer.0": 992.3379, "encoder_q-layer.1": 1022.0662, "encoder_q-layer.10": 349.1152, "encoder_q-layer.11": 901.6198, "encoder_q-layer.2": 1179.9845, "encoder_q-layer.3": 1241.9329, "encoder_q-layer.4": 1036.629, "encoder_q-layer.5": 1051.0206, "encoder_q-layer.6": 929.3427, "encoder_q-layer.7": 878.5806, "encoder_q-layer.8": 496.8496, "encoder_q-layer.9": 318.0041, "epoch": 0.06, "inbatch_neg_score": 0.3171, "inbatch_pos_score": 0.8447, "learning_rate": 4.876315789473684e-05, "loss": 4.1133, "norm_diff": 0.1404, "norm_loss": 0.0, "num_token_doc": 66.6944, "num_token_overlap": 11.6944, "num_token_query": 31.4353, "num_token_union": 65.1076, "num_word_context": 202.3466, "num_word_doc": 49.772, "num_word_query": 23.3578, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1483.1055, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3162, "query_norm": 1.5495, "queue_k_norm": 1.4045, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4353, "sent_len_1": 66.6944, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.1425, "stdk": 0.0466, "stdq": 0.0435, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 14700 }, { "accuracy": 42.334, "active_queue_size": 16384.0, "cl_loss": 4.1347, "doc_norm": 1.4093, "encoder_q-embeddings": 1001.3156, "encoder_q-layer.0": 658.4723, "encoder_q-layer.1": 760.9753, "encoder_q-layer.10": 320.9224, "encoder_q-layer.11": 790.1544, "encoder_q-layer.2": 937.6426, "encoder_q-layer.3": 1006.3999, "encoder_q-layer.4": 1049.9636, "encoder_q-layer.5": 1187.2847, "encoder_q-layer.6": 1273.1819, "encoder_q-layer.7": 1444.0588, "encoder_q-layer.8": 794.9689, "encoder_q-layer.9": 303.4027, "epoch": 0.06, "inbatch_neg_score": 0.3366, "inbatch_pos_score": 0.8657, "learning_rate": 4.873684210526316e-05, "loss": 4.1347, "norm_diff": 0.0985, "norm_loss": 0.0, "num_token_doc": 66.7828, "num_token_overlap": 11.6769, "num_token_query": 31.3052, "num_token_union": 65.0849, "num_word_context": 202.356, "num_word_doc": 49.8458, "num_word_query": 23.2624, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1419.7096, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.335, "query_norm": 1.5078, "queue_k_norm": 1.4059, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3052, "sent_len_1": 66.7828, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.1775, "stdk": 0.0466, "stdq": 0.0435, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 14800 }, { "accuracy": 43.6523, "active_queue_size": 16384.0, "cl_loss": 4.1138, "doc_norm": 1.4057, "encoder_q-embeddings": 1265.6128, "encoder_q-layer.0": 871.4995, "encoder_q-layer.1": 1014.0472, "encoder_q-layer.10": 294.6478, "encoder_q-layer.11": 843.1122, "encoder_q-layer.2": 1063.6116, "encoder_q-layer.3": 1069.2944, "encoder_q-layer.4": 1036.8027, "encoder_q-layer.5": 998.6819, "encoder_q-layer.6": 991.2288, "encoder_q-layer.7": 862.2176, "encoder_q-layer.8": 678.7743, "encoder_q-layer.9": 389.6417, "epoch": 0.06, "inbatch_neg_score": 0.3497, "inbatch_pos_score": 0.8794, "learning_rate": 4.871052631578948e-05, "loss": 4.1138, "norm_diff": 0.0571, "norm_loss": 0.0, "num_token_doc": 66.9078, "num_token_overlap": 11.7468, "num_token_query": 31.5327, "num_token_union": 65.2365, "num_word_context": 202.3507, "num_word_doc": 49.8916, "num_word_query": 23.4368, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1389.2343, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3496, "query_norm": 1.4628, "queue_k_norm": 1.405, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5327, "sent_len_1": 66.9078, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.0625, "stdk": 0.0464, "stdq": 0.0431, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 14900 }, { "accuracy": 40.8203, "active_queue_size": 16384.0, "cl_loss": 4.145, "doc_norm": 1.4019, "encoder_q-embeddings": 766.3836, "encoder_q-layer.0": 519.3179, "encoder_q-layer.1": 596.3865, "encoder_q-layer.10": 300.7934, "encoder_q-layer.11": 861.8022, "encoder_q-layer.2": 681.7211, "encoder_q-layer.3": 771.252, "encoder_q-layer.4": 848.1283, "encoder_q-layer.5": 790.6493, "encoder_q-layer.6": 677.291, "encoder_q-layer.7": 593.4908, "encoder_q-layer.8": 473.5273, "encoder_q-layer.9": 284.648, "epoch": 0.07, "inbatch_neg_score": 0.3515, "inbatch_pos_score": 0.8691, "learning_rate": 4.868421052631579e-05, "loss": 4.145, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.8234, "num_token_overlap": 11.6851, "num_token_query": 31.3761, "num_token_union": 65.1202, "num_word_context": 202.3071, "num_word_doc": 49.8436, "num_word_query": 23.3217, "postclip_grad_norm": 1.0, "preclip_grad_norm": 973.3314, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3513, "query_norm": 1.4702, "queue_k_norm": 1.4056, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3761, "sent_len_1": 66.8234, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.2063, "stdk": 0.0462, "stdq": 0.0437, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 15000 }, { "accuracy": 45.3125, "active_queue_size": 16384.0, "cl_loss": 4.1654, "doc_norm": 1.4116, "encoder_q-embeddings": 823.1271, "encoder_q-layer.0": 568.905, "encoder_q-layer.1": 645.0121, "encoder_q-layer.10": 306.7122, "encoder_q-layer.11": 875.6485, "encoder_q-layer.2": 772.54, "encoder_q-layer.3": 864.9798, "encoder_q-layer.4": 866.2765, "encoder_q-layer.5": 827.169, "encoder_q-layer.6": 813.9854, "encoder_q-layer.7": 721.8887, "encoder_q-layer.8": 622.4892, "encoder_q-layer.9": 416.0352, "epoch": 0.07, "inbatch_neg_score": 0.3458, "inbatch_pos_score": 0.8926, "learning_rate": 4.865789473684211e-05, "loss": 4.1654, "norm_diff": 0.0415, "norm_loss": 0.0, "num_token_doc": 66.7826, "num_token_overlap": 11.6694, "num_token_query": 31.331, "num_token_union": 65.1075, "num_word_context": 202.3559, "num_word_doc": 49.8344, "num_word_query": 23.2631, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1093.5345, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3457, "query_norm": 1.4531, "queue_k_norm": 1.4062, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.331, "sent_len_1": 66.7826, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.8363, "stdk": 0.0465, "stdq": 0.0435, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15100 }, { "accuracy": 43.2129, "active_queue_size": 16384.0, "cl_loss": 4.162, "doc_norm": 1.4041, "encoder_q-embeddings": 601.416, "encoder_q-layer.0": 423.3934, "encoder_q-layer.1": 490.442, "encoder_q-layer.10": 300.7068, "encoder_q-layer.11": 959.7146, "encoder_q-layer.2": 568.4876, "encoder_q-layer.3": 631.449, "encoder_q-layer.4": 690.8093, "encoder_q-layer.5": 796.8056, "encoder_q-layer.6": 774.8477, "encoder_q-layer.7": 781.8669, "encoder_q-layer.8": 531.7261, "encoder_q-layer.9": 280.7806, "epoch": 0.07, "inbatch_neg_score": 0.3186, "inbatch_pos_score": 0.8604, "learning_rate": 4.863157894736842e-05, "loss": 4.162, "norm_diff": 0.0332, "norm_loss": 0.0, "num_token_doc": 66.6866, "num_token_overlap": 11.7154, "num_token_query": 31.4269, "num_token_union": 65.0667, "num_word_context": 202.172, "num_word_doc": 49.7687, "num_word_query": 23.3452, "postclip_grad_norm": 1.0, "preclip_grad_norm": 932.6933, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3181, "query_norm": 1.4373, "queue_k_norm": 1.4073, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4269, "sent_len_1": 66.6866, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.4925, "stdk": 0.0462, "stdq": 0.0437, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15200 }, { "accuracy": 43.5547, "active_queue_size": 16384.0, "cl_loss": 4.1512, "doc_norm": 1.4119, "encoder_q-embeddings": 4586.3281, "encoder_q-layer.0": 3486.9895, "encoder_q-layer.1": 3429.3391, "encoder_q-layer.10": 772.2046, "encoder_q-layer.11": 2050.8037, "encoder_q-layer.2": 4142.0249, "encoder_q-layer.3": 4368.7173, "encoder_q-layer.4": 4103.3101, "encoder_q-layer.5": 4049.5081, "encoder_q-layer.6": 3899.0222, "encoder_q-layer.7": 3562.3093, "encoder_q-layer.8": 2680.0745, "encoder_q-layer.9": 1053.8466, "epoch": 0.07, "inbatch_neg_score": 0.2993, "inbatch_pos_score": 0.8477, "learning_rate": 4.860526315789474e-05, "loss": 4.1512, "norm_diff": 0.0339, "norm_loss": 0.0, "num_token_doc": 66.8929, "num_token_overlap": 11.6945, "num_token_query": 31.4124, "num_token_union": 65.2282, "num_word_context": 202.5643, "num_word_doc": 49.8792, "num_word_query": 23.3376, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5271.3571, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2998, "query_norm": 1.4458, "queue_k_norm": 1.4078, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4124, "sent_len_1": 66.8929, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.4825, "stdk": 0.0465, "stdq": 0.0443, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15300 }, { "accuracy": 44.9707, "active_queue_size": 16384.0, "cl_loss": 4.1447, "doc_norm": 1.4093, "encoder_q-embeddings": 4335.5728, "encoder_q-layer.0": 3197.5918, "encoder_q-layer.1": 3427.1416, "encoder_q-layer.10": 632.2378, "encoder_q-layer.11": 1849.8435, "encoder_q-layer.2": 4115.5806, "encoder_q-layer.3": 4043.7988, "encoder_q-layer.4": 3688.0559, "encoder_q-layer.5": 3111.9739, "encoder_q-layer.6": 2747.0903, "encoder_q-layer.7": 2163.2854, "encoder_q-layer.8": 960.3823, "encoder_q-layer.9": 594.8848, "epoch": 0.07, "inbatch_neg_score": 0.2977, "inbatch_pos_score": 0.8345, "learning_rate": 4.857894736842106e-05, "loss": 4.1447, "norm_diff": 0.0213, "norm_loss": 0.0, "num_token_doc": 66.9054, "num_token_overlap": 11.6681, "num_token_query": 31.4124, "num_token_union": 65.2194, "num_word_context": 202.7411, "num_word_doc": 49.9123, "num_word_query": 23.3346, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4544.0426, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2988, "query_norm": 1.3895, "queue_k_norm": 1.4083, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4124, "sent_len_1": 66.9054, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.955, "stdk": 0.0463, "stdq": 0.0427, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 15400 }, { "accuracy": 44.1895, "active_queue_size": 16384.0, "cl_loss": 4.1221, "doc_norm": 1.4093, "encoder_q-embeddings": 843.0743, "encoder_q-layer.0": 575.9454, "encoder_q-layer.1": 637.6025, "encoder_q-layer.10": 635.2148, "encoder_q-layer.11": 1813.5272, "encoder_q-layer.2": 731.7136, "encoder_q-layer.3": 776.5207, "encoder_q-layer.4": 816.1678, "encoder_q-layer.5": 863.7488, "encoder_q-layer.6": 889.1948, "encoder_q-layer.7": 894.3333, "encoder_q-layer.8": 791.1005, "encoder_q-layer.9": 555.3735, "epoch": 0.07, "inbatch_neg_score": 0.2654, "inbatch_pos_score": 0.8105, "learning_rate": 4.855263157894737e-05, "loss": 4.1221, "norm_diff": 0.0158, "norm_loss": 0.0, "num_token_doc": 66.8537, "num_token_overlap": 11.6921, "num_token_query": 31.3982, "num_token_union": 65.1645, "num_word_context": 202.1844, "num_word_doc": 49.873, "num_word_query": 23.3177, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1308.7907, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2646, "query_norm": 1.4251, "queue_k_norm": 1.4056, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3982, "sent_len_1": 66.8537, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.3038, "stdk": 0.0464, "stdq": 0.0443, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15500 }, { "accuracy": 42.8223, "active_queue_size": 16384.0, "cl_loss": 4.1369, "doc_norm": 1.405, "encoder_q-embeddings": 1895.7908, "encoder_q-layer.0": 1378.0952, "encoder_q-layer.1": 1433.1897, "encoder_q-layer.10": 588.6588, "encoder_q-layer.11": 1683.9709, "encoder_q-layer.2": 1552.6649, "encoder_q-layer.3": 1562.806, "encoder_q-layer.4": 1575.8145, "encoder_q-layer.5": 1554.7391, "encoder_q-layer.6": 1300.4496, "encoder_q-layer.7": 1311.5142, "encoder_q-layer.8": 914.8396, "encoder_q-layer.9": 527.5235, "epoch": 0.07, "inbatch_neg_score": 0.2545, "inbatch_pos_score": 0.7852, "learning_rate": 4.852631578947369e-05, "loss": 4.1369, "norm_diff": 0.0238, "norm_loss": 0.0, "num_token_doc": 66.7795, "num_token_overlap": 11.653, "num_token_query": 31.2879, "num_token_union": 65.0779, "num_word_context": 202.4097, "num_word_doc": 49.8271, "num_word_query": 23.2307, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2125.0731, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2544, "query_norm": 1.3812, "queue_k_norm": 1.4038, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2879, "sent_len_1": 66.7795, "sent_len_max_0": 127.98, "sent_len_max_1": 208.5225, "stdk": 0.0463, "stdq": 0.0429, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15600 }, { "accuracy": 43.1152, "active_queue_size": 16384.0, "cl_loss": 4.1125, "doc_norm": 1.3995, "encoder_q-embeddings": 1021.9699, "encoder_q-layer.0": 705.3054, "encoder_q-layer.1": 806.9128, "encoder_q-layer.10": 561.676, "encoder_q-layer.11": 1593.7676, "encoder_q-layer.2": 962.4338, "encoder_q-layer.3": 1012.4657, "encoder_q-layer.4": 1113.6599, "encoder_q-layer.5": 1097.9653, "encoder_q-layer.6": 1122.4672, "encoder_q-layer.7": 1150.8696, "encoder_q-layer.8": 852.4541, "encoder_q-layer.9": 526.9712, "epoch": 0.07, "inbatch_neg_score": 0.2225, "inbatch_pos_score": 0.7559, "learning_rate": 4.85e-05, "loss": 4.1125, "norm_diff": 0.0255, "norm_loss": 0.0, "num_token_doc": 67.0142, "num_token_overlap": 11.7101, "num_token_query": 31.4524, "num_token_union": 65.2687, "num_word_context": 202.5741, "num_word_doc": 50.0277, "num_word_query": 23.3736, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1486.421, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2225, "query_norm": 1.374, "queue_k_norm": 1.4003, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4524, "sent_len_1": 67.0142, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.525, "stdk": 0.0462, "stdq": 0.0429, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15700 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 4.1288, "doc_norm": 1.3951, "encoder_q-embeddings": 17792.6465, "encoder_q-layer.0": 12745.5723, "encoder_q-layer.1": 13364.0869, "encoder_q-layer.10": 791.4532, "encoder_q-layer.11": 1996.6624, "encoder_q-layer.2": 17452.1133, "encoder_q-layer.3": 20893.3984, "encoder_q-layer.4": 22482.1406, "encoder_q-layer.5": 22474.8203, "encoder_q-layer.6": 25807.4043, "encoder_q-layer.7": 17657.4824, "encoder_q-layer.8": 7747.4922, "encoder_q-layer.9": 1223.0056, "epoch": 0.07, "inbatch_neg_score": 0.2247, "inbatch_pos_score": 0.7515, "learning_rate": 4.847368421052632e-05, "loss": 4.1288, "norm_diff": 0.0133, "norm_loss": 0.0, "num_token_doc": 66.8911, "num_token_overlap": 11.6614, "num_token_query": 31.3598, "num_token_union": 65.201, "num_word_context": 202.6079, "num_word_doc": 49.9478, "num_word_query": 23.2909, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24646.069, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.2256, "query_norm": 1.3819, "queue_k_norm": 1.3964, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3598, "sent_len_1": 66.8911, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.4988, "stdk": 0.0462, "stdq": 0.0422, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 15800 }, { "accuracy": 41.2598, "active_queue_size": 16384.0, "cl_loss": 4.1228, "doc_norm": 1.3891, "encoder_q-embeddings": 872.7041, "encoder_q-layer.0": 610.5565, "encoder_q-layer.1": 718.2017, "encoder_q-layer.10": 579.4019, "encoder_q-layer.11": 1690.0852, "encoder_q-layer.2": 819.1313, "encoder_q-layer.3": 878.16, "encoder_q-layer.4": 939.8047, "encoder_q-layer.5": 880.863, "encoder_q-layer.6": 891.4373, "encoder_q-layer.7": 850.4754, "encoder_q-layer.8": 659.1393, "encoder_q-layer.9": 498.5835, "epoch": 0.07, "inbatch_neg_score": 0.2136, "inbatch_pos_score": 0.7236, "learning_rate": 4.8447368421052637e-05, "loss": 4.1228, "norm_diff": 0.0194, "norm_loss": 0.0, "num_token_doc": 66.5643, "num_token_overlap": 11.6459, "num_token_query": 31.38, "num_token_union": 65.0208, "num_word_context": 202.1226, "num_word_doc": 49.6952, "num_word_query": 23.3242, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1330.1349, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2131, "query_norm": 1.3697, "queue_k_norm": 1.3917, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.38, "sent_len_1": 66.5643, "sent_len_max_0": 127.9613, "sent_len_max_1": 207.1387, "stdk": 0.0461, "stdq": 0.0418, "stdqueue_k": 0.0462, "stdqueue_q": 0.0, "step": 15900 }, { "accuracy": 42.334, "active_queue_size": 16384.0, "cl_loss": 4.0997, "doc_norm": 1.3894, "encoder_q-embeddings": 739.1585, "encoder_q-layer.0": 507.4954, "encoder_q-layer.1": 544.4019, "encoder_q-layer.10": 582.1752, "encoder_q-layer.11": 1568.6562, "encoder_q-layer.2": 585.7203, "encoder_q-layer.3": 627.5257, "encoder_q-layer.4": 666.2478, "encoder_q-layer.5": 595.564, "encoder_q-layer.6": 644.3365, "encoder_q-layer.7": 707.5834, "encoder_q-layer.8": 666.4623, "encoder_q-layer.9": 592.508, "epoch": 0.07, "inbatch_neg_score": 0.1976, "inbatch_pos_score": 0.7104, "learning_rate": 4.842105263157895e-05, "loss": 4.0997, "norm_diff": 0.0162, "norm_loss": 0.0, "num_token_doc": 66.8831, "num_token_overlap": 11.7207, "num_token_query": 31.3186, "num_token_union": 65.0921, "num_word_context": 201.9296, "num_word_doc": 49.8899, "num_word_query": 23.2586, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1117.5631, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1975, "query_norm": 1.3732, "queue_k_norm": 1.3905, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3186, "sent_len_1": 66.8831, "sent_len_max_0": 128.0, "sent_len_max_1": 209.4538, "stdk": 0.0462, "stdq": 0.0426, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 16000 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 4.0868, "doc_norm": 1.3852, "encoder_q-embeddings": 2031.9226, "encoder_q-layer.0": 1492.9504, "encoder_q-layer.1": 1591.5781, "encoder_q-layer.10": 568.7751, "encoder_q-layer.11": 1545.9862, "encoder_q-layer.2": 1851.3792, "encoder_q-layer.3": 1851.1897, "encoder_q-layer.4": 1810.366, "encoder_q-layer.5": 1807.6298, "encoder_q-layer.6": 1906.2694, "encoder_q-layer.7": 1633.8499, "encoder_q-layer.8": 1079.5822, "encoder_q-layer.9": 655.3618, "epoch": 0.07, "inbatch_neg_score": 0.1895, "inbatch_pos_score": 0.7271, "learning_rate": 4.8394736842105266e-05, "loss": 4.0868, "norm_diff": 0.019, "norm_loss": 0.0, "num_token_doc": 66.7508, "num_token_overlap": 11.7228, "num_token_query": 31.4316, "num_token_union": 65.1199, "num_word_context": 202.3023, "num_word_doc": 49.7986, "num_word_query": 23.3591, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2387.762, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1907, "query_norm": 1.4012, "queue_k_norm": 1.3848, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4316, "sent_len_1": 66.7508, "sent_len_max_0": 127.98, "sent_len_max_1": 209.6325, "stdk": 0.0462, "stdq": 0.0438, "stdqueue_k": 0.0462, "stdqueue_q": 0.0, "step": 16100 }, { "accuracy": 42.627, "active_queue_size": 16384.0, "cl_loss": 4.1013, "doc_norm": 1.3852, "encoder_q-embeddings": 1160.4325, "encoder_q-layer.0": 746.697, "encoder_q-layer.1": 859.9602, "encoder_q-layer.10": 584.5869, "encoder_q-layer.11": 1526.8136, "encoder_q-layer.2": 949.4326, "encoder_q-layer.3": 983.3744, "encoder_q-layer.4": 996.3372, "encoder_q-layer.5": 874.6693, "encoder_q-layer.6": 908.3744, "encoder_q-layer.7": 811.131, "encoder_q-layer.8": 722.5118, "encoder_q-layer.9": 533.0117, "epoch": 0.07, "inbatch_neg_score": 0.1949, "inbatch_pos_score": 0.7124, "learning_rate": 4.836842105263158e-05, "loss": 4.1013, "norm_diff": 0.0267, "norm_loss": 0.0, "num_token_doc": 66.7898, "num_token_overlap": 11.7175, "num_token_query": 31.4265, "num_token_union": 65.1348, "num_word_context": 202.5384, "num_word_doc": 49.8396, "num_word_query": 23.3414, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1429.2515, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.194, "query_norm": 1.3585, "queue_k_norm": 1.3844, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4265, "sent_len_1": 66.7898, "sent_len_max_0": 127.935, "sent_len_max_1": 209.4462, "stdk": 0.0463, "stdq": 0.0425, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 16200 }, { "accuracy": 43.5059, "active_queue_size": 16384.0, "cl_loss": 4.1011, "doc_norm": 1.3814, "encoder_q-embeddings": 7518.8188, "encoder_q-layer.0": 4997.7251, "encoder_q-layer.1": 5539.7554, "encoder_q-layer.10": 593.5551, "encoder_q-layer.11": 1538.3535, "encoder_q-layer.2": 6347.29, "encoder_q-layer.3": 6495.5596, "encoder_q-layer.4": 6568.8784, "encoder_q-layer.5": 6266.8799, "encoder_q-layer.6": 6362.6465, "encoder_q-layer.7": 5462.7847, "encoder_q-layer.8": 2557.4641, "encoder_q-layer.9": 698.6951, "epoch": 0.07, "inbatch_neg_score": 0.1966, "inbatch_pos_score": 0.7266, "learning_rate": 4.8342105263157896e-05, "loss": 4.1011, "norm_diff": 0.0272, "norm_loss": 0.0, "num_token_doc": 66.6657, "num_token_overlap": 11.6508, "num_token_query": 31.278, "num_token_union": 65.0203, "num_word_context": 202.3926, "num_word_doc": 49.7353, "num_word_query": 23.2244, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8288.1743, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.196, "query_norm": 1.3542, "queue_k_norm": 1.3803, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.278, "sent_len_1": 66.6657, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.7612, "stdk": 0.0463, "stdq": 0.0428, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 16300 }, { "accuracy": 42.4316, "active_queue_size": 16384.0, "cl_loss": 4.107, "doc_norm": 1.3726, "encoder_q-embeddings": 635.5885, "encoder_q-layer.0": 423.3734, "encoder_q-layer.1": 469.5789, "encoder_q-layer.10": 590.2613, "encoder_q-layer.11": 1562.4863, "encoder_q-layer.2": 540.7576, "encoder_q-layer.3": 578.0402, "encoder_q-layer.4": 624.5166, "encoder_q-layer.5": 606.4855, "encoder_q-layer.6": 631.2817, "encoder_q-layer.7": 637.8359, "encoder_q-layer.8": 598.8554, "encoder_q-layer.9": 517.3447, "epoch": 0.07, "inbatch_neg_score": 0.1867, "inbatch_pos_score": 0.7178, "learning_rate": 4.8315789473684215e-05, "loss": 4.107, "norm_diff": 0.0079, "norm_loss": 0.0, "num_token_doc": 66.8181, "num_token_overlap": 11.6754, "num_token_query": 31.3033, "num_token_union": 65.0738, "num_word_context": 202.4407, "num_word_doc": 49.8509, "num_word_query": 23.2411, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1064.5673, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1853, "query_norm": 1.3651, "queue_k_norm": 1.3764, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3033, "sent_len_1": 66.8181, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.4062, "stdk": 0.046, "stdq": 0.0435, "stdqueue_k": 0.0462, "stdqueue_q": 0.0, "step": 16400 }, { "accuracy": 43.2617, "active_queue_size": 16384.0, "cl_loss": 4.1208, "doc_norm": 1.3755, "encoder_q-embeddings": 1689.6238, "encoder_q-layer.0": 1181.2297, "encoder_q-layer.1": 1431.5458, "encoder_q-layer.10": 542.687, "encoder_q-layer.11": 1552.0629, "encoder_q-layer.2": 1611.9762, "encoder_q-layer.3": 1652.5367, "encoder_q-layer.4": 1643.1299, "encoder_q-layer.5": 1668.0305, "encoder_q-layer.6": 1737.3439, "encoder_q-layer.7": 1574.5264, "encoder_q-layer.8": 1004.9988, "encoder_q-layer.9": 530.3363, "epoch": 0.07, "inbatch_neg_score": 0.1846, "inbatch_pos_score": 0.7075, "learning_rate": 4.8289473684210526e-05, "loss": 4.1208, "norm_diff": 0.0262, "norm_loss": 0.0, "num_token_doc": 66.7938, "num_token_overlap": 11.6688, "num_token_query": 31.3418, "num_token_union": 65.1069, "num_word_context": 202.3617, "num_word_doc": 49.8326, "num_word_query": 23.2856, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2142.7758, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1844, "query_norm": 1.3493, "queue_k_norm": 1.3777, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3418, "sent_len_1": 66.7938, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.0337, "stdk": 0.0462, "stdq": 0.0427, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 16500 }, { "accuracy": 42.0898, "active_queue_size": 16384.0, "cl_loss": 4.0887, "doc_norm": 1.3722, "encoder_q-embeddings": 1390.0585, "encoder_q-layer.0": 966.8552, "encoder_q-layer.1": 1017.8852, "encoder_q-layer.10": 592.6943, "encoder_q-layer.11": 1518.5923, "encoder_q-layer.2": 1093.8477, "encoder_q-layer.3": 1131.2904, "encoder_q-layer.4": 1121.9962, "encoder_q-layer.5": 1039.2012, "encoder_q-layer.6": 1160.1995, "encoder_q-layer.7": 1077.4738, "encoder_q-layer.8": 665.0046, "encoder_q-layer.9": 496.0797, "epoch": 0.07, "inbatch_neg_score": 0.2062, "inbatch_pos_score": 0.7334, "learning_rate": 4.8263157894736845e-05, "loss": 4.0887, "norm_diff": 0.0111, "norm_loss": 0.0, "num_token_doc": 66.7578, "num_token_overlap": 11.71, "num_token_query": 31.4003, "num_token_union": 65.1022, "num_word_context": 201.9768, "num_word_doc": 49.8068, "num_word_query": 23.3117, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1633.1076, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2061, "query_norm": 1.3611, "queue_k_norm": 1.3736, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4003, "sent_len_1": 66.7578, "sent_len_max_0": 127.98, "sent_len_max_1": 207.305, "stdk": 0.0462, "stdq": 0.0434, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 16600 }, { "accuracy": 42.2852, "active_queue_size": 16384.0, "cl_loss": 4.0856, "doc_norm": 1.3727, "encoder_q-embeddings": 359.2075, "encoder_q-layer.0": 241.0007, "encoder_q-layer.1": 268.4745, "encoder_q-layer.10": 296.2804, "encoder_q-layer.11": 790.449, "encoder_q-layer.2": 306.6953, "encoder_q-layer.3": 328.7669, "encoder_q-layer.4": 339.0804, "encoder_q-layer.5": 333.8298, "encoder_q-layer.6": 320.5016, "encoder_q-layer.7": 334.037, "encoder_q-layer.8": 297.3492, "encoder_q-layer.9": 254.1742, "epoch": 0.07, "inbatch_neg_score": 0.2058, "inbatch_pos_score": 0.7271, "learning_rate": 4.8236842105263156e-05, "loss": 4.0856, "norm_diff": 0.0234, "norm_loss": 0.0, "num_token_doc": 66.8445, "num_token_overlap": 11.6861, "num_token_query": 31.3946, "num_token_union": 65.1711, "num_word_context": 202.6152, "num_word_doc": 49.8731, "num_word_query": 23.3015, "postclip_grad_norm": 1.0, "preclip_grad_norm": 575.0924, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2051, "query_norm": 1.3502, "queue_k_norm": 1.3732, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3946, "sent_len_1": 66.8445, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.3875, "stdk": 0.0463, "stdq": 0.0427, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 16700 }, { "accuracy": 41.7969, "active_queue_size": 16384.0, "cl_loss": 4.0929, "doc_norm": 1.369, "encoder_q-embeddings": 563.4551, "encoder_q-layer.0": 374.1193, "encoder_q-layer.1": 413.6701, "encoder_q-layer.10": 302.6491, "encoder_q-layer.11": 715.7269, "encoder_q-layer.2": 496.4408, "encoder_q-layer.3": 537.5888, "encoder_q-layer.4": 588.02, "encoder_q-layer.5": 562.867, "encoder_q-layer.6": 525.3736, "encoder_q-layer.7": 464.1288, "encoder_q-layer.8": 360.6768, "encoder_q-layer.9": 256.7769, "epoch": 0.07, "inbatch_neg_score": 0.1796, "inbatch_pos_score": 0.7051, "learning_rate": 4.821052631578948e-05, "loss": 4.0929, "norm_diff": 0.0112, "norm_loss": 0.0, "num_token_doc": 66.8131, "num_token_overlap": 11.6886, "num_token_query": 31.3933, "num_token_union": 65.1508, "num_word_context": 202.2975, "num_word_doc": 49.8501, "num_word_query": 23.3067, "postclip_grad_norm": 1.0, "preclip_grad_norm": 730.1328, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1792, "query_norm": 1.3665, "queue_k_norm": 1.3717, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3933, "sent_len_1": 66.8131, "sent_len_max_0": 127.99, "sent_len_max_1": 207.3375, "stdk": 0.0462, "stdq": 0.0435, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 16800 }, { "accuracy": 42.6758, "active_queue_size": 16384.0, "cl_loss": 4.1107, "doc_norm": 1.3706, "encoder_q-embeddings": 695.6586, "encoder_q-layer.0": 503.3016, "encoder_q-layer.1": 551.1547, "encoder_q-layer.10": 279.8119, "encoder_q-layer.11": 718.9792, "encoder_q-layer.2": 638.771, "encoder_q-layer.3": 718.0173, "encoder_q-layer.4": 772.9726, "encoder_q-layer.5": 746.8663, "encoder_q-layer.6": 680.9299, "encoder_q-layer.7": 537.6611, "encoder_q-layer.8": 302.761, "encoder_q-layer.9": 246.8068, "epoch": 0.07, "inbatch_neg_score": 0.2005, "inbatch_pos_score": 0.7271, "learning_rate": 4.818421052631579e-05, "loss": 4.1107, "norm_diff": 0.0201, "norm_loss": 0.0, "num_token_doc": 66.6796, "num_token_overlap": 11.6652, "num_token_query": 31.3849, "num_token_union": 65.0758, "num_word_context": 202.4179, "num_word_doc": 49.7779, "num_word_query": 23.3126, "postclip_grad_norm": 1.0, "preclip_grad_norm": 904.9383, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2002, "query_norm": 1.3512, "queue_k_norm": 1.3704, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3849, "sent_len_1": 66.6796, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.5288, "stdk": 0.0463, "stdq": 0.0428, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 16900 }, { "accuracy": 43.0664, "active_queue_size": 16384.0, "cl_loss": 4.1198, "doc_norm": 1.369, "encoder_q-embeddings": 2960.2234, "encoder_q-layer.0": 2350.9465, "encoder_q-layer.1": 2464.0547, "encoder_q-layer.10": 283.1652, "encoder_q-layer.11": 746.0554, "encoder_q-layer.2": 2814.5029, "encoder_q-layer.3": 2927.2031, "encoder_q-layer.4": 2940.9253, "encoder_q-layer.5": 2630.7454, "encoder_q-layer.6": 2384.4307, "encoder_q-layer.7": 1564.3091, "encoder_q-layer.8": 946.2562, "encoder_q-layer.9": 317.8517, "epoch": 0.07, "inbatch_neg_score": 0.2224, "inbatch_pos_score": 0.7437, "learning_rate": 4.8157894736842105e-05, "loss": 4.1198, "norm_diff": 0.0143, "norm_loss": 0.0, "num_token_doc": 66.8578, "num_token_overlap": 11.7037, "num_token_query": 31.4556, "num_token_union": 65.1998, "num_word_context": 202.4578, "num_word_doc": 49.8902, "num_word_query": 23.3845, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3246.4187, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2233, "query_norm": 1.3548, "queue_k_norm": 1.369, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4556, "sent_len_1": 66.8578, "sent_len_max_0": 127.985, "sent_len_max_1": 207.9913, "stdk": 0.0463, "stdq": 0.0427, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 17000 }, { "accuracy": 43.0176, "active_queue_size": 16384.0, "cl_loss": 4.1232, "doc_norm": 1.3669, "encoder_q-embeddings": 687.6707, "encoder_q-layer.0": 483.8665, "encoder_q-layer.1": 558.4407, "encoder_q-layer.10": 267.4439, "encoder_q-layer.11": 697.7075, "encoder_q-layer.2": 641.2332, "encoder_q-layer.3": 652.851, "encoder_q-layer.4": 683.3694, "encoder_q-layer.5": 637.4586, "encoder_q-layer.6": 588.2197, "encoder_q-layer.7": 413.5982, "encoder_q-layer.8": 299.5232, "encoder_q-layer.9": 239.1088, "epoch": 0.07, "inbatch_neg_score": 0.2173, "inbatch_pos_score": 0.7529, "learning_rate": 4.813157894736842e-05, "loss": 4.1232, "norm_diff": 0.0129, "norm_loss": 0.0, "num_token_doc": 66.8107, "num_token_overlap": 11.6602, "num_token_query": 31.3451, "num_token_union": 65.1147, "num_word_context": 202.1661, "num_word_doc": 49.8372, "num_word_query": 23.2842, "postclip_grad_norm": 1.0, "preclip_grad_norm": 844.2234, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2175, "query_norm": 1.3541, "queue_k_norm": 1.3661, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3451, "sent_len_1": 66.8107, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.0637, "stdk": 0.0463, "stdq": 0.043, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 17100 }, { "accuracy": 42.4805, "active_queue_size": 16384.0, "cl_loss": 4.1205, "doc_norm": 1.3679, "encoder_q-embeddings": 1355.6279, "encoder_q-layer.0": 991.4022, "encoder_q-layer.1": 1071.6667, "encoder_q-layer.10": 321.6118, "encoder_q-layer.11": 811.043, "encoder_q-layer.2": 1172.0039, "encoder_q-layer.3": 1099.9167, "encoder_q-layer.4": 1118.9777, "encoder_q-layer.5": 1047.5842, "encoder_q-layer.6": 1031.4893, "encoder_q-layer.7": 772.9918, "encoder_q-layer.8": 394.0213, "encoder_q-layer.9": 302.0095, "epoch": 0.07, "inbatch_neg_score": 0.204, "inbatch_pos_score": 0.7173, "learning_rate": 4.8105263157894735e-05, "loss": 4.1205, "norm_diff": 0.0398, "norm_loss": 0.0, "num_token_doc": 66.8225, "num_token_overlap": 11.6494, "num_token_query": 31.4007, "num_token_union": 65.1825, "num_word_context": 202.7061, "num_word_doc": 49.9084, "num_word_query": 23.329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1440.2828, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2035, "query_norm": 1.3282, "queue_k_norm": 1.3674, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4007, "sent_len_1": 66.8225, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.2088, "stdk": 0.0463, "stdq": 0.042, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 17200 }, { "accuracy": 40.8203, "active_queue_size": 16384.0, "cl_loss": 4.1223, "doc_norm": 1.3684, "encoder_q-embeddings": 523.1234, "encoder_q-layer.0": 349.1053, "encoder_q-layer.1": 405.399, "encoder_q-layer.10": 289.8599, "encoder_q-layer.11": 774.669, "encoder_q-layer.2": 466.4166, "encoder_q-layer.3": 476.2384, "encoder_q-layer.4": 472.0635, "encoder_q-layer.5": 457.2511, "encoder_q-layer.6": 431.1833, "encoder_q-layer.7": 431.9869, "encoder_q-layer.8": 348.7939, "encoder_q-layer.9": 257.2593, "epoch": 0.08, "inbatch_neg_score": 0.2153, "inbatch_pos_score": 0.7383, "learning_rate": 4.807894736842106e-05, "loss": 4.1223, "norm_diff": 0.0082, "norm_loss": 0.0, "num_token_doc": 66.7592, "num_token_overlap": 11.6682, "num_token_query": 31.3062, "num_token_union": 65.1112, "num_word_context": 202.6159, "num_word_doc": 49.8348, "num_word_query": 23.2437, "postclip_grad_norm": 1.0, "preclip_grad_norm": 689.7586, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2163, "query_norm": 1.3624, "queue_k_norm": 1.3665, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3062, "sent_len_1": 66.7592, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.3363, "stdk": 0.0464, "stdq": 0.0434, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 17300 }, { "accuracy": 45.7031, "active_queue_size": 16384.0, "cl_loss": 4.1086, "doc_norm": 1.3641, "encoder_q-embeddings": 522.8871, "encoder_q-layer.0": 364.6477, "encoder_q-layer.1": 407.0751, "encoder_q-layer.10": 274.0203, "encoder_q-layer.11": 772.9064, "encoder_q-layer.2": 474.8488, "encoder_q-layer.3": 487.3918, "encoder_q-layer.4": 462.2295, "encoder_q-layer.5": 480.9181, "encoder_q-layer.6": 462.6602, "encoder_q-layer.7": 391.1481, "encoder_q-layer.8": 315.7628, "encoder_q-layer.9": 242.282, "epoch": 0.08, "inbatch_neg_score": 0.225, "inbatch_pos_score": 0.7676, "learning_rate": 4.805263157894737e-05, "loss": 4.1086, "norm_diff": 0.0117, "norm_loss": 0.0, "num_token_doc": 66.7313, "num_token_overlap": 11.6771, "num_token_query": 31.3935, "num_token_union": 65.1052, "num_word_context": 202.2515, "num_word_doc": 49.7924, "num_word_query": 23.3009, "postclip_grad_norm": 1.0, "preclip_grad_norm": 693.1352, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.225, "query_norm": 1.3524, "queue_k_norm": 1.3661, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3935, "sent_len_1": 66.7313, "sent_len_max_0": 127.985, "sent_len_max_1": 209.8988, "stdk": 0.0462, "stdq": 0.0431, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 17400 }, { "accuracy": 44.2383, "active_queue_size": 16384.0, "cl_loss": 4.1011, "doc_norm": 1.3688, "encoder_q-embeddings": 8213.8125, "encoder_q-layer.0": 7294.4707, "encoder_q-layer.1": 6239.9556, "encoder_q-layer.10": 292.2781, "encoder_q-layer.11": 764.5257, "encoder_q-layer.2": 7046.5029, "encoder_q-layer.3": 8049.0874, "encoder_q-layer.4": 7185.9067, "encoder_q-layer.5": 6437.2031, "encoder_q-layer.6": 6640.6006, "encoder_q-layer.7": 5778.8755, "encoder_q-layer.8": 2566.5801, "encoder_q-layer.9": 404.0981, "epoch": 0.08, "inbatch_neg_score": 0.22, "inbatch_pos_score": 0.7603, "learning_rate": 4.802631578947368e-05, "loss": 4.1011, "norm_diff": 0.0213, "norm_loss": 0.0, "num_token_doc": 66.7799, "num_token_overlap": 11.6871, "num_token_query": 31.4574, "num_token_union": 65.1484, "num_word_context": 202.2299, "num_word_doc": 49.8303, "num_word_query": 23.3919, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9119.6172, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2201, "query_norm": 1.3475, "queue_k_norm": 1.3668, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4574, "sent_len_1": 66.7799, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.8663, "stdk": 0.0464, "stdq": 0.0431, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 17500 }, { "accuracy": 42.9199, "active_queue_size": 16384.0, "cl_loss": 4.0997, "doc_norm": 1.3643, "encoder_q-embeddings": 1066.7844, "encoder_q-layer.0": 778.5038, "encoder_q-layer.1": 904.8112, "encoder_q-layer.10": 276.4572, "encoder_q-layer.11": 745.0288, "encoder_q-layer.2": 1055.0813, "encoder_q-layer.3": 1139.7809, "encoder_q-layer.4": 1274.2365, "encoder_q-layer.5": 1360.3556, "encoder_q-layer.6": 1342.3271, "encoder_q-layer.7": 960.5065, "encoder_q-layer.8": 422.5908, "encoder_q-layer.9": 261.5349, "epoch": 0.08, "inbatch_neg_score": 0.2169, "inbatch_pos_score": 0.7383, "learning_rate": 4.8e-05, "loss": 4.0997, "norm_diff": 0.017, "norm_loss": 0.0, "num_token_doc": 66.869, "num_token_overlap": 11.6676, "num_token_query": 31.3119, "num_token_union": 65.1571, "num_word_context": 202.3544, "num_word_doc": 49.8702, "num_word_query": 23.2542, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1409.2957, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2167, "query_norm": 1.3478, "queue_k_norm": 1.3674, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3119, "sent_len_1": 66.869, "sent_len_max_0": 127.97, "sent_len_max_1": 208.475, "stdk": 0.0462, "stdq": 0.0431, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 17600 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 4.0971, "doc_norm": 1.3661, "encoder_q-embeddings": 1233.3773, "encoder_q-layer.0": 844.8933, "encoder_q-layer.1": 921.981, "encoder_q-layer.10": 287.666, "encoder_q-layer.11": 803.8293, "encoder_q-layer.2": 1083.5156, "encoder_q-layer.3": 1123.5721, "encoder_q-layer.4": 1137.6624, "encoder_q-layer.5": 1198.1086, "encoder_q-layer.6": 1227.8879, "encoder_q-layer.7": 1076.0505, "encoder_q-layer.8": 597.3989, "encoder_q-layer.9": 314.3547, "epoch": 0.08, "inbatch_neg_score": 0.2161, "inbatch_pos_score": 0.7588, "learning_rate": 4.797368421052632e-05, "loss": 4.0971, "norm_diff": 0.0097, "norm_loss": 0.0, "num_token_doc": 66.7464, "num_token_overlap": 11.638, "num_token_query": 31.2706, "num_token_union": 65.066, "num_word_context": 202.2715, "num_word_doc": 49.7809, "num_word_query": 23.2138, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1450.1066, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2159, "query_norm": 1.3567, "queue_k_norm": 1.3664, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2706, "sent_len_1": 66.7464, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.2088, "stdk": 0.0463, "stdq": 0.0434, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 17700 }, { "accuracy": 43.8477, "active_queue_size": 16384.0, "cl_loss": 4.0885, "doc_norm": 1.368, "encoder_q-embeddings": 1955.2086, "encoder_q-layer.0": 1413.9451, "encoder_q-layer.1": 1560.3097, "encoder_q-layer.10": 288.2582, "encoder_q-layer.11": 687.5544, "encoder_q-layer.2": 1961.9438, "encoder_q-layer.3": 1879.2795, "encoder_q-layer.4": 1956.2955, "encoder_q-layer.5": 1661.9019, "encoder_q-layer.6": 1655.0796, "encoder_q-layer.7": 1367.4429, "encoder_q-layer.8": 936.5543, "encoder_q-layer.9": 644.5466, "epoch": 0.08, "inbatch_neg_score": 0.219, "inbatch_pos_score": 0.7466, "learning_rate": 4.794736842105264e-05, "loss": 4.0885, "norm_diff": 0.0293, "norm_loss": 0.0, "num_token_doc": 66.8271, "num_token_overlap": 11.6959, "num_token_query": 31.3798, "num_token_union": 65.1134, "num_word_context": 202.5805, "num_word_doc": 49.8438, "num_word_query": 23.3056, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2296.6048, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2189, "query_norm": 1.3386, "queue_k_norm": 1.3641, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3798, "sent_len_1": 66.8271, "sent_len_max_0": 127.99, "sent_len_max_1": 210.9387, "stdk": 0.0464, "stdq": 0.0426, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 17800 }, { "accuracy": 41.6504, "active_queue_size": 16384.0, "cl_loss": 4.1096, "doc_norm": 1.3664, "encoder_q-embeddings": 1569.7783, "encoder_q-layer.0": 1270.0958, "encoder_q-layer.1": 1219.7058, "encoder_q-layer.10": 300.8496, "encoder_q-layer.11": 837.1967, "encoder_q-layer.2": 1506.0884, "encoder_q-layer.3": 1470.2251, "encoder_q-layer.4": 1475.2758, "encoder_q-layer.5": 1412.2354, "encoder_q-layer.6": 1404.8743, "encoder_q-layer.7": 1443.9918, "encoder_q-layer.8": 1214.8992, "encoder_q-layer.9": 532.2303, "epoch": 0.08, "inbatch_neg_score": 0.2325, "inbatch_pos_score": 0.748, "learning_rate": 4.792105263157895e-05, "loss": 4.1096, "norm_diff": 0.0208, "norm_loss": 0.0, "num_token_doc": 66.8619, "num_token_overlap": 11.6744, "num_token_query": 31.3762, "num_token_union": 65.176, "num_word_context": 202.4028, "num_word_doc": 49.8669, "num_word_query": 23.3032, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1922.5541, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2318, "query_norm": 1.3455, "queue_k_norm": 1.3657, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3762, "sent_len_1": 66.8619, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.61, "stdk": 0.0463, "stdq": 0.0425, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 17900 }, { "accuracy": 42.2852, "active_queue_size": 16384.0, "cl_loss": 4.0747, "doc_norm": 1.365, "encoder_q-embeddings": 578.6993, "encoder_q-layer.0": 408.5897, "encoder_q-layer.1": 453.808, "encoder_q-layer.10": 294.3984, "encoder_q-layer.11": 764.4114, "encoder_q-layer.2": 535.0958, "encoder_q-layer.3": 555.8932, "encoder_q-layer.4": 568.8792, "encoder_q-layer.5": 599.5187, "encoder_q-layer.6": 568.1118, "encoder_q-layer.7": 513.854, "encoder_q-layer.8": 391.6376, "encoder_q-layer.9": 296.3904, "epoch": 0.08, "inbatch_neg_score": 0.2341, "inbatch_pos_score": 0.7603, "learning_rate": 4.789473684210526e-05, "loss": 4.0747, "norm_diff": 0.0175, "norm_loss": 0.0, "num_token_doc": 66.7977, "num_token_overlap": 11.6834, "num_token_query": 31.4131, "num_token_union": 65.1228, "num_word_context": 202.2286, "num_word_doc": 49.8009, "num_word_query": 23.3195, "postclip_grad_norm": 1.0, "preclip_grad_norm": 771.5463, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2336, "query_norm": 1.3622, "queue_k_norm": 1.3679, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4131, "sent_len_1": 66.7977, "sent_len_max_0": 127.97, "sent_len_max_1": 208.9787, "stdk": 0.0462, "stdq": 0.0434, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18000 }, { "accuracy": 43.6035, "active_queue_size": 16384.0, "cl_loss": 4.0946, "doc_norm": 1.3698, "encoder_q-embeddings": 307.44, "encoder_q-layer.0": 215.55, "encoder_q-layer.1": 231.6579, "encoder_q-layer.10": 290.9773, "encoder_q-layer.11": 693.5764, "encoder_q-layer.2": 267.063, "encoder_q-layer.3": 286.3109, "encoder_q-layer.4": 307.468, "encoder_q-layer.5": 294.765, "encoder_q-layer.6": 300.0302, "encoder_q-layer.7": 289.7333, "encoder_q-layer.8": 284.0622, "encoder_q-layer.9": 245.9699, "epoch": 0.08, "inbatch_neg_score": 0.2298, "inbatch_pos_score": 0.7461, "learning_rate": 4.786842105263158e-05, "loss": 4.0946, "norm_diff": 0.0523, "norm_loss": 0.0, "num_token_doc": 66.8267, "num_token_overlap": 11.6763, "num_token_query": 31.4167, "num_token_union": 65.2017, "num_word_context": 202.5853, "num_word_doc": 49.8944, "num_word_query": 23.3385, "postclip_grad_norm": 1.0, "preclip_grad_norm": 498.5096, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2295, "query_norm": 1.3175, "queue_k_norm": 1.3677, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4167, "sent_len_1": 66.8267, "sent_len_max_0": 127.97, "sent_len_max_1": 206.7937, "stdk": 0.0464, "stdq": 0.0417, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18100 }, { "accuracy": 43.8965, "active_queue_size": 16384.0, "cl_loss": 4.0772, "doc_norm": 1.3676, "encoder_q-embeddings": 1374.6372, "encoder_q-layer.0": 905.1879, "encoder_q-layer.1": 902.3215, "encoder_q-layer.10": 264.2696, "encoder_q-layer.11": 669.3331, "encoder_q-layer.2": 1016.7043, "encoder_q-layer.3": 1052.8599, "encoder_q-layer.4": 1124.6023, "encoder_q-layer.5": 1381.4692, "encoder_q-layer.6": 1375.6771, "encoder_q-layer.7": 1134.9229, "encoder_q-layer.8": 669.1782, "encoder_q-layer.9": 321.0709, "epoch": 0.08, "inbatch_neg_score": 0.2296, "inbatch_pos_score": 0.7554, "learning_rate": 4.78421052631579e-05, "loss": 4.0772, "norm_diff": 0.0196, "norm_loss": 0.0, "num_token_doc": 66.8407, "num_token_overlap": 11.6943, "num_token_query": 31.3837, "num_token_union": 65.1501, "num_word_context": 202.3335, "num_word_doc": 49.871, "num_word_query": 23.3062, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1519.316, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2295, "query_norm": 1.348, "queue_k_norm": 1.3667, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3837, "sent_len_1": 66.8407, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.2763, "stdk": 0.0463, "stdq": 0.0429, "stdqueue_k": 0.0463, "stdqueue_q": 0.0, "step": 18200 }, { "accuracy": 43.2129, "active_queue_size": 16384.0, "cl_loss": 4.0837, "doc_norm": 1.3685, "encoder_q-embeddings": 904.8439, "encoder_q-layer.0": 703.1746, "encoder_q-layer.1": 767.6199, "encoder_q-layer.10": 316.4326, "encoder_q-layer.11": 773.3074, "encoder_q-layer.2": 943.8508, "encoder_q-layer.3": 947.8101, "encoder_q-layer.4": 1038.4276, "encoder_q-layer.5": 882.2514, "encoder_q-layer.6": 945.7324, "encoder_q-layer.7": 886.3559, "encoder_q-layer.8": 516.9452, "encoder_q-layer.9": 303.6912, "epoch": 0.08, "inbatch_neg_score": 0.2243, "inbatch_pos_score": 0.7544, "learning_rate": 4.7815789473684216e-05, "loss": 4.0837, "norm_diff": 0.0119, "norm_loss": 0.0, "num_token_doc": 66.8648, "num_token_overlap": 11.6595, "num_token_query": 31.3458, "num_token_union": 65.1606, "num_word_context": 202.3546, "num_word_doc": 49.871, "num_word_query": 23.2753, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1207.5128, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2253, "query_norm": 1.3626, "queue_k_norm": 1.3674, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3458, "sent_len_1": 66.8648, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.5462, "stdk": 0.0464, "stdq": 0.0432, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18300 }, { "accuracy": 44.3848, "active_queue_size": 16384.0, "cl_loss": 4.0842, "doc_norm": 1.3656, "encoder_q-embeddings": 597.0292, "encoder_q-layer.0": 433.1633, "encoder_q-layer.1": 470.1239, "encoder_q-layer.10": 266.1712, "encoder_q-layer.11": 679.8602, "encoder_q-layer.2": 510.8625, "encoder_q-layer.3": 568.7863, "encoder_q-layer.4": 523.7233, "encoder_q-layer.5": 491.5171, "encoder_q-layer.6": 593.1503, "encoder_q-layer.7": 516.516, "encoder_q-layer.8": 287.7396, "encoder_q-layer.9": 232.0471, "epoch": 0.08, "inbatch_neg_score": 0.2262, "inbatch_pos_score": 0.7617, "learning_rate": 4.778947368421053e-05, "loss": 4.0842, "norm_diff": 0.0126, "norm_loss": 0.0, "num_token_doc": 66.7924, "num_token_overlap": 11.7192, "num_token_query": 31.4983, "num_token_union": 65.1855, "num_word_context": 202.4446, "num_word_doc": 49.8554, "num_word_query": 23.3997, "postclip_grad_norm": 1.0, "preclip_grad_norm": 753.2362, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2263, "query_norm": 1.3575, "queue_k_norm": 1.3678, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4983, "sent_len_1": 66.7924, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.4, "stdk": 0.0463, "stdq": 0.0429, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18400 }, { "accuracy": 42.1387, "active_queue_size": 16384.0, "cl_loss": 4.0603, "doc_norm": 1.3667, "encoder_q-embeddings": 548.2544, "encoder_q-layer.0": 387.5377, "encoder_q-layer.1": 442.7375, "encoder_q-layer.10": 269.1984, "encoder_q-layer.11": 679.2957, "encoder_q-layer.2": 524.4604, "encoder_q-layer.3": 546.5949, "encoder_q-layer.4": 544.2615, "encoder_q-layer.5": 507.4744, "encoder_q-layer.6": 462.6424, "encoder_q-layer.7": 469.1997, "encoder_q-layer.8": 363.3138, "encoder_q-layer.9": 237.797, "epoch": 0.08, "inbatch_neg_score": 0.2326, "inbatch_pos_score": 0.748, "learning_rate": 4.7763157894736846e-05, "loss": 4.0603, "norm_diff": 0.013, "norm_loss": 0.0, "num_token_doc": 66.8988, "num_token_overlap": 11.7123, "num_token_query": 31.4889, "num_token_union": 65.2073, "num_word_context": 202.3582, "num_word_doc": 49.9324, "num_word_query": 23.3947, "postclip_grad_norm": 1.0, "preclip_grad_norm": 723.3427, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2314, "query_norm": 1.3556, "queue_k_norm": 1.3717, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4889, "sent_len_1": 66.8988, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.3013, "stdk": 0.0463, "stdq": 0.0425, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 18500 }, { "accuracy": 44.5801, "active_queue_size": 16384.0, "cl_loss": 4.0756, "doc_norm": 1.3697, "encoder_q-embeddings": 319.0392, "encoder_q-layer.0": 201.2096, "encoder_q-layer.1": 219.8392, "encoder_q-layer.10": 303.0598, "encoder_q-layer.11": 723.8431, "encoder_q-layer.2": 244.393, "encoder_q-layer.3": 246.6458, "encoder_q-layer.4": 275.7031, "encoder_q-layer.5": 243.7769, "encoder_q-layer.6": 259.7458, "encoder_q-layer.7": 261.4296, "encoder_q-layer.8": 285.1097, "encoder_q-layer.9": 249.3619, "epoch": 0.08, "inbatch_neg_score": 0.2251, "inbatch_pos_score": 0.7598, "learning_rate": 4.773684210526316e-05, "loss": 4.0756, "norm_diff": 0.0116, "norm_loss": 0.0, "num_token_doc": 66.8988, "num_token_overlap": 11.6765, "num_token_query": 31.3633, "num_token_union": 65.1807, "num_word_context": 202.4197, "num_word_doc": 49.95, "num_word_query": 23.2769, "postclip_grad_norm": 1.0, "preclip_grad_norm": 487.2492, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2249, "query_norm": 1.3687, "queue_k_norm": 1.3679, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3633, "sent_len_1": 66.8988, "sent_len_max_0": 127.9688, "sent_len_max_1": 208.82, "stdk": 0.0464, "stdq": 0.0431, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18600 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 4.0843, "doc_norm": 1.3674, "encoder_q-embeddings": 10607.7354, "encoder_q-layer.0": 7897.7427, "encoder_q-layer.1": 6520.5933, "encoder_q-layer.10": 537.5276, "encoder_q-layer.11": 1342.3582, "encoder_q-layer.2": 6604.9736, "encoder_q-layer.3": 6318.8584, "encoder_q-layer.4": 6081.8174, "encoder_q-layer.5": 5504.8789, "encoder_q-layer.6": 5066.3345, "encoder_q-layer.7": 5008.3794, "encoder_q-layer.8": 2244.7166, "encoder_q-layer.9": 647.2902, "epoch": 0.08, "inbatch_neg_score": 0.2305, "inbatch_pos_score": 0.7627, "learning_rate": 4.7710526315789476e-05, "loss": 4.0843, "norm_diff": 0.0144, "norm_loss": 0.0, "num_token_doc": 66.7739, "num_token_overlap": 11.651, "num_token_query": 31.2668, "num_token_union": 65.0872, "num_word_context": 202.2477, "num_word_doc": 49.8459, "num_word_query": 23.2233, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9140.1911, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2296, "query_norm": 1.3738, "queue_k_norm": 1.3684, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2668, "sent_len_1": 66.7739, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.4787, "stdk": 0.0463, "stdq": 0.0433, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18700 }, { "accuracy": 44.1406, "active_queue_size": 16384.0, "cl_loss": 4.075, "doc_norm": 1.367, "encoder_q-embeddings": 7583.8232, "encoder_q-layer.0": 5532.4692, "encoder_q-layer.1": 5569.373, "encoder_q-layer.10": 496.3929, "encoder_q-layer.11": 1338.5858, "encoder_q-layer.2": 5500.0444, "encoder_q-layer.3": 5641.6899, "encoder_q-layer.4": 5210.8989, "encoder_q-layer.5": 5168.4644, "encoder_q-layer.6": 5927.3467, "encoder_q-layer.7": 5039.3496, "encoder_q-layer.8": 1698.0198, "encoder_q-layer.9": 654.233, "epoch": 0.08, "inbatch_neg_score": 0.2297, "inbatch_pos_score": 0.7661, "learning_rate": 4.7684210526315794e-05, "loss": 4.075, "norm_diff": 0.0063, "norm_loss": 0.0, "num_token_doc": 66.6628, "num_token_overlap": 11.657, "num_token_query": 31.4136, "num_token_union": 65.0889, "num_word_context": 202.2141, "num_word_doc": 49.7619, "num_word_query": 23.3404, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7631.4023, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2291, "query_norm": 1.3647, "queue_k_norm": 1.3689, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4136, "sent_len_1": 66.6628, "sent_len_max_0": 128.0, "sent_len_max_1": 208.6225, "stdk": 0.0463, "stdq": 0.0434, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 18800 }, { "accuracy": 44.0918, "active_queue_size": 16384.0, "cl_loss": 4.0836, "doc_norm": 1.3702, "encoder_q-embeddings": 571.5017, "encoder_q-layer.0": 374.3783, "encoder_q-layer.1": 411.1729, "encoder_q-layer.10": 550.9172, "encoder_q-layer.11": 1396.7736, "encoder_q-layer.2": 465.7209, "encoder_q-layer.3": 502.5346, "encoder_q-layer.4": 552.7943, "encoder_q-layer.5": 502.3086, "encoder_q-layer.6": 527.5689, "encoder_q-layer.7": 539.6254, "encoder_q-layer.8": 563.9259, "encoder_q-layer.9": 500.0521, "epoch": 0.08, "inbatch_neg_score": 0.231, "inbatch_pos_score": 0.7539, "learning_rate": 4.7657894736842106e-05, "loss": 4.0836, "norm_diff": 0.0498, "norm_loss": 0.0, "num_token_doc": 66.7441, "num_token_overlap": 11.663, "num_token_query": 31.4308, "num_token_union": 65.1593, "num_word_context": 202.3789, "num_word_doc": 49.8039, "num_word_query": 23.3472, "postclip_grad_norm": 1.0, "preclip_grad_norm": 950.0478, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2312, "query_norm": 1.3204, "queue_k_norm": 1.3699, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4308, "sent_len_1": 66.7441, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.8388, "stdk": 0.0464, "stdq": 0.0419, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 18900 }, { "accuracy": 42.3828, "active_queue_size": 16384.0, "cl_loss": 4.0621, "doc_norm": 1.3662, "encoder_q-embeddings": 648.7047, "encoder_q-layer.0": 417.2681, "encoder_q-layer.1": 439.5641, "encoder_q-layer.10": 604.6113, "encoder_q-layer.11": 1352.55, "encoder_q-layer.2": 518.6839, "encoder_q-layer.3": 536.5275, "encoder_q-layer.4": 545.0141, "encoder_q-layer.5": 544.9903, "encoder_q-layer.6": 565.8282, "encoder_q-layer.7": 531.2578, "encoder_q-layer.8": 576.3676, "encoder_q-layer.9": 523.085, "epoch": 0.08, "inbatch_neg_score": 0.2309, "inbatch_pos_score": 0.7534, "learning_rate": 4.7631578947368424e-05, "loss": 4.0621, "norm_diff": 0.0318, "norm_loss": 0.0, "num_token_doc": 66.9509, "num_token_overlap": 11.6859, "num_token_query": 31.3699, "num_token_union": 65.2302, "num_word_context": 202.615, "num_word_doc": 49.9561, "num_word_query": 23.2976, "postclip_grad_norm": 1.0, "preclip_grad_norm": 972.0613, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2299, "query_norm": 1.3345, "queue_k_norm": 1.369, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3699, "sent_len_1": 66.9509, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.4837, "stdk": 0.0463, "stdq": 0.0428, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 19000 }, { "accuracy": 43.7012, "active_queue_size": 16384.0, "cl_loss": 4.0616, "doc_norm": 1.3635, "encoder_q-embeddings": 769.3344, "encoder_q-layer.0": 509.3141, "encoder_q-layer.1": 547.9827, "encoder_q-layer.10": 496.5003, "encoder_q-layer.11": 1375.8452, "encoder_q-layer.2": 627.2411, "encoder_q-layer.3": 671.9254, "encoder_q-layer.4": 692.3532, "encoder_q-layer.5": 648.303, "encoder_q-layer.6": 711.2276, "encoder_q-layer.7": 683.7892, "encoder_q-layer.8": 634.5733, "encoder_q-layer.9": 486.0373, "epoch": 0.08, "inbatch_neg_score": 0.2291, "inbatch_pos_score": 0.7734, "learning_rate": 4.760526315789474e-05, "loss": 4.0616, "norm_diff": 0.0152, "norm_loss": 0.0, "num_token_doc": 66.8038, "num_token_overlap": 11.6649, "num_token_query": 31.3648, "num_token_union": 65.1508, "num_word_context": 202.2922, "num_word_doc": 49.8395, "num_word_query": 23.281, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1097.356, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2292, "query_norm": 1.351, "queue_k_norm": 1.3688, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3648, "sent_len_1": 66.8038, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.8862, "stdk": 0.0462, "stdq": 0.0433, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 19100 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 4.046, "doc_norm": 1.3698, "encoder_q-embeddings": 842.2598, "encoder_q-layer.0": 576.4596, "encoder_q-layer.1": 655.4059, "encoder_q-layer.10": 533.9494, "encoder_q-layer.11": 1484.386, "encoder_q-layer.2": 646.8159, "encoder_q-layer.3": 626.1403, "encoder_q-layer.4": 616.641, "encoder_q-layer.5": 572.3205, "encoder_q-layer.6": 555.3946, "encoder_q-layer.7": 518.3709, "encoder_q-layer.8": 531.2342, "encoder_q-layer.9": 484.1949, "epoch": 0.08, "inbatch_neg_score": 0.2323, "inbatch_pos_score": 0.7627, "learning_rate": 4.7578947368421054e-05, "loss": 4.046, "norm_diff": 0.0248, "norm_loss": 0.0, "num_token_doc": 66.6943, "num_token_overlap": 11.6694, "num_token_query": 31.3087, "num_token_union": 65.0496, "num_word_context": 202.2532, "num_word_doc": 49.7945, "num_word_query": 23.2527, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1106.4444, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2327, "query_norm": 1.345, "queue_k_norm": 1.3693, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3087, "sent_len_1": 66.6943, "sent_len_max_0": 127.985, "sent_len_max_1": 207.9888, "stdk": 0.0464, "stdq": 0.0429, "stdqueue_k": 0.0464, "stdqueue_q": 0.0, "step": 19200 }, { "accuracy": 44.0918, "active_queue_size": 16384.0, "cl_loss": 4.0645, "doc_norm": 1.3716, "encoder_q-embeddings": 1043.1705, "encoder_q-layer.0": 724.8469, "encoder_q-layer.1": 810.8912, "encoder_q-layer.10": 585.238, "encoder_q-layer.11": 1517.3474, "encoder_q-layer.2": 966.7158, "encoder_q-layer.3": 1016.9283, "encoder_q-layer.4": 1054.9791, "encoder_q-layer.5": 1074.0128, "encoder_q-layer.6": 1213.2228, "encoder_q-layer.7": 1172.826, "encoder_q-layer.8": 957.9542, "encoder_q-layer.9": 536.1222, "epoch": 0.08, "inbatch_neg_score": 0.2233, "inbatch_pos_score": 0.7554, "learning_rate": 4.755263157894737e-05, "loss": 4.0645, "norm_diff": 0.0474, "norm_loss": 0.0, "num_token_doc": 66.6052, "num_token_overlap": 11.6384, "num_token_query": 31.243, "num_token_union": 64.9855, "num_word_context": 202.0554, "num_word_doc": 49.6965, "num_word_query": 23.1901, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1528.7122, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2241, "query_norm": 1.3243, "queue_k_norm": 1.3709, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.243, "sent_len_1": 66.6052, "sent_len_max_0": 127.975, "sent_len_max_1": 208.7562, "stdk": 0.0465, "stdq": 0.0422, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 19300 }, { "accuracy": 42.1875, "active_queue_size": 16384.0, "cl_loss": 4.0771, "doc_norm": 1.3694, "encoder_q-embeddings": 7744.4077, "encoder_q-layer.0": 5478.3418, "encoder_q-layer.1": 5185.5996, "encoder_q-layer.10": 566.1162, "encoder_q-layer.11": 1568.2235, "encoder_q-layer.2": 6584.0469, "encoder_q-layer.3": 5965.9722, "encoder_q-layer.4": 7351.7012, "encoder_q-layer.5": 6125.2915, "encoder_q-layer.6": 6260.0469, "encoder_q-layer.7": 5327.0352, "encoder_q-layer.8": 2027.6882, "encoder_q-layer.9": 603.3134, "epoch": 0.08, "inbatch_neg_score": 0.2397, "inbatch_pos_score": 0.7646, "learning_rate": 4.7526315789473684e-05, "loss": 4.0771, "norm_diff": 0.0179, "norm_loss": 0.0, "num_token_doc": 66.7113, "num_token_overlap": 11.6816, "num_token_query": 31.3728, "num_token_union": 65.0664, "num_word_context": 202.1335, "num_word_doc": 49.7955, "num_word_query": 23.3036, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8206.1622, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2395, "query_norm": 1.3515, "queue_k_norm": 1.3728, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3728, "sent_len_1": 66.7113, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.7562, "stdk": 0.0464, "stdq": 0.0429, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 19400 }, { "accuracy": 42.1387, "active_queue_size": 16384.0, "cl_loss": 4.0609, "doc_norm": 1.3681, "encoder_q-embeddings": 2024.6318, "encoder_q-layer.0": 1478.8903, "encoder_q-layer.1": 1582.9989, "encoder_q-layer.10": 663.114, "encoder_q-layer.11": 1579.7957, "encoder_q-layer.2": 1871.4474, "encoder_q-layer.3": 2111.4844, "encoder_q-layer.4": 2292.156, "encoder_q-layer.5": 2167.5566, "encoder_q-layer.6": 2238.3306, "encoder_q-layer.7": 1855.1538, "encoder_q-layer.8": 1159.3267, "encoder_q-layer.9": 613.0695, "epoch": 0.08, "inbatch_neg_score": 0.2336, "inbatch_pos_score": 0.7568, "learning_rate": 4.75e-05, "loss": 4.0609, "norm_diff": 0.0257, "norm_loss": 0.0, "num_token_doc": 66.8044, "num_token_overlap": 11.7435, "num_token_query": 31.452, "num_token_union": 65.1601, "num_word_context": 202.2804, "num_word_doc": 49.8766, "num_word_query": 23.3775, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2626.3814, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2327, "query_norm": 1.3425, "queue_k_norm": 1.3725, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.452, "sent_len_1": 66.8044, "sent_len_max_0": 127.9688, "sent_len_max_1": 207.595, "stdk": 0.0463, "stdq": 0.0428, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 19500 }, { "accuracy": 44.2383, "active_queue_size": 16384.0, "cl_loss": 4.0605, "doc_norm": 1.3709, "encoder_q-embeddings": 3088.5945, "encoder_q-layer.0": 2414.4553, "encoder_q-layer.1": 2703.8965, "encoder_q-layer.10": 522.3032, "encoder_q-layer.11": 1391.1746, "encoder_q-layer.2": 2776.3977, "encoder_q-layer.3": 3405.4456, "encoder_q-layer.4": 3517.1157, "encoder_q-layer.5": 2771.342, "encoder_q-layer.6": 2728.8379, "encoder_q-layer.7": 2117.7378, "encoder_q-layer.8": 991.5846, "encoder_q-layer.9": 502.7256, "epoch": 0.09, "inbatch_neg_score": 0.2212, "inbatch_pos_score": 0.7622, "learning_rate": 4.747368421052632e-05, "loss": 4.0605, "norm_diff": 0.0281, "norm_loss": 0.0, "num_token_doc": 66.8778, "num_token_overlap": 11.7173, "num_token_query": 31.3975, "num_token_union": 65.1513, "num_word_context": 202.3932, "num_word_doc": 49.9216, "num_word_query": 23.3165, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3723.7861, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2214, "query_norm": 1.3428, "queue_k_norm": 1.3711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3975, "sent_len_1": 66.8778, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.1425, "stdk": 0.0465, "stdq": 0.0432, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 19600 }, { "accuracy": 44.0918, "active_queue_size": 16384.0, "cl_loss": 4.0496, "doc_norm": 1.3747, "encoder_q-embeddings": 4269.0195, "encoder_q-layer.0": 2948.1611, "encoder_q-layer.1": 3394.6997, "encoder_q-layer.10": 522.4406, "encoder_q-layer.11": 1396.4254, "encoder_q-layer.2": 3459.7268, "encoder_q-layer.3": 3200.4453, "encoder_q-layer.4": 2797.4856, "encoder_q-layer.5": 2522.2349, "encoder_q-layer.6": 2276.2773, "encoder_q-layer.7": 1547.4672, "encoder_q-layer.8": 1065.5334, "encoder_q-layer.9": 541.0366, "epoch": 0.09, "inbatch_neg_score": 0.2221, "inbatch_pos_score": 0.7656, "learning_rate": 4.744736842105263e-05, "loss": 4.0496, "norm_diff": 0.0376, "norm_loss": 0.0, "num_token_doc": 66.8456, "num_token_overlap": 11.7212, "num_token_query": 31.4703, "num_token_union": 65.1883, "num_word_context": 202.4195, "num_word_doc": 49.8849, "num_word_query": 23.3654, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4010.9353, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2227, "query_norm": 1.3372, "queue_k_norm": 1.3713, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4703, "sent_len_1": 66.8456, "sent_len_max_0": 127.9663, "sent_len_max_1": 210.5, "stdk": 0.0466, "stdq": 0.043, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 19700 }, { "accuracy": 42.4316, "active_queue_size": 16384.0, "cl_loss": 4.0576, "doc_norm": 1.3723, "encoder_q-embeddings": 806.361, "encoder_q-layer.0": 532.8212, "encoder_q-layer.1": 600.1796, "encoder_q-layer.10": 557.8206, "encoder_q-layer.11": 1610.3848, "encoder_q-layer.2": 701.9592, "encoder_q-layer.3": 758.0491, "encoder_q-layer.4": 753.5715, "encoder_q-layer.5": 733.2882, "encoder_q-layer.6": 761.4733, "encoder_q-layer.7": 752.238, "encoder_q-layer.8": 663.3089, "encoder_q-layer.9": 492.3429, "epoch": 0.09, "inbatch_neg_score": 0.2304, "inbatch_pos_score": 0.75, "learning_rate": 4.742105263157895e-05, "loss": 4.0576, "norm_diff": 0.0504, "norm_loss": 0.0, "num_token_doc": 66.6513, "num_token_overlap": 11.652, "num_token_query": 31.3183, "num_token_union": 65.0327, "num_word_context": 202.2368, "num_word_doc": 49.739, "num_word_query": 23.2547, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1248.3293, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2312, "query_norm": 1.3219, "queue_k_norm": 1.3705, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3183, "sent_len_1": 66.6513, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.5762, "stdk": 0.0465, "stdq": 0.0421, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 19800 }, { "accuracy": 44.6289, "active_queue_size": 16384.0, "cl_loss": 4.0451, "doc_norm": 1.3762, "encoder_q-embeddings": 2075.2502, "encoder_q-layer.0": 1569.3301, "encoder_q-layer.1": 1652.7974, "encoder_q-layer.10": 529.993, "encoder_q-layer.11": 1431.9558, "encoder_q-layer.2": 1857.0902, "encoder_q-layer.3": 2111.1853, "encoder_q-layer.4": 2310.9351, "encoder_q-layer.5": 2183.7512, "encoder_q-layer.6": 2202.0229, "encoder_q-layer.7": 1618.5854, "encoder_q-layer.8": 821.6088, "encoder_q-layer.9": 482.4488, "epoch": 0.09, "inbatch_neg_score": 0.2314, "inbatch_pos_score": 0.77, "learning_rate": 4.739473684210526e-05, "loss": 4.0451, "norm_diff": 0.0415, "norm_loss": 0.0, "num_token_doc": 66.8019, "num_token_overlap": 11.626, "num_token_query": 31.2676, "num_token_union": 65.143, "num_word_context": 202.4532, "num_word_doc": 49.8638, "num_word_query": 23.2153, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2567.7456, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2323, "query_norm": 1.3347, "queue_k_norm": 1.3727, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2676, "sent_len_1": 66.8019, "sent_len_max_0": 127.9862, "sent_len_max_1": 206.955, "stdk": 0.0467, "stdq": 0.0429, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 19900 }, { "accuracy": 43.9453, "active_queue_size": 16384.0, "cl_loss": 4.0488, "doc_norm": 1.3718, "encoder_q-embeddings": 767.9477, "encoder_q-layer.0": 541.313, "encoder_q-layer.1": 577.2363, "encoder_q-layer.10": 555.855, "encoder_q-layer.11": 1479.1096, "encoder_q-layer.2": 594.7791, "encoder_q-layer.3": 644.6556, "encoder_q-layer.4": 645.4051, "encoder_q-layer.5": 651.1931, "encoder_q-layer.6": 741.4681, "encoder_q-layer.7": 629.8629, "encoder_q-layer.8": 564.4584, "encoder_q-layer.9": 478.0804, "epoch": 0.09, "inbatch_neg_score": 0.242, "inbatch_pos_score": 0.7695, "learning_rate": 4.736842105263158e-05, "loss": 4.0488, "norm_diff": 0.0449, "norm_loss": 0.0, "num_token_doc": 66.6054, "num_token_overlap": 11.6567, "num_token_query": 31.3447, "num_token_union": 65.0296, "num_word_context": 202.1924, "num_word_doc": 49.7105, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1139.9897, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2432, "query_norm": 1.3269, "queue_k_norm": 1.3715, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3447, "sent_len_1": 66.6054, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.1763, "stdk": 0.0465, "stdq": 0.0424, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 20000 }, { "dev_runtime": 28.2956, "dev_samples_per_second": 1.131, "dev_steps_per_second": 0.035, "epoch": 0.09, "step": 20000, "test_accuracy": 91.552734375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.4999617636203766, "test_doc_norm": 1.30553138256073, "test_inbatch_neg_score": 0.4570072889328003, "test_inbatch_pos_score": 1.329084038734436, "test_loss": 0.4999617636203766, "test_loss_align": 1.018460750579834, "test_loss_unif": 3.8772196769714355, "test_loss_unif_q@queue": 3.8772194385528564, "test_norm_diff": 0.0889127254486084, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.23076561093330383, "test_query_norm": 1.394444227218628, "test_queue_k_norm": 1.3718971014022827, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04029650613665581, "test_stdq": 0.04139195755124092, "test_stdqueue_k": 0.04652399942278862, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.2956, "dev_samples_per_second": 1.131, "dev_steps_per_second": 0.035, "epoch": 0.09, "eval_beir-arguana_ndcg@10": 0.28266, "eval_beir-arguana_recall@10": 0.48791, "eval_beir-arguana_recall@100": 0.7845, "eval_beir-arguana_recall@20": 0.60313, "eval_beir-avg_ndcg@10": 0.34264075, "eval_beir-avg_recall@10": 0.40821874999999996, "eval_beir-avg_recall@100": 0.5829099999999999, "eval_beir-avg_recall@20": 0.4643495000000001, "eval_beir-cqadupstack_ndcg@10": 0.2245975, "eval_beir-cqadupstack_recall@10": 0.3134075, "eval_beir-cqadupstack_recall@100": 0.54279, "eval_beir-cqadupstack_recall@20": 0.380225, "eval_beir-fiqa_ndcg@10": 0.20748, "eval_beir-fiqa_recall@10": 0.2615, "eval_beir-fiqa_recall@100": 0.50843, "eval_beir-fiqa_recall@20": 0.3249, "eval_beir-nfcorpus_ndcg@10": 0.27587, "eval_beir-nfcorpus_recall@10": 0.13492, "eval_beir-nfcorpus_recall@100": 0.25177, "eval_beir-nfcorpus_recall@20": 0.16111, "eval_beir-nq_ndcg@10": 0.25906, "eval_beir-nq_recall@10": 0.41862, "eval_beir-nq_recall@100": 0.75263, "eval_beir-nq_recall@20": 0.53124, "eval_beir-quora_ndcg@10": 0.74115, "eval_beir-quora_recall@10": 0.85902, "eval_beir-quora_recall@100": 0.96946, "eval_beir-quora_recall@20": 0.90667, "eval_beir-scidocs_ndcg@10": 0.13742, "eval_beir-scidocs_recall@10": 0.14338, "eval_beir-scidocs_recall@100": 0.33547, "eval_beir-scidocs_recall@20": 0.192, "eval_beir-scifact_ndcg@10": 0.60932, "eval_beir-scifact_recall@10": 0.76833, "eval_beir-scifact_recall@100": 0.90489, "eval_beir-scifact_recall@20": 0.82578, "eval_beir-trec-covid_ndcg@10": 0.49779, "eval_beir-trec-covid_recall@10": 0.562, "eval_beir-trec-covid_recall@100": 0.3826, "eval_beir-trec-covid_recall@20": 0.526, "eval_beir-webis-touche2020_ndcg@10": 0.19106, "eval_beir-webis-touche2020_recall@10": 0.1331, "eval_beir-webis-touche2020_recall@100": 0.39656, "eval_beir-webis-touche2020_recall@20": 0.19244, "eval_senteval-avg_sts": 0.7503991228900059, "eval_senteval-sickr_spearman": 0.7159367083950615, "eval_senteval-stsb_spearman": 0.7848615373849502, "step": 20000, "test_accuracy": 91.552734375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.4999617636203766, "test_doc_norm": 1.30553138256073, "test_inbatch_neg_score": 0.4570072889328003, "test_inbatch_pos_score": 1.329084038734436, "test_loss": 0.4999617636203766, "test_loss_align": 1.018460750579834, "test_loss_unif": 3.8772196769714355, "test_loss_unif_q@queue": 3.8772194385528564, "test_norm_diff": 0.0889127254486084, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.23076561093330383, "test_query_norm": 1.394444227218628, "test_queue_k_norm": 1.3718971014022827, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04029650613665581, "test_stdq": 0.04139195755124092, "test_stdqueue_k": 0.04652399942278862, "test_stdqueue_q": 0.0 }, { "accuracy": 42.5781, "active_queue_size": 16384.0, "cl_loss": 4.0485, "doc_norm": 1.37, "encoder_q-embeddings": 1393.6886, "encoder_q-layer.0": 1056.9993, "encoder_q-layer.1": 1119.1616, "encoder_q-layer.10": 500.6381, "encoder_q-layer.11": 1304.6318, "encoder_q-layer.2": 1428.9631, "encoder_q-layer.3": 1544.5984, "encoder_q-layer.4": 1554.864, "encoder_q-layer.5": 1380.9557, "encoder_q-layer.6": 1145.2455, "encoder_q-layer.7": 880.7158, "encoder_q-layer.8": 536.96, "encoder_q-layer.9": 442.3408, "epoch": 0.09, "inbatch_neg_score": 0.2327, "inbatch_pos_score": 0.7617, "learning_rate": 4.73421052631579e-05, "loss": 4.0485, "norm_diff": 0.0314, "norm_loss": 0.0, "num_token_doc": 66.7948, "num_token_overlap": 11.691, "num_token_query": 31.4394, "num_token_union": 65.1939, "num_word_context": 202.4601, "num_word_doc": 49.8528, "num_word_query": 23.3375, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1777.6968, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2318, "query_norm": 1.3385, "queue_k_norm": 1.3736, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4394, "sent_len_1": 66.7948, "sent_len_max_0": 127.98, "sent_len_max_1": 207.0725, "stdk": 0.0464, "stdq": 0.0432, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20100 }, { "accuracy": 42.8711, "active_queue_size": 16384.0, "cl_loss": 4.04, "doc_norm": 1.3724, "encoder_q-embeddings": 791.4343, "encoder_q-layer.0": 542.033, "encoder_q-layer.1": 587.5309, "encoder_q-layer.10": 654.8328, "encoder_q-layer.11": 1499.0833, "encoder_q-layer.2": 655.9055, "encoder_q-layer.3": 688.0049, "encoder_q-layer.4": 765.9728, "encoder_q-layer.5": 747.3056, "encoder_q-layer.6": 782.7783, "encoder_q-layer.7": 700.0533, "encoder_q-layer.8": 657.373, "encoder_q-layer.9": 572.7188, "epoch": 0.09, "inbatch_neg_score": 0.2305, "inbatch_pos_score": 0.7622, "learning_rate": 4.731578947368421e-05, "loss": 4.04, "norm_diff": 0.033, "norm_loss": 0.0, "num_token_doc": 66.834, "num_token_overlap": 11.674, "num_token_query": 31.3278, "num_token_union": 65.1321, "num_word_context": 202.51, "num_word_doc": 49.8578, "num_word_query": 23.2722, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1177.5213, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2308, "query_norm": 1.3394, "queue_k_norm": 1.3739, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3278, "sent_len_1": 66.834, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.82, "stdk": 0.0465, "stdq": 0.0434, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20200 }, { "accuracy": 43.457, "active_queue_size": 16384.0, "cl_loss": 4.0433, "doc_norm": 1.3728, "encoder_q-embeddings": 326.0425, "encoder_q-layer.0": 226.7356, "encoder_q-layer.1": 255.8504, "encoder_q-layer.10": 272.7678, "encoder_q-layer.11": 738.6999, "encoder_q-layer.2": 284.2432, "encoder_q-layer.3": 268.0921, "encoder_q-layer.4": 285.626, "encoder_q-layer.5": 267.8694, "encoder_q-layer.6": 265.5244, "encoder_q-layer.7": 259.5269, "encoder_q-layer.8": 284.3301, "encoder_q-layer.9": 244.1092, "epoch": 0.09, "inbatch_neg_score": 0.2427, "inbatch_pos_score": 0.7588, "learning_rate": 4.728947368421053e-05, "loss": 4.0433, "norm_diff": 0.065, "norm_loss": 0.0, "num_token_doc": 66.6495, "num_token_overlap": 11.6549, "num_token_query": 31.3338, "num_token_union": 65.0396, "num_word_context": 202.237, "num_word_doc": 49.7596, "num_word_query": 23.2671, "postclip_grad_norm": 1.0, "preclip_grad_norm": 512.9628, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2434, "query_norm": 1.3078, "queue_k_norm": 1.3722, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3338, "sent_len_1": 66.6495, "sent_len_max_0": 127.9475, "sent_len_max_1": 210.2488, "stdk": 0.0465, "stdq": 0.0419, "stdqueue_k": 0.0465, "stdqueue_q": 0.0, "step": 20300 }, { "accuracy": 41.1621, "active_queue_size": 16384.0, "cl_loss": 4.0328, "doc_norm": 1.3749, "encoder_q-embeddings": 294.907, "encoder_q-layer.0": 189.924, "encoder_q-layer.1": 213.0555, "encoder_q-layer.10": 252.3193, "encoder_q-layer.11": 704.0092, "encoder_q-layer.2": 252.1244, "encoder_q-layer.3": 253.7212, "encoder_q-layer.4": 285.2926, "encoder_q-layer.5": 279.3953, "encoder_q-layer.6": 283.6273, "encoder_q-layer.7": 280.7432, "encoder_q-layer.8": 259.8993, "encoder_q-layer.9": 218.5689, "epoch": 0.09, "inbatch_neg_score": 0.2396, "inbatch_pos_score": 0.7539, "learning_rate": 4.726315789473684e-05, "loss": 4.0328, "norm_diff": 0.0514, "norm_loss": 0.0, "num_token_doc": 66.7436, "num_token_overlap": 11.6885, "num_token_query": 31.3971, "num_token_union": 65.0763, "num_word_context": 202.1099, "num_word_doc": 49.8025, "num_word_query": 23.3233, "postclip_grad_norm": 1.0, "preclip_grad_norm": 495.7062, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2402, "query_norm": 1.3235, "queue_k_norm": 1.3747, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3971, "sent_len_1": 66.7436, "sent_len_max_0": 127.98, "sent_len_max_1": 209.5462, "stdk": 0.0466, "stdq": 0.0426, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20400 }, { "accuracy": 44.2871, "active_queue_size": 16384.0, "cl_loss": 4.0331, "doc_norm": 1.3727, "encoder_q-embeddings": 1657.2239, "encoder_q-layer.0": 901.1868, "encoder_q-layer.1": 727.9326, "encoder_q-layer.10": 262.1692, "encoder_q-layer.11": 740.1703, "encoder_q-layer.2": 580.6187, "encoder_q-layer.3": 380.1385, "encoder_q-layer.4": 323.7832, "encoder_q-layer.5": 321.2024, "encoder_q-layer.6": 321.2729, "encoder_q-layer.7": 332.7237, "encoder_q-layer.8": 300.9543, "encoder_q-layer.9": 231.8975, "epoch": 0.09, "inbatch_neg_score": 0.2379, "inbatch_pos_score": 0.7817, "learning_rate": 4.7236842105263166e-05, "loss": 4.0331, "norm_diff": 0.0392, "norm_loss": 0.0, "num_token_doc": 66.9165, "num_token_overlap": 11.7007, "num_token_query": 31.4338, "num_token_union": 65.209, "num_word_context": 202.3383, "num_word_doc": 49.9062, "num_word_query": 23.3542, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1168.974, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2374, "query_norm": 1.3335, "queue_k_norm": 1.3761, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4338, "sent_len_1": 66.9165, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.8913, "stdk": 0.0465, "stdq": 0.0431, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20500 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 4.0285, "doc_norm": 1.3804, "encoder_q-embeddings": 907.0698, "encoder_q-layer.0": 629.6179, "encoder_q-layer.1": 616.9177, "encoder_q-layer.10": 266.3861, "encoder_q-layer.11": 703.0386, "encoder_q-layer.2": 716.14, "encoder_q-layer.3": 695.4691, "encoder_q-layer.4": 614.0417, "encoder_q-layer.5": 637.7635, "encoder_q-layer.6": 492.7838, "encoder_q-layer.7": 401.8128, "encoder_q-layer.8": 283.8896, "encoder_q-layer.9": 231.2769, "epoch": 0.09, "inbatch_neg_score": 0.2371, "inbatch_pos_score": 0.7842, "learning_rate": 4.721052631578948e-05, "loss": 4.0285, "norm_diff": 0.062, "norm_loss": 0.0, "num_token_doc": 66.8826, "num_token_overlap": 11.6739, "num_token_query": 31.4238, "num_token_union": 65.2208, "num_word_context": 202.4697, "num_word_doc": 49.93, "num_word_query": 23.3451, "postclip_grad_norm": 1.0, "preclip_grad_norm": 916.2767, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2373, "query_norm": 1.3184, "queue_k_norm": 1.3766, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4238, "sent_len_1": 66.8826, "sent_len_max_0": 127.98, "sent_len_max_1": 208.2262, "stdk": 0.0467, "stdq": 0.0425, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20600 }, { "accuracy": 42.6758, "active_queue_size": 16384.0, "cl_loss": 4.0385, "doc_norm": 1.3764, "encoder_q-embeddings": 429.8504, "encoder_q-layer.0": 287.4117, "encoder_q-layer.1": 318.418, "encoder_q-layer.10": 267.738, "encoder_q-layer.11": 715.2799, "encoder_q-layer.2": 362.5862, "encoder_q-layer.3": 385.4177, "encoder_q-layer.4": 399.9413, "encoder_q-layer.5": 410.528, "encoder_q-layer.6": 434.9132, "encoder_q-layer.7": 363.5621, "encoder_q-layer.8": 265.4635, "encoder_q-layer.9": 220.0101, "epoch": 0.09, "inbatch_neg_score": 0.2393, "inbatch_pos_score": 0.751, "learning_rate": 4.718421052631579e-05, "loss": 4.0385, "norm_diff": 0.0613, "norm_loss": 0.0, "num_token_doc": 66.8424, "num_token_overlap": 11.6532, "num_token_query": 31.288, "num_token_union": 65.1233, "num_word_context": 202.4169, "num_word_doc": 49.8527, "num_word_query": 23.2442, "postclip_grad_norm": 1.0, "preclip_grad_norm": 611.5835, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2393, "query_norm": 1.3151, "queue_k_norm": 1.3776, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.288, "sent_len_1": 66.8424, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.5087, "stdk": 0.0466, "stdq": 0.0424, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 20700 }, { "accuracy": 43.2617, "active_queue_size": 16384.0, "cl_loss": 4.0397, "doc_norm": 1.3724, "encoder_q-embeddings": 429.8464, "encoder_q-layer.0": 297.3635, "encoder_q-layer.1": 334.0823, "encoder_q-layer.10": 280.0142, "encoder_q-layer.11": 752.4022, "encoder_q-layer.2": 404.9613, "encoder_q-layer.3": 384.8599, "encoder_q-layer.4": 368.4422, "encoder_q-layer.5": 352.6096, "encoder_q-layer.6": 349.9116, "encoder_q-layer.7": 316.199, "encoder_q-layer.8": 296.722, "encoder_q-layer.9": 242.769, "epoch": 0.09, "inbatch_neg_score": 0.2416, "inbatch_pos_score": 0.7642, "learning_rate": 4.715789473684211e-05, "loss": 4.0397, "norm_diff": 0.0428, "norm_loss": 0.0, "num_token_doc": 66.6972, "num_token_overlap": 11.674, "num_token_query": 31.4328, "num_token_union": 65.0909, "num_word_context": 202.3783, "num_word_doc": 49.7773, "num_word_query": 23.3511, "postclip_grad_norm": 1.0, "preclip_grad_norm": 604.7454, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2404, "query_norm": 1.3296, "queue_k_norm": 1.3802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4328, "sent_len_1": 66.6972, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.0863, "stdk": 0.0464, "stdq": 0.0427, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 20800 }, { "accuracy": 43.5059, "active_queue_size": 16384.0, "cl_loss": 4.0255, "doc_norm": 1.3809, "encoder_q-embeddings": 545.1492, "encoder_q-layer.0": 383.5509, "encoder_q-layer.1": 427.6687, "encoder_q-layer.10": 306.382, "encoder_q-layer.11": 804.5112, "encoder_q-layer.2": 490.725, "encoder_q-layer.3": 499.6156, "encoder_q-layer.4": 523.5703, "encoder_q-layer.5": 518.8109, "encoder_q-layer.6": 529.4799, "encoder_q-layer.7": 462.4318, "encoder_q-layer.8": 381.1913, "encoder_q-layer.9": 285.9657, "epoch": 0.09, "inbatch_neg_score": 0.2479, "inbatch_pos_score": 0.7651, "learning_rate": 4.713157894736842e-05, "loss": 4.0255, "norm_diff": 0.0871, "norm_loss": 0.0, "num_token_doc": 66.7064, "num_token_overlap": 11.6675, "num_token_query": 31.3609, "num_token_union": 65.0756, "num_word_context": 202.6183, "num_word_doc": 49.777, "num_word_query": 23.2997, "postclip_grad_norm": 1.0, "preclip_grad_norm": 744.1676, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2468, "query_norm": 1.2939, "queue_k_norm": 1.3782, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3609, "sent_len_1": 66.7064, "sent_len_max_0": 127.99, "sent_len_max_1": 208.83, "stdk": 0.0467, "stdq": 0.0413, "stdqueue_k": 0.0466, "stdqueue_q": 0.0, "step": 20900 }, { "accuracy": 44.9707, "active_queue_size": 16384.0, "cl_loss": 4.015, "doc_norm": 1.3808, "encoder_q-embeddings": 412.3219, "encoder_q-layer.0": 269.7297, "encoder_q-layer.1": 290.6502, "encoder_q-layer.10": 287.3152, "encoder_q-layer.11": 698.269, "encoder_q-layer.2": 351.1871, "encoder_q-layer.3": 391.9489, "encoder_q-layer.4": 401.1783, "encoder_q-layer.5": 385.6622, "encoder_q-layer.6": 398.9194, "encoder_q-layer.7": 391.4098, "encoder_q-layer.8": 304.0455, "encoder_q-layer.9": 239.1502, "epoch": 0.09, "inbatch_neg_score": 0.2382, "inbatch_pos_score": 0.7788, "learning_rate": 4.7105263157894744e-05, "loss": 4.015, "norm_diff": 0.0434, "norm_loss": 0.0, "num_token_doc": 66.862, "num_token_overlap": 11.7363, "num_token_query": 31.4943, "num_token_union": 65.1821, "num_word_context": 202.4309, "num_word_doc": 49.8723, "num_word_query": 23.397, "postclip_grad_norm": 1.0, "preclip_grad_norm": 596.6183, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2375, "query_norm": 1.3374, "queue_k_norm": 1.3813, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4943, "sent_len_1": 66.862, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.3762, "stdk": 0.0467, "stdq": 0.0431, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 21000 }, { "accuracy": 43.9453, "active_queue_size": 16384.0, "cl_loss": 4.0246, "doc_norm": 1.3793, "encoder_q-embeddings": 364.373, "encoder_q-layer.0": 241.1812, "encoder_q-layer.1": 267.2601, "encoder_q-layer.10": 284.654, "encoder_q-layer.11": 724.8091, "encoder_q-layer.2": 296.2347, "encoder_q-layer.3": 313.6662, "encoder_q-layer.4": 330.7151, "encoder_q-layer.5": 341.958, "encoder_q-layer.6": 368.5368, "encoder_q-layer.7": 396.347, "encoder_q-layer.8": 345.0226, "encoder_q-layer.9": 236.403, "epoch": 0.09, "inbatch_neg_score": 0.2333, "inbatch_pos_score": 0.7607, "learning_rate": 4.7078947368421056e-05, "loss": 4.0246, "norm_diff": 0.0487, "norm_loss": 0.0, "num_token_doc": 66.6632, "num_token_overlap": 11.665, "num_token_query": 31.3777, "num_token_union": 65.0881, "num_word_context": 202.3898, "num_word_doc": 49.7307, "num_word_query": 23.3084, "postclip_grad_norm": 1.0, "preclip_grad_norm": 571.9443, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2332, "query_norm": 1.3306, "queue_k_norm": 1.3797, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3777, "sent_len_1": 66.6632, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.8975, "stdk": 0.0466, "stdq": 0.0431, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 21100 }, { "accuracy": 43.6035, "active_queue_size": 16384.0, "cl_loss": 4.0247, "doc_norm": 1.3805, "encoder_q-embeddings": 419.9934, "encoder_q-layer.0": 292.1616, "encoder_q-layer.1": 335.8393, "encoder_q-layer.10": 253.2278, "encoder_q-layer.11": 758.3459, "encoder_q-layer.2": 392.4514, "encoder_q-layer.3": 421.9703, "encoder_q-layer.4": 452.053, "encoder_q-layer.5": 449.0335, "encoder_q-layer.6": 444.5878, "encoder_q-layer.7": 411.6963, "encoder_q-layer.8": 368.1635, "encoder_q-layer.9": 235.6741, "epoch": 0.09, "inbatch_neg_score": 0.2365, "inbatch_pos_score": 0.7651, "learning_rate": 4.705263157894737e-05, "loss": 4.0247, "norm_diff": 0.0537, "norm_loss": 0.0, "num_token_doc": 66.9459, "num_token_overlap": 11.6853, "num_token_query": 31.4174, "num_token_union": 65.2411, "num_word_context": 202.5521, "num_word_doc": 49.9824, "num_word_query": 23.3538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 659.0457, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2367, "query_norm": 1.3267, "queue_k_norm": 1.3794, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4174, "sent_len_1": 66.9459, "sent_len_max_0": 127.99, "sent_len_max_1": 207.2875, "stdk": 0.0467, "stdq": 0.0426, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 21200 }, { "accuracy": 45.3613, "active_queue_size": 16384.0, "cl_loss": 4.0198, "doc_norm": 1.3808, "encoder_q-embeddings": 342.4696, "encoder_q-layer.0": 230.9749, "encoder_q-layer.1": 253.3746, "encoder_q-layer.10": 272.3789, "encoder_q-layer.11": 698.5577, "encoder_q-layer.2": 294.0509, "encoder_q-layer.3": 314.7853, "encoder_q-layer.4": 300.1844, "encoder_q-layer.5": 295.5903, "encoder_q-layer.6": 306.7529, "encoder_q-layer.7": 276.6792, "encoder_q-layer.8": 287.6674, "encoder_q-layer.9": 235.8706, "epoch": 0.09, "inbatch_neg_score": 0.2302, "inbatch_pos_score": 0.7832, "learning_rate": 4.7026315789473686e-05, "loss": 4.0198, "norm_diff": 0.0534, "norm_loss": 0.0, "num_token_doc": 66.6783, "num_token_overlap": 11.6343, "num_token_query": 31.3522, "num_token_union": 65.0959, "num_word_context": 202.1404, "num_word_doc": 49.7669, "num_word_query": 23.2755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 524.9737, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2306, "query_norm": 1.3275, "queue_k_norm": 1.3822, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3522, "sent_len_1": 66.6783, "sent_len_max_0": 127.995, "sent_len_max_1": 205.4925, "stdk": 0.0467, "stdq": 0.043, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 21300 }, { "accuracy": 43.1641, "active_queue_size": 16384.0, "cl_loss": 4.0321, "doc_norm": 1.3794, "encoder_q-embeddings": 422.2675, "encoder_q-layer.0": 305.114, "encoder_q-layer.1": 337.0635, "encoder_q-layer.10": 254.4336, "encoder_q-layer.11": 734.6227, "encoder_q-layer.2": 392.5885, "encoder_q-layer.3": 422.3456, "encoder_q-layer.4": 445.1218, "encoder_q-layer.5": 401.898, "encoder_q-layer.6": 354.3693, "encoder_q-layer.7": 327.3499, "encoder_q-layer.8": 306.0019, "encoder_q-layer.9": 237.4364, "epoch": 0.09, "inbatch_neg_score": 0.2198, "inbatch_pos_score": 0.7427, "learning_rate": 4.7e-05, "loss": 4.0321, "norm_diff": 0.06, "norm_loss": 0.0, "num_token_doc": 66.9755, "num_token_overlap": 11.6657, "num_token_query": 31.3652, "num_token_union": 65.2485, "num_word_context": 202.6454, "num_word_doc": 49.9508, "num_word_query": 23.2972, "postclip_grad_norm": 1.0, "preclip_grad_norm": 619.2952, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2202, "query_norm": 1.3194, "queue_k_norm": 1.38, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3652, "sent_len_1": 66.9755, "sent_len_max_0": 127.9575, "sent_len_max_1": 207.8388, "stdk": 0.0466, "stdq": 0.0428, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 21400 }, { "accuracy": 42.7734, "active_queue_size": 16384.0, "cl_loss": 4.0236, "doc_norm": 1.3834, "encoder_q-embeddings": 536.8879, "encoder_q-layer.0": 385.8084, "encoder_q-layer.1": 444.8593, "encoder_q-layer.10": 281.4864, "encoder_q-layer.11": 729.6723, "encoder_q-layer.2": 501.5022, "encoder_q-layer.3": 566.2431, "encoder_q-layer.4": 563.6492, "encoder_q-layer.5": 591.4336, "encoder_q-layer.6": 502.0146, "encoder_q-layer.7": 426.6293, "encoder_q-layer.8": 352.4507, "encoder_q-layer.9": 256.4014, "epoch": 0.09, "inbatch_neg_score": 0.2151, "inbatch_pos_score": 0.7349, "learning_rate": 4.697368421052632e-05, "loss": 4.0236, "norm_diff": 0.0739, "norm_loss": 0.0, "num_token_doc": 66.6974, "num_token_overlap": 11.6373, "num_token_query": 31.3617, "num_token_union": 65.1019, "num_word_context": 202.3543, "num_word_doc": 49.7764, "num_word_query": 23.3044, "postclip_grad_norm": 1.0, "preclip_grad_norm": 748.3092, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2147, "query_norm": 1.3095, "queue_k_norm": 1.3815, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3617, "sent_len_1": 66.6974, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.5137, "stdk": 0.0468, "stdq": 0.0428, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 21500 }, { "accuracy": 43.2617, "active_queue_size": 16384.0, "cl_loss": 4.0236, "doc_norm": 1.3769, "encoder_q-embeddings": 1023.2269, "encoder_q-layer.0": 702.1454, "encoder_q-layer.1": 746.0784, "encoder_q-layer.10": 257.1463, "encoder_q-layer.11": 705.1056, "encoder_q-layer.2": 914.3267, "encoder_q-layer.3": 1009.2734, "encoder_q-layer.4": 1225.8948, "encoder_q-layer.5": 1345.8787, "encoder_q-layer.6": 979.9724, "encoder_q-layer.7": 941.589, "encoder_q-layer.8": 480.1207, "encoder_q-layer.9": 263.8483, "epoch": 0.09, "inbatch_neg_score": 0.2107, "inbatch_pos_score": 0.7393, "learning_rate": 4.6947368421052634e-05, "loss": 4.0236, "norm_diff": 0.0636, "norm_loss": 0.0, "num_token_doc": 66.6564, "num_token_overlap": 11.6597, "num_token_query": 31.2725, "num_token_union": 64.9994, "num_word_context": 202.2837, "num_word_doc": 49.7546, "num_word_query": 23.2209, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1317.3459, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2101, "query_norm": 1.3133, "queue_k_norm": 1.3794, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2725, "sent_len_1": 66.6564, "sent_len_max_0": 127.9788, "sent_len_max_1": 206.735, "stdk": 0.0466, "stdq": 0.043, "stdqueue_k": 0.0467, "stdqueue_q": 0.0, "step": 21600 }, { "accuracy": 44.1895, "active_queue_size": 16384.0, "cl_loss": 4.0143, "doc_norm": 1.3798, "encoder_q-embeddings": 457.8099, "encoder_q-layer.0": 311.6125, "encoder_q-layer.1": 352.4633, "encoder_q-layer.10": 243.6246, "encoder_q-layer.11": 664.0568, "encoder_q-layer.2": 436.0172, "encoder_q-layer.3": 441.6349, "encoder_q-layer.4": 447.404, "encoder_q-layer.5": 381.3961, "encoder_q-layer.6": 337.0116, "encoder_q-layer.7": 320.8466, "encoder_q-layer.8": 281.6505, "encoder_q-layer.9": 223.748, "epoch": 0.09, "inbatch_neg_score": 0.2062, "inbatch_pos_score": 0.7441, "learning_rate": 4.6921052631578946e-05, "loss": 4.0143, "norm_diff": 0.0645, "norm_loss": 0.0, "num_token_doc": 66.7165, "num_token_overlap": 11.6835, "num_token_query": 31.3318, "num_token_union": 65.0529, "num_word_context": 202.2769, "num_word_doc": 49.7833, "num_word_query": 23.2731, "postclip_grad_norm": 1.0, "preclip_grad_norm": 601.1863, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2061, "query_norm": 1.3153, "queue_k_norm": 1.3811, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3318, "sent_len_1": 66.7165, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.8713, "stdk": 0.0467, "stdq": 0.0433, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 21700 }, { "accuracy": 46.1914, "active_queue_size": 16384.0, "cl_loss": 3.996, "doc_norm": 1.3794, "encoder_q-embeddings": 618.5956, "encoder_q-layer.0": 460.3917, "encoder_q-layer.1": 446.2235, "encoder_q-layer.10": 239.1885, "encoder_q-layer.11": 667.2806, "encoder_q-layer.2": 493.9839, "encoder_q-layer.3": 551.5432, "encoder_q-layer.4": 534.8137, "encoder_q-layer.5": 553.1906, "encoder_q-layer.6": 476.6974, "encoder_q-layer.7": 331.7195, "encoder_q-layer.8": 279.4747, "encoder_q-layer.9": 221.41, "epoch": 0.09, "inbatch_neg_score": 0.1971, "inbatch_pos_score": 0.7451, "learning_rate": 4.6894736842105264e-05, "loss": 3.996, "norm_diff": 0.0819, "norm_loss": 0.0, "num_token_doc": 66.6838, "num_token_overlap": 11.7287, "num_token_query": 31.4661, "num_token_union": 65.0831, "num_word_context": 202.0924, "num_word_doc": 49.7141, "num_word_query": 23.3718, "postclip_grad_norm": 1.0, "preclip_grad_norm": 725.8494, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1978, "query_norm": 1.2976, "queue_k_norm": 1.3797, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4661, "sent_len_1": 66.6838, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.6725, "stdk": 0.0468, "stdq": 0.0427, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 21800 }, { "accuracy": 44.1895, "active_queue_size": 16384.0, "cl_loss": 4.0063, "doc_norm": 1.38, "encoder_q-embeddings": 937.6085, "encoder_q-layer.0": 700.9157, "encoder_q-layer.1": 750.5759, "encoder_q-layer.10": 256.5702, "encoder_q-layer.11": 683.6167, "encoder_q-layer.2": 816.2384, "encoder_q-layer.3": 862.3259, "encoder_q-layer.4": 1064.7086, "encoder_q-layer.5": 985.3499, "encoder_q-layer.6": 924.5995, "encoder_q-layer.7": 841.5535, "encoder_q-layer.8": 492.8742, "encoder_q-layer.9": 249.0365, "epoch": 0.1, "inbatch_neg_score": 0.1932, "inbatch_pos_score": 0.7319, "learning_rate": 4.686842105263158e-05, "loss": 4.0063, "norm_diff": 0.0732, "norm_loss": 0.0, "num_token_doc": 66.8353, "num_token_overlap": 11.6764, "num_token_query": 31.4254, "num_token_union": 65.1705, "num_word_context": 202.5785, "num_word_doc": 49.8626, "num_word_query": 23.3525, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1173.987, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1927, "query_norm": 1.3069, "queue_k_norm": 1.3791, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4254, "sent_len_1": 66.8353, "sent_len_max_0": 127.99, "sent_len_max_1": 211.3575, "stdk": 0.0468, "stdq": 0.0431, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 21900 }, { "accuracy": 43.6035, "active_queue_size": 16384.0, "cl_loss": 4.0142, "doc_norm": 1.3816, "encoder_q-embeddings": 269.9725, "encoder_q-layer.0": 176.9893, "encoder_q-layer.1": 185.8117, "encoder_q-layer.10": 271.9829, "encoder_q-layer.11": 681.3361, "encoder_q-layer.2": 204.9063, "encoder_q-layer.3": 220.7168, "encoder_q-layer.4": 232.9363, "encoder_q-layer.5": 224.1022, "encoder_q-layer.6": 239.7521, "encoder_q-layer.7": 244.4081, "encoder_q-layer.8": 262.0571, "encoder_q-layer.9": 229.1361, "epoch": 0.1, "inbatch_neg_score": 0.1986, "inbatch_pos_score": 0.731, "learning_rate": 4.68421052631579e-05, "loss": 4.0142, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 67.0441, "num_token_overlap": 11.6908, "num_token_query": 31.4109, "num_token_union": 65.2856, "num_word_context": 202.7496, "num_word_doc": 50.0472, "num_word_query": 23.3446, "postclip_grad_norm": 1.0, "preclip_grad_norm": 463.2127, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1981, "query_norm": 1.3098, "queue_k_norm": 1.3769, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4109, "sent_len_1": 67.0441, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.9038, "stdk": 0.0469, "stdq": 0.043, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 22000 }, { "accuracy": 44.0918, "active_queue_size": 16384.0, "cl_loss": 4.002, "doc_norm": 1.3742, "encoder_q-embeddings": 425.2613, "encoder_q-layer.0": 296.4056, "encoder_q-layer.1": 326.3358, "encoder_q-layer.10": 245.6094, "encoder_q-layer.11": 635.5563, "encoder_q-layer.2": 398.8632, "encoder_q-layer.3": 420.0312, "encoder_q-layer.4": 478.9178, "encoder_q-layer.5": 501.8036, "encoder_q-layer.6": 415.5693, "encoder_q-layer.7": 303.3814, "encoder_q-layer.8": 271.5108, "encoder_q-layer.9": 211.7924, "epoch": 0.1, "inbatch_neg_score": 0.1955, "inbatch_pos_score": 0.7295, "learning_rate": 4.681578947368421e-05, "loss": 4.002, "norm_diff": 0.0669, "norm_loss": 0.0, "num_token_doc": 66.8438, "num_token_overlap": 11.6929, "num_token_query": 31.4125, "num_token_union": 65.1983, "num_word_context": 202.3366, "num_word_doc": 49.8665, "num_word_query": 23.3202, "postclip_grad_norm": 1.0, "preclip_grad_norm": 606.0785, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1957, "query_norm": 1.3073, "queue_k_norm": 1.3758, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4125, "sent_len_1": 66.8438, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.4613, "stdk": 0.0467, "stdq": 0.0429, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 22100 }, { "accuracy": 44.7754, "active_queue_size": 16384.0, "cl_loss": 3.996, "doc_norm": 1.3787, "encoder_q-embeddings": 236.7559, "encoder_q-layer.0": 151.6267, "encoder_q-layer.1": 163.3874, "encoder_q-layer.10": 244.0213, "encoder_q-layer.11": 696.1094, "encoder_q-layer.2": 182.9207, "encoder_q-layer.3": 193.5858, "encoder_q-layer.4": 198.9253, "encoder_q-layer.5": 205.4427, "encoder_q-layer.6": 225.2859, "encoder_q-layer.7": 211.4608, "encoder_q-layer.8": 234.7064, "encoder_q-layer.9": 216.3167, "epoch": 0.1, "inbatch_neg_score": 0.1958, "inbatch_pos_score": 0.7383, "learning_rate": 4.678947368421053e-05, "loss": 3.996, "norm_diff": 0.0689, "norm_loss": 0.0, "num_token_doc": 66.8073, "num_token_overlap": 11.6939, "num_token_query": 31.501, "num_token_union": 65.2336, "num_word_context": 202.2081, "num_word_doc": 49.8493, "num_word_query": 23.3998, "postclip_grad_norm": 1.0, "preclip_grad_norm": 445.2091, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.196, "query_norm": 1.3097, "queue_k_norm": 1.3779, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.501, "sent_len_1": 66.8073, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.3325, "stdk": 0.0469, "stdq": 0.0428, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22200 }, { "accuracy": 45.3125, "active_queue_size": 16384.0, "cl_loss": 3.997, "doc_norm": 1.3787, "encoder_q-embeddings": 535.7493, "encoder_q-layer.0": 370.4314, "encoder_q-layer.1": 404.4338, "encoder_q-layer.10": 535.3961, "encoder_q-layer.11": 1475.5536, "encoder_q-layer.2": 469.6897, "encoder_q-layer.3": 485.4842, "encoder_q-layer.4": 535.9646, "encoder_q-layer.5": 510.0936, "encoder_q-layer.6": 523.2844, "encoder_q-layer.7": 488.1361, "encoder_q-layer.8": 567.7261, "encoder_q-layer.9": 498.0872, "epoch": 0.1, "inbatch_neg_score": 0.1986, "inbatch_pos_score": 0.7344, "learning_rate": 4.676315789473684e-05, "loss": 3.997, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.8512, "num_token_overlap": 11.6921, "num_token_query": 31.4383, "num_token_union": 65.1938, "num_word_context": 202.4546, "num_word_doc": 49.8644, "num_word_query": 23.3418, "postclip_grad_norm": 1.0, "preclip_grad_norm": 985.9016, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1991, "query_norm": 1.2937, "queue_k_norm": 1.3739, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4383, "sent_len_1": 66.8512, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.4437, "stdk": 0.0469, "stdq": 0.0424, "stdqueue_k": 0.0468, "stdqueue_q": 0.0, "step": 22300 }, { "accuracy": 43.7012, "active_queue_size": 16384.0, "cl_loss": 3.9902, "doc_norm": 1.3734, "encoder_q-embeddings": 845.1653, "encoder_q-layer.0": 608.7545, "encoder_q-layer.1": 663.7355, "encoder_q-layer.10": 527.4588, "encoder_q-layer.11": 1457.736, "encoder_q-layer.2": 685.8138, "encoder_q-layer.3": 699.0323, "encoder_q-layer.4": 627.6661, "encoder_q-layer.5": 633.4286, "encoder_q-layer.6": 610.3799, "encoder_q-layer.7": 504.3256, "encoder_q-layer.8": 516.8823, "encoder_q-layer.9": 446.0824, "epoch": 0.1, "inbatch_neg_score": 0.1943, "inbatch_pos_score": 0.7266, "learning_rate": 4.673684210526316e-05, "loss": 3.9902, "norm_diff": 0.0908, "norm_loss": 0.0, "num_token_doc": 66.8255, "num_token_overlap": 11.6853, "num_token_query": 31.3284, "num_token_union": 65.1046, "num_word_context": 202.1102, "num_word_doc": 49.8605, "num_word_query": 23.2614, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1150.0404, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1936, "query_norm": 1.2826, "queue_k_norm": 1.375, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3284, "sent_len_1": 66.8255, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.3187, "stdk": 0.0467, "stdq": 0.0421, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22400 }, { "accuracy": 44.8242, "active_queue_size": 16384.0, "cl_loss": 3.9726, "doc_norm": 1.3765, "encoder_q-embeddings": 533.8687, "encoder_q-layer.0": 363.2279, "encoder_q-layer.1": 389.7845, "encoder_q-layer.10": 489.0467, "encoder_q-layer.11": 1381.8457, "encoder_q-layer.2": 477.7176, "encoder_q-layer.3": 482.7429, "encoder_q-layer.4": 501.3714, "encoder_q-layer.5": 482.7228, "encoder_q-layer.6": 473.7027, "encoder_q-layer.7": 475.5347, "encoder_q-layer.8": 505.5698, "encoder_q-layer.9": 450.8216, "epoch": 0.1, "inbatch_neg_score": 0.2017, "inbatch_pos_score": 0.75, "learning_rate": 4.671052631578948e-05, "loss": 3.9726, "norm_diff": 0.0685, "norm_loss": 0.0, "num_token_doc": 66.897, "num_token_overlap": 11.7154, "num_token_query": 31.4481, "num_token_union": 65.2119, "num_word_context": 202.3479, "num_word_doc": 49.9138, "num_word_query": 23.3662, "postclip_grad_norm": 1.0, "preclip_grad_norm": 938.2554, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2013, "query_norm": 1.308, "queue_k_norm": 1.3768, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4481, "sent_len_1": 66.897, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.3388, "stdk": 0.0469, "stdq": 0.043, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22500 }, { "accuracy": 45.3613, "active_queue_size": 16384.0, "cl_loss": 3.9969, "doc_norm": 1.3717, "encoder_q-embeddings": 598.5904, "encoder_q-layer.0": 401.2482, "encoder_q-layer.1": 451.6981, "encoder_q-layer.10": 543.5546, "encoder_q-layer.11": 1390.8475, "encoder_q-layer.2": 527.3488, "encoder_q-layer.3": 572.0406, "encoder_q-layer.4": 585.8296, "encoder_q-layer.5": 580.2453, "encoder_q-layer.6": 615.8586, "encoder_q-layer.7": 586.3115, "encoder_q-layer.8": 570.41, "encoder_q-layer.9": 496.6744, "epoch": 0.1, "inbatch_neg_score": 0.1977, "inbatch_pos_score": 0.7446, "learning_rate": 4.668421052631579e-05, "loss": 3.9969, "norm_diff": 0.073, "norm_loss": 0.0, "num_token_doc": 66.7166, "num_token_overlap": 11.6528, "num_token_query": 31.2481, "num_token_union": 65.0399, "num_word_context": 202.172, "num_word_doc": 49.7994, "num_word_query": 23.2163, "postclip_grad_norm": 1.0, "preclip_grad_norm": 996.3137, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1981, "query_norm": 1.2987, "queue_k_norm": 1.3766, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2481, "sent_len_1": 66.7166, "sent_len_max_0": 128.0, "sent_len_max_1": 209.75, "stdk": 0.0467, "stdq": 0.0426, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22600 }, { "accuracy": 45.0684, "active_queue_size": 16384.0, "cl_loss": 3.9842, "doc_norm": 1.3755, "encoder_q-embeddings": 555.7371, "encoder_q-layer.0": 396.6278, "encoder_q-layer.1": 407.2003, "encoder_q-layer.10": 495.104, "encoder_q-layer.11": 1317.1532, "encoder_q-layer.2": 436.9043, "encoder_q-layer.3": 444.0514, "encoder_q-layer.4": 449.9103, "encoder_q-layer.5": 453.9521, "encoder_q-layer.6": 445.5097, "encoder_q-layer.7": 439.3712, "encoder_q-layer.8": 492.1139, "encoder_q-layer.9": 445.0031, "epoch": 0.1, "inbatch_neg_score": 0.2072, "inbatch_pos_score": 0.749, "learning_rate": 4.665789473684211e-05, "loss": 3.9842, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.7802, "num_token_overlap": 11.6614, "num_token_query": 31.3885, "num_token_union": 65.1648, "num_word_context": 202.5328, "num_word_doc": 49.8155, "num_word_query": 23.3251, "postclip_grad_norm": 1.0, "preclip_grad_norm": 912.9367, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2076, "query_norm": 1.293, "queue_k_norm": 1.375, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3885, "sent_len_1": 66.7802, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.86, "stdk": 0.0469, "stdq": 0.0422, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22700 }, { "accuracy": 44.9707, "active_queue_size": 16384.0, "cl_loss": 3.9791, "doc_norm": 1.3775, "encoder_q-embeddings": 654.4832, "encoder_q-layer.0": 468.1518, "encoder_q-layer.1": 510.2227, "encoder_q-layer.10": 520.7351, "encoder_q-layer.11": 1325.6036, "encoder_q-layer.2": 599.7917, "encoder_q-layer.3": 576.0689, "encoder_q-layer.4": 582.4247, "encoder_q-layer.5": 566.9827, "encoder_q-layer.6": 571.8221, "encoder_q-layer.7": 509.025, "encoder_q-layer.8": 530.4488, "encoder_q-layer.9": 471.2952, "epoch": 0.1, "inbatch_neg_score": 0.2163, "inbatch_pos_score": 0.7607, "learning_rate": 4.663157894736842e-05, "loss": 3.9791, "norm_diff": 0.046, "norm_loss": 0.0, "num_token_doc": 66.9317, "num_token_overlap": 11.7106, "num_token_query": 31.484, "num_token_union": 65.2838, "num_word_context": 202.4389, "num_word_doc": 49.9353, "num_word_query": 23.3942, "postclip_grad_norm": 1.0, "preclip_grad_norm": 992.825, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2163, "query_norm": 1.3314, "queue_k_norm": 1.3763, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.484, "sent_len_1": 66.9317, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.9075, "stdk": 0.0469, "stdq": 0.0435, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 22800 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 4.0055, "doc_norm": 1.3805, "encoder_q-embeddings": 513.8492, "encoder_q-layer.0": 332.0457, "encoder_q-layer.1": 356.2608, "encoder_q-layer.10": 530.9333, "encoder_q-layer.11": 1445.3439, "encoder_q-layer.2": 393.6687, "encoder_q-layer.3": 408.3295, "encoder_q-layer.4": 430.6608, "encoder_q-layer.5": 432.4726, "encoder_q-layer.6": 451.6646, "encoder_q-layer.7": 490.6778, "encoder_q-layer.8": 524.9384, "encoder_q-layer.9": 463.8948, "epoch": 0.1, "inbatch_neg_score": 0.2158, "inbatch_pos_score": 0.7642, "learning_rate": 4.660526315789474e-05, "loss": 4.0055, "norm_diff": 0.0693, "norm_loss": 0.0, "num_token_doc": 66.6098, "num_token_overlap": 11.6718, "num_token_query": 31.3569, "num_token_union": 65.0391, "num_word_context": 202.3033, "num_word_doc": 49.7086, "num_word_query": 23.2874, "postclip_grad_norm": 1.0, "preclip_grad_norm": 927.7357, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2172, "query_norm": 1.3112, "queue_k_norm": 1.378, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3569, "sent_len_1": 66.6098, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.6738, "stdk": 0.047, "stdq": 0.0427, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 22900 }, { "accuracy": 44.043, "active_queue_size": 16384.0, "cl_loss": 3.9698, "doc_norm": 1.3819, "encoder_q-embeddings": 490.3489, "encoder_q-layer.0": 314.2878, "encoder_q-layer.1": 348.7464, "encoder_q-layer.10": 529.7026, "encoder_q-layer.11": 1318.3594, "encoder_q-layer.2": 393.7744, "encoder_q-layer.3": 395.4687, "encoder_q-layer.4": 420.7794, "encoder_q-layer.5": 419.3048, "encoder_q-layer.6": 457.8877, "encoder_q-layer.7": 478.4963, "encoder_q-layer.8": 499.7484, "encoder_q-layer.9": 461.8991, "epoch": 0.1, "inbatch_neg_score": 0.2216, "inbatch_pos_score": 0.772, "learning_rate": 4.657894736842106e-05, "loss": 3.9698, "norm_diff": 0.055, "norm_loss": 0.0, "num_token_doc": 66.8335, "num_token_overlap": 11.6732, "num_token_query": 31.4147, "num_token_union": 65.1731, "num_word_context": 202.3846, "num_word_doc": 49.8809, "num_word_query": 23.3327, "postclip_grad_norm": 1.0, "preclip_grad_norm": 862.5785, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.222, "query_norm": 1.3269, "queue_k_norm": 1.3768, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4147, "sent_len_1": 66.8335, "sent_len_max_0": 127.995, "sent_len_max_1": 208.5387, "stdk": 0.0471, "stdq": 0.043, "stdqueue_k": 0.0469, "stdqueue_q": 0.0, "step": 23000 }, { "accuracy": 45.1172, "active_queue_size": 16384.0, "cl_loss": 3.9816, "doc_norm": 1.3787, "encoder_q-embeddings": 954.4375, "encoder_q-layer.0": 679.8049, "encoder_q-layer.1": 649.5247, "encoder_q-layer.10": 518.6021, "encoder_q-layer.11": 1356.3196, "encoder_q-layer.2": 690.1258, "encoder_q-layer.3": 706.3157, "encoder_q-layer.4": 715.3725, "encoder_q-layer.5": 677.6031, "encoder_q-layer.6": 643.7446, "encoder_q-layer.7": 595.0536, "encoder_q-layer.8": 594.9708, "encoder_q-layer.9": 477.1049, "epoch": 0.1, "inbatch_neg_score": 0.2163, "inbatch_pos_score": 0.7476, "learning_rate": 4.655263157894737e-05, "loss": 3.9816, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.699, "num_token_overlap": 11.6364, "num_token_query": 31.3028, "num_token_union": 65.0612, "num_word_context": 202.0857, "num_word_doc": 49.7773, "num_word_query": 23.2437, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1162.8676, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2172, "query_norm": 1.3051, "queue_k_norm": 1.3792, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3028, "sent_len_1": 66.699, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.3363, "stdk": 0.0469, "stdq": 0.0424, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 23100 }, { "accuracy": 46.582, "active_queue_size": 16384.0, "cl_loss": 3.9655, "doc_norm": 1.3789, "encoder_q-embeddings": 476.1818, "encoder_q-layer.0": 320.9952, "encoder_q-layer.1": 328.3491, "encoder_q-layer.10": 491.5834, "encoder_q-layer.11": 1327.804, "encoder_q-layer.2": 355.8557, "encoder_q-layer.3": 369.4676, "encoder_q-layer.4": 397.9005, "encoder_q-layer.5": 384.8106, "encoder_q-layer.6": 420.9275, "encoder_q-layer.7": 465.1112, "encoder_q-layer.8": 524.3212, "encoder_q-layer.9": 457.7055, "epoch": 0.1, "inbatch_neg_score": 0.2203, "inbatch_pos_score": 0.7754, "learning_rate": 4.652631578947369e-05, "loss": 3.9655, "norm_diff": 0.0521, "norm_loss": 0.0, "num_token_doc": 66.8099, "num_token_overlap": 11.71, "num_token_query": 31.4314, "num_token_union": 65.136, "num_word_context": 202.3095, "num_word_doc": 49.8527, "num_word_query": 23.3358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 863.4586, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2209, "query_norm": 1.3268, "queue_k_norm": 1.3794, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4314, "sent_len_1": 66.8099, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.775, "stdk": 0.0469, "stdq": 0.0433, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 23200 }, { "accuracy": 45.6055, "active_queue_size": 16384.0, "cl_loss": 3.9724, "doc_norm": 1.3757, "encoder_q-embeddings": 844.8682, "encoder_q-layer.0": 636.3945, "encoder_q-layer.1": 677.9799, "encoder_q-layer.10": 499.5093, "encoder_q-layer.11": 1343.5397, "encoder_q-layer.2": 701.5591, "encoder_q-layer.3": 715.3497, "encoder_q-layer.4": 748.3585, "encoder_q-layer.5": 607.3199, "encoder_q-layer.6": 650.2624, "encoder_q-layer.7": 629.5871, "encoder_q-layer.8": 592.278, "encoder_q-layer.9": 456.1558, "epoch": 0.1, "inbatch_neg_score": 0.215, "inbatch_pos_score": 0.7627, "learning_rate": 4.6500000000000005e-05, "loss": 3.9724, "norm_diff": 0.0652, "norm_loss": 0.0, "num_token_doc": 66.6716, "num_token_overlap": 11.715, "num_token_query": 31.4456, "num_token_union": 65.01, "num_word_context": 202.2857, "num_word_doc": 49.7298, "num_word_query": 23.3592, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1144.5811, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2151, "query_norm": 1.3105, "queue_k_norm": 1.3787, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4456, "sent_len_1": 66.6716, "sent_len_max_0": 128.0, "sent_len_max_1": 211.7425, "stdk": 0.0468, "stdq": 0.043, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 23300 }, { "accuracy": 42.6758, "active_queue_size": 16384.0, "cl_loss": 3.9711, "doc_norm": 1.3809, "encoder_q-embeddings": 450.9623, "encoder_q-layer.0": 313.1239, "encoder_q-layer.1": 334.3956, "encoder_q-layer.10": 496.0883, "encoder_q-layer.11": 1329.882, "encoder_q-layer.2": 389.0153, "encoder_q-layer.3": 403.0391, "encoder_q-layer.4": 409.0519, "encoder_q-layer.5": 427.5261, "encoder_q-layer.6": 455.2746, "encoder_q-layer.7": 454.6057, "encoder_q-layer.8": 510.9487, "encoder_q-layer.9": 442.8723, "epoch": 0.1, "inbatch_neg_score": 0.2188, "inbatch_pos_score": 0.7417, "learning_rate": 4.647368421052632e-05, "loss": 3.9711, "norm_diff": 0.0896, "norm_loss": 0.0, "num_token_doc": 66.7782, "num_token_overlap": 11.6833, "num_token_query": 31.4975, "num_token_union": 65.1794, "num_word_context": 202.4503, "num_word_doc": 49.8397, "num_word_query": 23.3936, "postclip_grad_norm": 1.0, "preclip_grad_norm": 873.5275, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2195, "query_norm": 1.2913, "queue_k_norm": 1.3802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4975, "sent_len_1": 66.7782, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.65, "stdk": 0.047, "stdq": 0.0422, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 23400 }, { "accuracy": 44.1406, "active_queue_size": 16384.0, "cl_loss": 3.9542, "doc_norm": 1.38, "encoder_q-embeddings": 493.5925, "encoder_q-layer.0": 332.5387, "encoder_q-layer.1": 368.7432, "encoder_q-layer.10": 500.184, "encoder_q-layer.11": 1345.3582, "encoder_q-layer.2": 423.5547, "encoder_q-layer.3": 423.9776, "encoder_q-layer.4": 431.2545, "encoder_q-layer.5": 392.8558, "encoder_q-layer.6": 409.33, "encoder_q-layer.7": 431.9549, "encoder_q-layer.8": 495.123, "encoder_q-layer.9": 445.9513, "epoch": 0.1, "inbatch_neg_score": 0.2136, "inbatch_pos_score": 0.7549, "learning_rate": 4.6447368421052635e-05, "loss": 3.9542, "norm_diff": 0.0789, "norm_loss": 0.0, "num_token_doc": 66.7855, "num_token_overlap": 11.7036, "num_token_query": 31.4278, "num_token_union": 65.1311, "num_word_context": 202.3877, "num_word_doc": 49.8412, "num_word_query": 23.3404, "postclip_grad_norm": 1.0, "preclip_grad_norm": 876.3955, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.214, "query_norm": 1.3011, "queue_k_norm": 1.381, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4278, "sent_len_1": 66.7855, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.2975, "stdk": 0.047, "stdq": 0.0428, "stdqueue_k": 0.047, "stdqueue_q": 0.0, "step": 23500 }, { "accuracy": 45.2637, "active_queue_size": 16384.0, "cl_loss": 3.9592, "doc_norm": 1.3832, "encoder_q-embeddings": 458.2129, "encoder_q-layer.0": 313.4966, "encoder_q-layer.1": 333.4073, "encoder_q-layer.10": 474.5751, "encoder_q-layer.11": 1246.1409, "encoder_q-layer.2": 378.2433, "encoder_q-layer.3": 391.0465, "encoder_q-layer.4": 402.4809, "encoder_q-layer.5": 389.9065, "encoder_q-layer.6": 438.1233, "encoder_q-layer.7": 428.7488, "encoder_q-layer.8": 459.6829, "encoder_q-layer.9": 417.9699, "epoch": 0.1, "inbatch_neg_score": 0.2027, "inbatch_pos_score": 0.7373, "learning_rate": 4.642105263157895e-05, "loss": 3.9592, "norm_diff": 0.0947, "norm_loss": 0.0, "num_token_doc": 66.4654, "num_token_overlap": 11.6403, "num_token_query": 31.3274, "num_token_union": 64.9477, "num_word_context": 202.0538, "num_word_doc": 49.5849, "num_word_query": 23.2699, "postclip_grad_norm": 1.0, "preclip_grad_norm": 809.8912, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2019, "query_norm": 1.2885, "queue_k_norm": 1.3826, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3274, "sent_len_1": 66.4654, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.5475, "stdk": 0.0471, "stdq": 0.0424, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 23600 }, { "accuracy": 45.5566, "active_queue_size": 16384.0, "cl_loss": 3.9604, "doc_norm": 1.3799, "encoder_q-embeddings": 470.1349, "encoder_q-layer.0": 315.8272, "encoder_q-layer.1": 336.7077, "encoder_q-layer.10": 473.3132, "encoder_q-layer.11": 1254.2482, "encoder_q-layer.2": 379.3166, "encoder_q-layer.3": 401.2919, "encoder_q-layer.4": 417.1395, "encoder_q-layer.5": 439.2184, "encoder_q-layer.6": 481.0352, "encoder_q-layer.7": 486.117, "encoder_q-layer.8": 476.4968, "encoder_q-layer.9": 419.8203, "epoch": 0.1, "inbatch_neg_score": 0.2012, "inbatch_pos_score": 0.75, "learning_rate": 4.6394736842105265e-05, "loss": 3.9604, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.5809, "num_token_overlap": 11.682, "num_token_query": 31.4476, "num_token_union": 65.0344, "num_word_context": 202.0639, "num_word_doc": 49.6965, "num_word_query": 23.3639, "postclip_grad_norm": 1.0, "preclip_grad_norm": 844.0211, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2009, "query_norm": 1.3055, "queue_k_norm": 1.3837, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4476, "sent_len_1": 66.5809, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.3413, "stdk": 0.047, "stdq": 0.043, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 23700 }, { "accuracy": 44.3848, "active_queue_size": 16384.0, "cl_loss": 3.9586, "doc_norm": 1.3819, "encoder_q-embeddings": 410.6508, "encoder_q-layer.0": 267.7379, "encoder_q-layer.1": 277.4492, "encoder_q-layer.10": 487.0483, "encoder_q-layer.11": 1197.6392, "encoder_q-layer.2": 300.6964, "encoder_q-layer.3": 312.0424, "encoder_q-layer.4": 338.0167, "encoder_q-layer.5": 344.441, "encoder_q-layer.6": 397.0326, "encoder_q-layer.7": 414.5733, "encoder_q-layer.8": 469.7, "encoder_q-layer.9": 415.3711, "epoch": 0.1, "inbatch_neg_score": 0.1948, "inbatch_pos_score": 0.73, "learning_rate": 4.6368421052631584e-05, "loss": 3.9586, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 67.1302, "num_token_overlap": 11.7218, "num_token_query": 31.403, "num_token_union": 65.3441, "num_word_context": 202.6848, "num_word_doc": 50.0906, "num_word_query": 23.3201, "postclip_grad_norm": 1.0, "preclip_grad_norm": 763.084, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1952, "query_norm": 1.31, "queue_k_norm": 1.3817, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.403, "sent_len_1": 67.1302, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.91, "stdk": 0.0471, "stdq": 0.043, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 23800 }, { "accuracy": 44.6777, "active_queue_size": 16384.0, "cl_loss": 3.9661, "doc_norm": 1.3874, "encoder_q-embeddings": 433.2526, "encoder_q-layer.0": 290.0998, "encoder_q-layer.1": 296.5546, "encoder_q-layer.10": 508.6073, "encoder_q-layer.11": 1301.901, "encoder_q-layer.2": 319.4274, "encoder_q-layer.3": 333.1525, "encoder_q-layer.4": 376.0369, "encoder_q-layer.5": 382.0076, "encoder_q-layer.6": 416.9037, "encoder_q-layer.7": 466.0414, "encoder_q-layer.8": 496.3148, "encoder_q-layer.9": 442.1878, "epoch": 0.1, "inbatch_neg_score": 0.1964, "inbatch_pos_score": 0.7324, "learning_rate": 4.6342105263157895e-05, "loss": 3.9661, "norm_diff": 0.0832, "norm_loss": 0.0, "num_token_doc": 66.7554, "num_token_overlap": 11.6499, "num_token_query": 31.3529, "num_token_union": 65.1186, "num_word_context": 202.1632, "num_word_doc": 49.8224, "num_word_query": 23.2822, "postclip_grad_norm": 1.0, "preclip_grad_norm": 844.6697, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1967, "query_norm": 1.3042, "queue_k_norm": 1.3801, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3529, "sent_len_1": 66.7554, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.6075, "stdk": 0.0473, "stdq": 0.0427, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 23900 }, { "accuracy": 44.2383, "active_queue_size": 16384.0, "cl_loss": 3.9467, "doc_norm": 1.3808, "encoder_q-embeddings": 575.5454, "encoder_q-layer.0": 387.6871, "encoder_q-layer.1": 436.5637, "encoder_q-layer.10": 518.1684, "encoder_q-layer.11": 1285.0637, "encoder_q-layer.2": 501.837, "encoder_q-layer.3": 543.213, "encoder_q-layer.4": 627.3862, "encoder_q-layer.5": 569.3096, "encoder_q-layer.6": 549.548, "encoder_q-layer.7": 523.0205, "encoder_q-layer.8": 577.4205, "encoder_q-layer.9": 463.3239, "epoch": 0.1, "inbatch_neg_score": 0.1992, "inbatch_pos_score": 0.73, "learning_rate": 4.6315789473684214e-05, "loss": 3.9467, "norm_diff": 0.0639, "norm_loss": 0.0, "num_token_doc": 66.8995, "num_token_overlap": 11.6948, "num_token_query": 31.438, "num_token_union": 65.2133, "num_word_context": 202.2733, "num_word_doc": 49.9194, "num_word_query": 23.3625, "postclip_grad_norm": 1.0, "preclip_grad_norm": 945.2875, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1992, "query_norm": 1.3169, "queue_k_norm": 1.3802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.438, "sent_len_1": 66.8995, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.935, "stdk": 0.047, "stdq": 0.043, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 24000 }, { "accuracy": 43.8965, "active_queue_size": 16384.0, "cl_loss": 3.9412, "doc_norm": 1.3821, "encoder_q-embeddings": 419.1386, "encoder_q-layer.0": 281.8195, "encoder_q-layer.1": 307.3056, "encoder_q-layer.10": 483.0849, "encoder_q-layer.11": 1292.5691, "encoder_q-layer.2": 342.7258, "encoder_q-layer.3": 345.8183, "encoder_q-layer.4": 360.8669, "encoder_q-layer.5": 365.6169, "encoder_q-layer.6": 395.4471, "encoder_q-layer.7": 415.7684, "encoder_q-layer.8": 493.7836, "encoder_q-layer.9": 437.0663, "epoch": 0.1, "inbatch_neg_score": 0.2023, "inbatch_pos_score": 0.7393, "learning_rate": 4.6289473684210525e-05, "loss": 3.9412, "norm_diff": 0.0632, "norm_loss": 0.0, "num_token_doc": 66.7251, "num_token_overlap": 11.7014, "num_token_query": 31.3935, "num_token_union": 65.0942, "num_word_context": 202.2694, "num_word_doc": 49.833, "num_word_query": 23.308, "postclip_grad_norm": 1.0, "preclip_grad_norm": 823.0626, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2021, "query_norm": 1.3189, "queue_k_norm": 1.3815, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3935, "sent_len_1": 66.7251, "sent_len_max_0": 127.9925, "sent_len_max_1": 206.09, "stdk": 0.0471, "stdq": 0.043, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 24100 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.9261, "doc_norm": 1.3823, "encoder_q-embeddings": 2153.6982, "encoder_q-layer.0": 1523.15, "encoder_q-layer.1": 1726.7325, "encoder_q-layer.10": 589.365, "encoder_q-layer.11": 1358.2925, "encoder_q-layer.2": 1971.0033, "encoder_q-layer.3": 1574.9398, "encoder_q-layer.4": 1351.2833, "encoder_q-layer.5": 968.8977, "encoder_q-layer.6": 837.9505, "encoder_q-layer.7": 899.1706, "encoder_q-layer.8": 979.9536, "encoder_q-layer.9": 687.1306, "epoch": 0.11, "inbatch_neg_score": 0.1962, "inbatch_pos_score": 0.7524, "learning_rate": 4.6263157894736844e-05, "loss": 3.9261, "norm_diff": 0.0635, "norm_loss": 0.0, "num_token_doc": 66.8461, "num_token_overlap": 11.6784, "num_token_query": 31.379, "num_token_union": 65.1742, "num_word_context": 202.2518, "num_word_doc": 49.8772, "num_word_query": 23.3001, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2159.5384, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.197, "query_norm": 1.3188, "queue_k_norm": 1.3812, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.379, "sent_len_1": 66.8461, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.0238, "stdk": 0.0472, "stdq": 0.0435, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 24200 }, { "accuracy": 45.459, "active_queue_size": 16384.0, "cl_loss": 3.9492, "doc_norm": 1.3806, "encoder_q-embeddings": 902.3877, "encoder_q-layer.0": 589.3226, "encoder_q-layer.1": 644.634, "encoder_q-layer.10": 937.3669, "encoder_q-layer.11": 2481.3516, "encoder_q-layer.2": 726.8474, "encoder_q-layer.3": 779.3361, "encoder_q-layer.4": 829.6958, "encoder_q-layer.5": 800.9423, "encoder_q-layer.6": 799.3784, "encoder_q-layer.7": 819.9165, "encoder_q-layer.8": 998.4172, "encoder_q-layer.9": 887.3995, "epoch": 0.11, "inbatch_neg_score": 0.2057, "inbatch_pos_score": 0.7666, "learning_rate": 4.623684210526316e-05, "loss": 3.9492, "norm_diff": 0.0594, "norm_loss": 0.0, "num_token_doc": 66.8802, "num_token_overlap": 11.6817, "num_token_query": 31.3031, "num_token_union": 65.1351, "num_word_context": 202.4208, "num_word_doc": 49.9116, "num_word_query": 23.2448, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1605.7465, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2068, "query_norm": 1.3213, "queue_k_norm": 1.3804, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3031, "sent_len_1": 66.8802, "sent_len_max_0": 127.9625, "sent_len_max_1": 207.28, "stdk": 0.0471, "stdq": 0.0436, "stdqueue_k": 0.0471, "stdqueue_q": 0.0, "step": 24300 }, { "accuracy": 44.4336, "active_queue_size": 16384.0, "cl_loss": 3.9396, "doc_norm": 1.383, "encoder_q-embeddings": 1220.0861, "encoder_q-layer.0": 808.5229, "encoder_q-layer.1": 966.66, "encoder_q-layer.10": 984.7592, "encoder_q-layer.11": 2528.6516, "encoder_q-layer.2": 1139.7728, "encoder_q-layer.3": 1278.0424, "encoder_q-layer.4": 1331.6852, "encoder_q-layer.5": 1450.3687, "encoder_q-layer.6": 1345.8956, "encoder_q-layer.7": 1329.3835, "encoder_q-layer.8": 1234.0233, "encoder_q-layer.9": 944.1099, "epoch": 0.11, "inbatch_neg_score": 0.2128, "inbatch_pos_score": 0.752, "learning_rate": 4.6210526315789473e-05, "loss": 3.9396, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 66.7054, "num_token_overlap": 11.6873, "num_token_query": 31.3768, "num_token_union": 65.0647, "num_word_context": 202.2071, "num_word_doc": 49.7674, "num_word_query": 23.3222, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2061.7901, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2133, "query_norm": 1.3002, "queue_k_norm": 1.3814, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3768, "sent_len_1": 66.7054, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.235, "stdk": 0.0472, "stdq": 0.0427, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24400 }, { "accuracy": 47.168, "active_queue_size": 16384.0, "cl_loss": 3.9286, "doc_norm": 1.3853, "encoder_q-embeddings": 971.9692, "encoder_q-layer.0": 653.0097, "encoder_q-layer.1": 691.5218, "encoder_q-layer.10": 1051.2888, "encoder_q-layer.11": 2585.3237, "encoder_q-layer.2": 782.6613, "encoder_q-layer.3": 820.2681, "encoder_q-layer.4": 922.4987, "encoder_q-layer.5": 877.6044, "encoder_q-layer.6": 1008.0175, "encoder_q-layer.7": 957.4943, "encoder_q-layer.8": 1055.4192, "encoder_q-layer.9": 934.4168, "epoch": 0.11, "inbatch_neg_score": 0.2068, "inbatch_pos_score": 0.752, "learning_rate": 4.618421052631579e-05, "loss": 3.9286, "norm_diff": 0.1083, "norm_loss": 0.0, "num_token_doc": 66.7073, "num_token_overlap": 11.6501, "num_token_query": 31.3036, "num_token_union": 65.0538, "num_word_context": 202.1694, "num_word_doc": 49.7868, "num_word_query": 23.2599, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1684.0294, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2069, "query_norm": 1.277, "queue_k_norm": 1.3825, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3036, "sent_len_1": 66.7073, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.875, "stdk": 0.0473, "stdq": 0.042, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24500 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.9343, "doc_norm": 1.3851, "encoder_q-embeddings": 785.551, "encoder_q-layer.0": 523.0787, "encoder_q-layer.1": 544.549, "encoder_q-layer.10": 970.905, "encoder_q-layer.11": 2397.3093, "encoder_q-layer.2": 604.5856, "encoder_q-layer.3": 624.8111, "encoder_q-layer.4": 645.4222, "encoder_q-layer.5": 678.3564, "encoder_q-layer.6": 722.2763, "encoder_q-layer.7": 850.9056, "encoder_q-layer.8": 977.9389, "encoder_q-layer.9": 880.6371, "epoch": 0.11, "inbatch_neg_score": 0.2059, "inbatch_pos_score": 0.7422, "learning_rate": 4.6157894736842103e-05, "loss": 3.9343, "norm_diff": 0.1067, "norm_loss": 0.0, "num_token_doc": 66.6666, "num_token_overlap": 11.6819, "num_token_query": 31.3283, "num_token_union": 65.0049, "num_word_context": 201.9124, "num_word_doc": 49.7253, "num_word_query": 23.2595, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1512.0403, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2062, "query_norm": 1.2784, "queue_k_norm": 1.3833, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3283, "sent_len_1": 66.6666, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.2287, "stdk": 0.0473, "stdq": 0.042, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24600 }, { "accuracy": 44.5801, "active_queue_size": 16384.0, "cl_loss": 3.931, "doc_norm": 1.3809, "encoder_q-embeddings": 7189.416, "encoder_q-layer.0": 4876.2739, "encoder_q-layer.1": 5921.5547, "encoder_q-layer.10": 944.5144, "encoder_q-layer.11": 2529.2832, "encoder_q-layer.2": 7430.4292, "encoder_q-layer.3": 8142.3213, "encoder_q-layer.4": 7999.606, "encoder_q-layer.5": 5911.5308, "encoder_q-layer.6": 3709.7312, "encoder_q-layer.7": 2730.0364, "encoder_q-layer.8": 1593.7775, "encoder_q-layer.9": 994.0895, "epoch": 0.11, "inbatch_neg_score": 0.21, "inbatch_pos_score": 0.7568, "learning_rate": 4.613157894736843e-05, "loss": 3.931, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.829, "num_token_overlap": 11.6799, "num_token_query": 31.3647, "num_token_union": 65.1203, "num_word_context": 202.5991, "num_word_doc": 49.8606, "num_word_query": 23.2873, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8049.8728, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2091, "query_norm": 1.3077, "queue_k_norm": 1.3817, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3647, "sent_len_1": 66.829, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.875, "stdk": 0.0471, "stdq": 0.0431, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24700 }, { "accuracy": 44.6289, "active_queue_size": 16384.0, "cl_loss": 3.9265, "doc_norm": 1.3813, "encoder_q-embeddings": 1249.0037, "encoder_q-layer.0": 865.4132, "encoder_q-layer.1": 922.8954, "encoder_q-layer.10": 991.1377, "encoder_q-layer.11": 2405.7947, "encoder_q-layer.2": 1009.1968, "encoder_q-layer.3": 1104.8771, "encoder_q-layer.4": 1049.6168, "encoder_q-layer.5": 1115.9049, "encoder_q-layer.6": 1189.5121, "encoder_q-layer.7": 969.4124, "encoder_q-layer.8": 971.2784, "encoder_q-layer.9": 836.322, "epoch": 0.11, "inbatch_neg_score": 0.2078, "inbatch_pos_score": 0.7466, "learning_rate": 4.610526315789474e-05, "loss": 3.9265, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7376, "num_token_overlap": 11.7095, "num_token_query": 31.462, "num_token_union": 65.1292, "num_word_context": 202.1311, "num_word_doc": 49.7909, "num_word_query": 23.3625, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1848.9083, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.208, "query_norm": 1.3017, "queue_k_norm": 1.3826, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.462, "sent_len_1": 66.7376, "sent_len_max_0": 127.995, "sent_len_max_1": 208.0125, "stdk": 0.0471, "stdq": 0.0427, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24800 }, { "accuracy": 45.4102, "active_queue_size": 16384.0, "cl_loss": 3.9203, "doc_norm": 1.3793, "encoder_q-embeddings": 959.7839, "encoder_q-layer.0": 643.1292, "encoder_q-layer.1": 675.9659, "encoder_q-layer.10": 942.4561, "encoder_q-layer.11": 2372.2014, "encoder_q-layer.2": 804.9452, "encoder_q-layer.3": 872.0398, "encoder_q-layer.4": 877.4159, "encoder_q-layer.5": 917.1202, "encoder_q-layer.6": 935.5795, "encoder_q-layer.7": 910.5636, "encoder_q-layer.8": 924.4496, "encoder_q-layer.9": 813.7699, "epoch": 0.11, "inbatch_neg_score": 0.2062, "inbatch_pos_score": 0.7549, "learning_rate": 4.607894736842105e-05, "loss": 3.9203, "norm_diff": 0.0691, "norm_loss": 0.0, "num_token_doc": 66.8281, "num_token_overlap": 11.6877, "num_token_query": 31.3833, "num_token_union": 65.132, "num_word_context": 202.2033, "num_word_doc": 49.8379, "num_word_query": 23.2983, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1607.5734, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2064, "query_norm": 1.3103, "queue_k_norm": 1.3823, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3833, "sent_len_1": 66.8281, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.1188, "stdk": 0.0471, "stdq": 0.043, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 24900 }, { "accuracy": 44.8242, "active_queue_size": 16384.0, "cl_loss": 3.925, "doc_norm": 1.3828, "encoder_q-embeddings": 1898.7686, "encoder_q-layer.0": 1363.729, "encoder_q-layer.1": 1436.6338, "encoder_q-layer.10": 1009.4399, "encoder_q-layer.11": 2370.0261, "encoder_q-layer.2": 1570.0846, "encoder_q-layer.3": 1458.5848, "encoder_q-layer.4": 1541.2789, "encoder_q-layer.5": 1339.9294, "encoder_q-layer.6": 1405.1686, "encoder_q-layer.7": 1192.5415, "encoder_q-layer.8": 1085.549, "encoder_q-layer.9": 886.7676, "epoch": 0.11, "inbatch_neg_score": 0.2107, "inbatch_pos_score": 0.7568, "learning_rate": 4.605263157894737e-05, "loss": 3.925, "norm_diff": 0.0531, "norm_loss": 0.0, "num_token_doc": 66.7963, "num_token_overlap": 11.7065, "num_token_query": 31.5028, "num_token_union": 65.2247, "num_word_context": 202.2836, "num_word_doc": 49.8383, "num_word_query": 23.4189, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2278.1332, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2109, "query_norm": 1.3298, "queue_k_norm": 1.3811, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5028, "sent_len_1": 66.7963, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.8063, "stdk": 0.0472, "stdq": 0.0434, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25000 }, { "accuracy": 45.459, "active_queue_size": 16384.0, "cl_loss": 3.9186, "doc_norm": 1.3882, "encoder_q-embeddings": 803.7178, "encoder_q-layer.0": 519.0836, "encoder_q-layer.1": 534.9113, "encoder_q-layer.10": 896.3745, "encoder_q-layer.11": 2321.8574, "encoder_q-layer.2": 580.3519, "encoder_q-layer.3": 615.6862, "encoder_q-layer.4": 633.5683, "encoder_q-layer.5": 629.7487, "encoder_q-layer.6": 697.2308, "encoder_q-layer.7": 766.963, "encoder_q-layer.8": 923.0839, "encoder_q-layer.9": 815.8109, "epoch": 0.11, "inbatch_neg_score": 0.2111, "inbatch_pos_score": 0.7686, "learning_rate": 4.602631578947368e-05, "loss": 3.9186, "norm_diff": 0.0586, "norm_loss": 0.0, "num_token_doc": 66.8696, "num_token_overlap": 11.712, "num_token_query": 31.4492, "num_token_union": 65.2098, "num_word_context": 202.6656, "num_word_doc": 49.8825, "num_word_query": 23.3748, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1478.1328, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2117, "query_norm": 1.3295, "queue_k_norm": 1.3826, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4492, "sent_len_1": 66.8696, "sent_len_max_0": 128.0, "sent_len_max_1": 209.0925, "stdk": 0.0474, "stdq": 0.0431, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25100 }, { "accuracy": 44.9219, "active_queue_size": 16384.0, "cl_loss": 3.9196, "doc_norm": 1.3756, "encoder_q-embeddings": 808.4509, "encoder_q-layer.0": 546.6625, "encoder_q-layer.1": 578.006, "encoder_q-layer.10": 1015.1192, "encoder_q-layer.11": 2601.1042, "encoder_q-layer.2": 645.7884, "encoder_q-layer.3": 681.6783, "encoder_q-layer.4": 711.905, "encoder_q-layer.5": 713.8618, "encoder_q-layer.6": 795.8074, "encoder_q-layer.7": 882.0034, "encoder_q-layer.8": 1054.4393, "encoder_q-layer.9": 970.4131, "epoch": 0.11, "inbatch_neg_score": 0.2222, "inbatch_pos_score": 0.7773, "learning_rate": 4.600000000000001e-05, "loss": 3.9196, "norm_diff": 0.032, "norm_loss": 0.0, "num_token_doc": 66.6245, "num_token_overlap": 11.6507, "num_token_query": 31.3825, "num_token_union": 65.0714, "num_word_context": 202.3486, "num_word_doc": 49.7512, "num_word_query": 23.3036, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1635.4471, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2228, "query_norm": 1.3437, "queue_k_norm": 1.384, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3825, "sent_len_1": 66.6245, "sent_len_max_0": 127.995, "sent_len_max_1": 208.32, "stdk": 0.0469, "stdq": 0.0436, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 25200 }, { "accuracy": 45.8496, "active_queue_size": 16384.0, "cl_loss": 3.9093, "doc_norm": 1.3811, "encoder_q-embeddings": 902.0591, "encoder_q-layer.0": 596.9617, "encoder_q-layer.1": 615.2878, "encoder_q-layer.10": 929.9505, "encoder_q-layer.11": 2469.4751, "encoder_q-layer.2": 673.6621, "encoder_q-layer.3": 710.4841, "encoder_q-layer.4": 729.9951, "encoder_q-layer.5": 717.0919, "encoder_q-layer.6": 829.8646, "encoder_q-layer.7": 850.9764, "encoder_q-layer.8": 1008.7498, "encoder_q-layer.9": 866.14, "epoch": 0.11, "inbatch_neg_score": 0.218, "inbatch_pos_score": 0.7681, "learning_rate": 4.597368421052632e-05, "loss": 3.9093, "norm_diff": 0.0588, "norm_loss": 0.0, "num_token_doc": 66.9028, "num_token_overlap": 11.6987, "num_token_query": 31.4015, "num_token_union": 65.1882, "num_word_context": 202.4823, "num_word_doc": 49.9267, "num_word_query": 23.3221, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1605.3548, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2178, "query_norm": 1.3223, "queue_k_norm": 1.3821, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4015, "sent_len_1": 66.9028, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.335, "stdk": 0.0471, "stdq": 0.0431, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25300 }, { "accuracy": 45.1172, "active_queue_size": 16384.0, "cl_loss": 3.917, "doc_norm": 1.3802, "encoder_q-embeddings": 1615.6678, "encoder_q-layer.0": 1093.3199, "encoder_q-layer.1": 1185.8475, "encoder_q-layer.10": 916.0598, "encoder_q-layer.11": 2463.6638, "encoder_q-layer.2": 1113.3348, "encoder_q-layer.3": 1195.455, "encoder_q-layer.4": 1178.7223, "encoder_q-layer.5": 1099.8712, "encoder_q-layer.6": 1260.9938, "encoder_q-layer.7": 1078.762, "encoder_q-layer.8": 1114.9652, "encoder_q-layer.9": 899.4346, "epoch": 0.11, "inbatch_neg_score": 0.2215, "inbatch_pos_score": 0.7646, "learning_rate": 4.594736842105264e-05, "loss": 3.917, "norm_diff": 0.0651, "norm_loss": 0.0, "num_token_doc": 66.7045, "num_token_overlap": 11.6625, "num_token_query": 31.3849, "num_token_union": 65.1142, "num_word_context": 201.941, "num_word_doc": 49.7743, "num_word_query": 23.3103, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2036.6538, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2211, "query_norm": 1.3151, "queue_k_norm": 1.3833, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3849, "sent_len_1": 66.7045, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.9725, "stdk": 0.0471, "stdq": 0.0429, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25400 }, { "accuracy": 46.5332, "active_queue_size": 16384.0, "cl_loss": 3.9111, "doc_norm": 1.3846, "encoder_q-embeddings": 1095.1422, "encoder_q-layer.0": 759.6048, "encoder_q-layer.1": 750.6917, "encoder_q-layer.10": 898.735, "encoder_q-layer.11": 2277.7312, "encoder_q-layer.2": 838.1682, "encoder_q-layer.3": 782.1965, "encoder_q-layer.4": 784.9915, "encoder_q-layer.5": 746.9537, "encoder_q-layer.6": 868.7855, "encoder_q-layer.7": 901.1422, "encoder_q-layer.8": 872.8975, "encoder_q-layer.9": 794.4095, "epoch": 0.11, "inbatch_neg_score": 0.2174, "inbatch_pos_score": 0.7666, "learning_rate": 4.592105263157895e-05, "loss": 3.9111, "norm_diff": 0.0828, "norm_loss": 0.0, "num_token_doc": 66.6901, "num_token_overlap": 11.6827, "num_token_query": 31.3575, "num_token_union": 65.0293, "num_word_context": 202.0075, "num_word_doc": 49.7495, "num_word_query": 23.2834, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1614.8018, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2179, "query_norm": 1.3017, "queue_k_norm": 1.3831, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3575, "sent_len_1": 66.6901, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.3288, "stdk": 0.0473, "stdq": 0.0427, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25500 }, { "accuracy": 45.4102, "active_queue_size": 16384.0, "cl_loss": 3.9259, "doc_norm": 1.3841, "encoder_q-embeddings": 818.5169, "encoder_q-layer.0": 532.5438, "encoder_q-layer.1": 568.9141, "encoder_q-layer.10": 1028.1326, "encoder_q-layer.11": 2520.5132, "encoder_q-layer.2": 640.3715, "encoder_q-layer.3": 662.8433, "encoder_q-layer.4": 677.6097, "encoder_q-layer.5": 679.7374, "encoder_q-layer.6": 772.0298, "encoder_q-layer.7": 818.4521, "encoder_q-layer.8": 1023.2676, "encoder_q-layer.9": 903.3578, "epoch": 0.11, "inbatch_neg_score": 0.2128, "inbatch_pos_score": 0.7598, "learning_rate": 4.589473684210526e-05, "loss": 3.9259, "norm_diff": 0.081, "norm_loss": 0.0, "num_token_doc": 66.8485, "num_token_overlap": 11.695, "num_token_query": 31.3808, "num_token_union": 65.144, "num_word_context": 202.3894, "num_word_doc": 49.8633, "num_word_query": 23.2959, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1569.6696, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2125, "query_norm": 1.3032, "queue_k_norm": 1.3852, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3808, "sent_len_1": 66.8485, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.5588, "stdk": 0.0472, "stdq": 0.0429, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 25600 }, { "accuracy": 46.0938, "active_queue_size": 16384.0, "cl_loss": 3.9127, "doc_norm": 1.3885, "encoder_q-embeddings": 802.4913, "encoder_q-layer.0": 534.7788, "encoder_q-layer.1": 559.9259, "encoder_q-layer.10": 959.5114, "encoder_q-layer.11": 2425.0898, "encoder_q-layer.2": 623.5104, "encoder_q-layer.3": 653.2607, "encoder_q-layer.4": 711.3623, "encoder_q-layer.5": 725.6348, "encoder_q-layer.6": 784.0021, "encoder_q-layer.7": 832.0309, "encoder_q-layer.8": 946.9066, "encoder_q-layer.9": 867.8762, "epoch": 0.11, "inbatch_neg_score": 0.2134, "inbatch_pos_score": 0.7715, "learning_rate": 4.5868421052631585e-05, "loss": 3.9127, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.882, "num_token_overlap": 11.6633, "num_token_query": 31.37, "num_token_union": 65.1723, "num_word_context": 202.5013, "num_word_doc": 49.9167, "num_word_query": 23.304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1519.2739, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2139, "query_norm": 1.3054, "queue_k_norm": 1.3832, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.37, "sent_len_1": 66.882, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.8725, "stdk": 0.0474, "stdq": 0.0428, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 25700 }, { "accuracy": 45.1172, "active_queue_size": 16384.0, "cl_loss": 3.9152, "doc_norm": 1.3841, "encoder_q-embeddings": 1561.8196, "encoder_q-layer.0": 1125.6497, "encoder_q-layer.1": 1155.5381, "encoder_q-layer.10": 925.6158, "encoder_q-layer.11": 2459.1423, "encoder_q-layer.2": 1233.2399, "encoder_q-layer.3": 1255.0514, "encoder_q-layer.4": 1260.6453, "encoder_q-layer.5": 1163.7737, "encoder_q-layer.6": 1132.5544, "encoder_q-layer.7": 1054.0236, "encoder_q-layer.8": 960.5087, "encoder_q-layer.9": 843.3105, "epoch": 0.11, "inbatch_neg_score": 0.2165, "inbatch_pos_score": 0.7651, "learning_rate": 4.58421052631579e-05, "loss": 3.9152, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.882, "num_token_overlap": 11.7012, "num_token_query": 31.4248, "num_token_union": 65.2006, "num_word_context": 202.31, "num_word_doc": 49.9086, "num_word_query": 23.3352, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2028.2107, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2167, "query_norm": 1.3046, "queue_k_norm": 1.3845, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4248, "sent_len_1": 66.882, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.765, "stdk": 0.0472, "stdq": 0.0425, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 25800 }, { "accuracy": 46.0938, "active_queue_size": 16384.0, "cl_loss": 3.9008, "doc_norm": 1.3884, "encoder_q-embeddings": 1381.2098, "encoder_q-layer.0": 1050.2478, "encoder_q-layer.1": 1046.0604, "encoder_q-layer.10": 917.6259, "encoder_q-layer.11": 2246.7476, "encoder_q-layer.2": 1147.3804, "encoder_q-layer.3": 1199.3792, "encoder_q-layer.4": 1244.3717, "encoder_q-layer.5": 1075.855, "encoder_q-layer.6": 1135.7936, "encoder_q-layer.7": 1000.1476, "encoder_q-layer.8": 1006.1265, "encoder_q-layer.9": 857.668, "epoch": 0.11, "inbatch_neg_score": 0.2146, "inbatch_pos_score": 0.7676, "learning_rate": 4.5815789473684215e-05, "loss": 3.9008, "norm_diff": 0.0707, "norm_loss": 0.0, "num_token_doc": 66.7523, "num_token_overlap": 11.6741, "num_token_query": 31.3681, "num_token_union": 65.1019, "num_word_context": 202.2082, "num_word_doc": 49.7998, "num_word_query": 23.3039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1873.547, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2142, "query_norm": 1.3176, "queue_k_norm": 1.3871, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3681, "sent_len_1": 66.7523, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.4175, "stdk": 0.0474, "stdq": 0.0429, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 25900 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.8969, "doc_norm": 1.3844, "encoder_q-embeddings": 863.0045, "encoder_q-layer.0": 532.3295, "encoder_q-layer.1": 549.6516, "encoder_q-layer.10": 921.4032, "encoder_q-layer.11": 2434.5488, "encoder_q-layer.2": 597.2752, "encoder_q-layer.3": 607.6912, "encoder_q-layer.4": 626.4806, "encoder_q-layer.5": 642.5186, "encoder_q-layer.6": 688.2863, "encoder_q-layer.7": 781.1033, "encoder_q-layer.8": 957.431, "encoder_q-layer.9": 857.6194, "epoch": 0.11, "inbatch_neg_score": 0.2141, "inbatch_pos_score": 0.771, "learning_rate": 4.5789473684210527e-05, "loss": 3.8969, "norm_diff": 0.0551, "norm_loss": 0.0, "num_token_doc": 66.8954, "num_token_overlap": 11.6862, "num_token_query": 31.4098, "num_token_union": 65.2029, "num_word_context": 202.4645, "num_word_doc": 49.9305, "num_word_query": 23.3348, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1502.788, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2146, "query_norm": 1.3293, "queue_k_norm": 1.3868, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4098, "sent_len_1": 66.8954, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.2663, "stdk": 0.0472, "stdq": 0.0431, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26000 }, { "accuracy": 47.0703, "active_queue_size": 16384.0, "cl_loss": 3.8912, "doc_norm": 1.3874, "encoder_q-embeddings": 816.8606, "encoder_q-layer.0": 544.3861, "encoder_q-layer.1": 555.8706, "encoder_q-layer.10": 921.3732, "encoder_q-layer.11": 2424.1907, "encoder_q-layer.2": 607.693, "encoder_q-layer.3": 650.3313, "encoder_q-layer.4": 677.6418, "encoder_q-layer.5": 679.2422, "encoder_q-layer.6": 759.0723, "encoder_q-layer.7": 830.8535, "encoder_q-layer.8": 985.262, "encoder_q-layer.9": 857.5377, "epoch": 0.11, "inbatch_neg_score": 0.2085, "inbatch_pos_score": 0.7754, "learning_rate": 4.5763157894736845e-05, "loss": 3.8912, "norm_diff": 0.0515, "norm_loss": 0.0, "num_token_doc": 66.7824, "num_token_overlap": 11.6516, "num_token_query": 31.3083, "num_token_union": 65.12, "num_word_context": 202.3964, "num_word_doc": 49.8382, "num_word_query": 23.252, "postclip_grad_norm": 1.0, "preclip_grad_norm": 1521.4864, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2092, "query_norm": 1.3359, "queue_k_norm": 1.3862, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3083, "sent_len_1": 66.7824, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.5087, "stdk": 0.0473, "stdq": 0.0431, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26100 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.8973, "doc_norm": 1.3847, "encoder_q-embeddings": 1681.8669, "encoder_q-layer.0": 1264.6896, "encoder_q-layer.1": 1166.2216, "encoder_q-layer.10": 933.5476, "encoder_q-layer.11": 2314.0002, "encoder_q-layer.2": 1555.493, "encoder_q-layer.3": 1439.7462, "encoder_q-layer.4": 1467.2651, "encoder_q-layer.5": 1343.1833, "encoder_q-layer.6": 1428.1166, "encoder_q-layer.7": 1455.5323, "encoder_q-layer.8": 1205.5272, "encoder_q-layer.9": 889.9179, "epoch": 0.11, "inbatch_neg_score": 0.2114, "inbatch_pos_score": 0.7471, "learning_rate": 4.573684210526316e-05, "loss": 3.8973, "norm_diff": 0.0366, "norm_loss": 0.0, "num_token_doc": 66.8879, "num_token_overlap": 11.6432, "num_token_query": 31.2916, "num_token_union": 65.1778, "num_word_context": 202.5869, "num_word_doc": 49.9388, "num_word_query": 23.2478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2183.3483, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2111, "query_norm": 1.348, "queue_k_norm": 1.3835, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2916, "sent_len_1": 66.8879, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.2063, "stdk": 0.0472, "stdq": 0.0427, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 26200 }, { "accuracy": 44.4824, "active_queue_size": 16384.0, "cl_loss": 3.8999, "doc_norm": 1.3872, "encoder_q-embeddings": 1798.4291, "encoder_q-layer.0": 1185.9178, "encoder_q-layer.1": 1235.8281, "encoder_q-layer.10": 1888.1993, "encoder_q-layer.11": 4396.7642, "encoder_q-layer.2": 1369.7896, "encoder_q-layer.3": 1431.8459, "encoder_q-layer.4": 1555.8848, "encoder_q-layer.5": 1597.5763, "encoder_q-layer.6": 1738.4158, "encoder_q-layer.7": 1833.694, "encoder_q-layer.8": 2098.5066, "encoder_q-layer.9": 1845.7261, "epoch": 0.11, "inbatch_neg_score": 0.2156, "inbatch_pos_score": 0.7598, "learning_rate": 4.5710526315789475e-05, "loss": 3.8999, "norm_diff": 0.0295, "norm_loss": 0.0, "num_token_doc": 66.7024, "num_token_overlap": 11.676, "num_token_query": 31.3606, "num_token_union": 65.0573, "num_word_context": 202.1611, "num_word_doc": 49.7444, "num_word_query": 23.2937, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3018.5525, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2161, "query_norm": 1.3577, "queue_k_norm": 1.3843, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3606, "sent_len_1": 66.7024, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.6763, "stdk": 0.0473, "stdq": 0.043, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 26300 }, { "accuracy": 45.5078, "active_queue_size": 16384.0, "cl_loss": 3.9039, "doc_norm": 1.393, "encoder_q-embeddings": 1765.999, "encoder_q-layer.0": 1156.595, "encoder_q-layer.1": 1257.325, "encoder_q-layer.10": 1872.8666, "encoder_q-layer.11": 4827.0049, "encoder_q-layer.2": 1486.8765, "encoder_q-layer.3": 1633.4403, "encoder_q-layer.4": 1792.012, "encoder_q-layer.5": 1716.3558, "encoder_q-layer.6": 1766.1256, "encoder_q-layer.7": 2117.8008, "encoder_q-layer.8": 2348.8877, "encoder_q-layer.9": 1896.2635, "epoch": 0.11, "inbatch_neg_score": 0.2256, "inbatch_pos_score": 0.7705, "learning_rate": 4.568421052631579e-05, "loss": 3.9039, "norm_diff": 0.0369, "norm_loss": 0.0, "num_token_doc": 66.7046, "num_token_overlap": 11.6375, "num_token_query": 31.3073, "num_token_union": 65.0806, "num_word_context": 202.3338, "num_word_doc": 49.8056, "num_word_query": 23.2623, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3269.8197, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2251, "query_norm": 1.3561, "queue_k_norm": 1.3861, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3073, "sent_len_1": 66.7046, "sent_len_max_0": 127.97, "sent_len_max_1": 206.9062, "stdk": 0.0475, "stdq": 0.0434, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26400 }, { "accuracy": 46.1426, "active_queue_size": 16384.0, "cl_loss": 3.8916, "doc_norm": 1.3922, "encoder_q-embeddings": 1701.994, "encoder_q-layer.0": 1073.1628, "encoder_q-layer.1": 1122.3273, "encoder_q-layer.10": 1944.4791, "encoder_q-layer.11": 4918.0923, "encoder_q-layer.2": 1263.4849, "encoder_q-layer.3": 1325.5345, "encoder_q-layer.4": 1404.9042, "encoder_q-layer.5": 1464.3796, "encoder_q-layer.6": 1659.0076, "encoder_q-layer.7": 1833.8721, "encoder_q-layer.8": 2217.7285, "encoder_q-layer.9": 1904.8936, "epoch": 0.11, "inbatch_neg_score": 0.2253, "inbatch_pos_score": 0.7842, "learning_rate": 4.5657894736842105e-05, "loss": 3.8916, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.678, "num_token_overlap": 11.6622, "num_token_query": 31.3051, "num_token_union": 65.0061, "num_word_context": 202.2645, "num_word_doc": 49.779, "num_word_query": 23.2611, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3136.0479, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2255, "query_norm": 1.3267, "queue_k_norm": 1.3865, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3051, "sent_len_1": 66.678, "sent_len_max_0": 127.99, "sent_len_max_1": 210.9888, "stdk": 0.0475, "stdq": 0.0429, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26500 }, { "accuracy": 46.582, "active_queue_size": 16384.0, "cl_loss": 3.8807, "doc_norm": 1.3839, "encoder_q-embeddings": 1547.269, "encoder_q-layer.0": 1029.871, "encoder_q-layer.1": 1096.983, "encoder_q-layer.10": 1843.5189, "encoder_q-layer.11": 4821.2524, "encoder_q-layer.2": 1185.9474, "encoder_q-layer.3": 1232.0857, "encoder_q-layer.4": 1344.7322, "encoder_q-layer.5": 1371.01, "encoder_q-layer.6": 1560.3165, "encoder_q-layer.7": 1762.1951, "encoder_q-layer.8": 1973.4246, "encoder_q-layer.9": 1763.9546, "epoch": 0.12, "inbatch_neg_score": 0.223, "inbatch_pos_score": 0.7788, "learning_rate": 4.563157894736842e-05, "loss": 3.8807, "norm_diff": 0.086, "norm_loss": 0.0, "num_token_doc": 66.9041, "num_token_overlap": 11.6703, "num_token_query": 31.3194, "num_token_union": 65.1664, "num_word_context": 201.9893, "num_word_doc": 49.8983, "num_word_query": 23.2467, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3018.5197, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2241, "query_norm": 1.2978, "queue_k_norm": 1.388, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3194, "sent_len_1": 66.9041, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.165, "stdk": 0.0472, "stdq": 0.0421, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26600 }, { "accuracy": 47.168, "active_queue_size": 16384.0, "cl_loss": 3.8881, "doc_norm": 1.389, "encoder_q-embeddings": 1534.9777, "encoder_q-layer.0": 966.5632, "encoder_q-layer.1": 1001.4577, "encoder_q-layer.10": 2041.2424, "encoder_q-layer.11": 4931.9062, "encoder_q-layer.2": 1133.8435, "encoder_q-layer.3": 1220.528, "encoder_q-layer.4": 1249.8658, "encoder_q-layer.5": 1310.3379, "encoder_q-layer.6": 1475.3959, "encoder_q-layer.7": 1746.0109, "encoder_q-layer.8": 2138.7263, "encoder_q-layer.9": 1941.0465, "epoch": 0.12, "inbatch_neg_score": 0.222, "inbatch_pos_score": 0.7852, "learning_rate": 4.560526315789474e-05, "loss": 3.8881, "norm_diff": 0.0855, "norm_loss": 0.0, "num_token_doc": 66.9454, "num_token_overlap": 11.6971, "num_token_query": 31.456, "num_token_union": 65.2361, "num_word_context": 202.546, "num_word_doc": 49.9428, "num_word_query": 23.3727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3048.2681, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2218, "query_norm": 1.3035, "queue_k_norm": 1.386, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.456, "sent_len_1": 66.9454, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.2375, "stdk": 0.0473, "stdq": 0.0426, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26700 }, { "accuracy": 44.4824, "active_queue_size": 16384.0, "cl_loss": 3.8769, "doc_norm": 1.3888, "encoder_q-embeddings": 1957.6473, "encoder_q-layer.0": 1287.6888, "encoder_q-layer.1": 1352.2765, "encoder_q-layer.10": 1870.6766, "encoder_q-layer.11": 5068.8081, "encoder_q-layer.2": 1516.1401, "encoder_q-layer.3": 1638.5846, "encoder_q-layer.4": 1611.6012, "encoder_q-layer.5": 1593.6949, "encoder_q-layer.6": 1702.9255, "encoder_q-layer.7": 1785.2488, "encoder_q-layer.8": 2029.5105, "encoder_q-layer.9": 1759.2488, "epoch": 0.12, "inbatch_neg_score": 0.2222, "inbatch_pos_score": 0.7524, "learning_rate": 4.557894736842105e-05, "loss": 3.8769, "norm_diff": 0.1, "norm_loss": 0.0, "num_token_doc": 66.6608, "num_token_overlap": 11.67, "num_token_query": 31.3386, "num_token_union": 65.0528, "num_word_context": 202.2755, "num_word_doc": 49.7443, "num_word_query": 23.2897, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3366.2254, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2213, "query_norm": 1.2888, "queue_k_norm": 1.3881, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3386, "sent_len_1": 66.6608, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.5563, "stdk": 0.0473, "stdq": 0.042, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26800 }, { "accuracy": 45.459, "active_queue_size": 16384.0, "cl_loss": 3.8771, "doc_norm": 1.3896, "encoder_q-embeddings": 2160.1255, "encoder_q-layer.0": 1555.4241, "encoder_q-layer.1": 1516.8036, "encoder_q-layer.10": 1941.0931, "encoder_q-layer.11": 5223.0005, "encoder_q-layer.2": 1809.2347, "encoder_q-layer.3": 1776.2214, "encoder_q-layer.4": 1840.407, "encoder_q-layer.5": 1943.8702, "encoder_q-layer.6": 2067.6987, "encoder_q-layer.7": 2044.2721, "encoder_q-layer.8": 2017.3922, "encoder_q-layer.9": 1675.5438, "epoch": 0.12, "inbatch_neg_score": 0.221, "inbatch_pos_score": 0.769, "learning_rate": 4.555263157894737e-05, "loss": 3.8771, "norm_diff": 0.0922, "norm_loss": 0.0, "num_token_doc": 66.8772, "num_token_overlap": 11.7132, "num_token_query": 31.3648, "num_token_union": 65.1446, "num_word_context": 202.3715, "num_word_doc": 49.9217, "num_word_query": 23.2994, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3620.9386, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2205, "query_norm": 1.2974, "queue_k_norm": 1.3876, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3648, "sent_len_1": 66.8772, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.1375, "stdk": 0.0473, "stdq": 0.0424, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 26900 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.8852, "doc_norm": 1.3931, "encoder_q-embeddings": 1575.6897, "encoder_q-layer.0": 1008.3804, "encoder_q-layer.1": 1058.3582, "encoder_q-layer.10": 2105.0596, "encoder_q-layer.11": 5023.3745, "encoder_q-layer.2": 1188.8577, "encoder_q-layer.3": 1272.3539, "encoder_q-layer.4": 1333.6094, "encoder_q-layer.5": 1343.5135, "encoder_q-layer.6": 1582.9546, "encoder_q-layer.7": 1721.0518, "encoder_q-layer.8": 2006.8182, "encoder_q-layer.9": 1834.1136, "epoch": 0.12, "inbatch_neg_score": 0.2107, "inbatch_pos_score": 0.749, "learning_rate": 4.552631578947369e-05, "loss": 3.8852, "norm_diff": 0.1092, "norm_loss": 0.0, "num_token_doc": 66.8076, "num_token_overlap": 11.6771, "num_token_query": 31.3521, "num_token_union": 65.1397, "num_word_context": 202.1491, "num_word_doc": 49.8588, "num_word_query": 23.2833, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3055.9636, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2111, "query_norm": 1.2839, "queue_k_norm": 1.389, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3521, "sent_len_1": 66.8076, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.3013, "stdk": 0.0474, "stdq": 0.042, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 27000 }, { "accuracy": 45.3125, "active_queue_size": 16384.0, "cl_loss": 3.8742, "doc_norm": 1.3912, "encoder_q-embeddings": 2218.1887, "encoder_q-layer.0": 1529.3163, "encoder_q-layer.1": 1723.377, "encoder_q-layer.10": 1984.8555, "encoder_q-layer.11": 5207.4116, "encoder_q-layer.2": 1929.72, "encoder_q-layer.3": 2144.1792, "encoder_q-layer.4": 2266.4551, "encoder_q-layer.5": 2169.3257, "encoder_q-layer.6": 2077.3218, "encoder_q-layer.7": 1940.8749, "encoder_q-layer.8": 2206.499, "encoder_q-layer.9": 1872.1279, "epoch": 0.12, "inbatch_neg_score": 0.2045, "inbatch_pos_score": 0.7446, "learning_rate": 4.55e-05, "loss": 3.8742, "norm_diff": 0.1012, "norm_loss": 0.0, "num_token_doc": 66.9366, "num_token_overlap": 11.7077, "num_token_query": 31.404, "num_token_union": 65.2025, "num_word_context": 202.3602, "num_word_doc": 49.9413, "num_word_query": 23.336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3708.478, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2047, "query_norm": 1.2899, "queue_k_norm": 1.3891, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.404, "sent_len_1": 66.9366, "sent_len_max_0": 127.97, "sent_len_max_1": 208.8212, "stdk": 0.0473, "stdq": 0.0423, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 27100 }, { "accuracy": 44.1406, "active_queue_size": 16384.0, "cl_loss": 3.8677, "doc_norm": 1.3848, "encoder_q-embeddings": 1711.1262, "encoder_q-layer.0": 1192.9957, "encoder_q-layer.1": 1231.3008, "encoder_q-layer.10": 1880.0571, "encoder_q-layer.11": 4915.2471, "encoder_q-layer.2": 1367.2883, "encoder_q-layer.3": 1407.1869, "encoder_q-layer.4": 1561.1761, "encoder_q-layer.5": 1586.9149, "encoder_q-layer.6": 1751.5488, "encoder_q-layer.7": 1898.7693, "encoder_q-layer.8": 2082.2358, "encoder_q-layer.9": 1793.8525, "epoch": 0.12, "inbatch_neg_score": 0.1924, "inbatch_pos_score": 0.7285, "learning_rate": 4.547368421052632e-05, "loss": 3.8677, "norm_diff": 0.0746, "norm_loss": 0.0, "num_token_doc": 66.7564, "num_token_overlap": 11.6876, "num_token_query": 31.3942, "num_token_union": 65.1237, "num_word_context": 202.4867, "num_word_doc": 49.8315, "num_word_query": 23.3315, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3224.6431, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1915, "query_norm": 1.3102, "queue_k_norm": 1.3893, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3942, "sent_len_1": 66.7564, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.3187, "stdk": 0.0471, "stdq": 0.0427, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 27200 }, { "accuracy": 43.6035, "active_queue_size": 16384.0, "cl_loss": 3.8909, "doc_norm": 1.3912, "encoder_q-embeddings": 3382.1123, "encoder_q-layer.0": 2476.0242, "encoder_q-layer.1": 2293.8525, "encoder_q-layer.10": 1839.5281, "encoder_q-layer.11": 4882.0581, "encoder_q-layer.2": 2375.186, "encoder_q-layer.3": 2509.7646, "encoder_q-layer.4": 2374.5015, "encoder_q-layer.5": 2322.5344, "encoder_q-layer.6": 2798.5618, "encoder_q-layer.7": 2743.7524, "encoder_q-layer.8": 2000.2687, "encoder_q-layer.9": 1715.6012, "epoch": 0.12, "inbatch_neg_score": 0.1758, "inbatch_pos_score": 0.709, "learning_rate": 4.544736842105263e-05, "loss": 3.8909, "norm_diff": 0.0925, "norm_loss": 0.0, "num_token_doc": 66.6426, "num_token_overlap": 11.6698, "num_token_query": 31.3416, "num_token_union": 65.0523, "num_word_context": 201.8349, "num_word_doc": 49.7346, "num_word_query": 23.2783, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4236.9857, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.176, "query_norm": 1.2987, "queue_k_norm": 1.3872, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3416, "sent_len_1": 66.6426, "sent_len_max_0": 127.985, "sent_len_max_1": 206.6587, "stdk": 0.0473, "stdq": 0.0424, "stdqueue_k": 0.0472, "stdqueue_q": 0.0, "step": 27300 }, { "accuracy": 45.5078, "active_queue_size": 16384.0, "cl_loss": 3.8795, "doc_norm": 1.3867, "encoder_q-embeddings": 1608.7731, "encoder_q-layer.0": 1099.5975, "encoder_q-layer.1": 1134.0613, "encoder_q-layer.10": 2044.9752, "encoder_q-layer.11": 4999.2134, "encoder_q-layer.2": 1231.6412, "encoder_q-layer.3": 1261.7583, "encoder_q-layer.4": 1358.517, "encoder_q-layer.5": 1431.6161, "encoder_q-layer.6": 1526.4514, "encoder_q-layer.7": 1681.7852, "encoder_q-layer.8": 2007.3231, "encoder_q-layer.9": 1753.3142, "epoch": 0.12, "inbatch_neg_score": 0.1666, "inbatch_pos_score": 0.7139, "learning_rate": 4.542105263157895e-05, "loss": 3.8795, "norm_diff": 0.0657, "norm_loss": 0.0, "num_token_doc": 66.716, "num_token_overlap": 11.7032, "num_token_query": 31.4244, "num_token_union": 65.1191, "num_word_context": 202.4673, "num_word_doc": 49.7766, "num_word_query": 23.3231, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3161.189, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1666, "query_norm": 1.321, "queue_k_norm": 1.3885, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4244, "sent_len_1": 66.716, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.1488, "stdk": 0.0472, "stdq": 0.0431, "stdqueue_k": 0.0473, "stdqueue_q": 0.0, "step": 27400 }, { "accuracy": 45.7031, "active_queue_size": 16384.0, "cl_loss": 3.8789, "doc_norm": 1.3928, "encoder_q-embeddings": 5341.2793, "encoder_q-layer.0": 3876.6467, "encoder_q-layer.1": 4034.8499, "encoder_q-layer.10": 1921.8202, "encoder_q-layer.11": 4873.5073, "encoder_q-layer.2": 4576.9736, "encoder_q-layer.3": 4560.7275, "encoder_q-layer.4": 5172.0049, "encoder_q-layer.5": 5156.8994, "encoder_q-layer.6": 5356.8218, "encoder_q-layer.7": 5009.3057, "encoder_q-layer.8": 5084.4731, "encoder_q-layer.9": 2512.9202, "epoch": 0.12, "inbatch_neg_score": 0.1542, "inbatch_pos_score": 0.6982, "learning_rate": 4.539473684210527e-05, "loss": 3.8789, "norm_diff": 0.0693, "norm_loss": 0.0, "num_token_doc": 66.8241, "num_token_overlap": 11.7045, "num_token_query": 31.4106, "num_token_union": 65.1394, "num_word_context": 202.407, "num_word_doc": 49.8692, "num_word_query": 23.3363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6879.9161, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1548, "query_norm": 1.3235, "queue_k_norm": 1.3919, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4106, "sent_len_1": 66.8241, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.1525, "stdk": 0.0474, "stdq": 0.0425, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 27500 }, { "accuracy": 47.2168, "active_queue_size": 16384.0, "cl_loss": 3.8605, "doc_norm": 1.3899, "encoder_q-embeddings": 1501.1927, "encoder_q-layer.0": 995.7547, "encoder_q-layer.1": 1022.7871, "encoder_q-layer.10": 2069.9407, "encoder_q-layer.11": 5077.0688, "encoder_q-layer.2": 1142.0969, "encoder_q-layer.3": 1223.2726, "encoder_q-layer.4": 1254.6464, "encoder_q-layer.5": 1291.9301, "encoder_q-layer.6": 1437.3129, "encoder_q-layer.7": 1642.7753, "encoder_q-layer.8": 1940.1677, "encoder_q-layer.9": 1844.6764, "epoch": 0.12, "inbatch_neg_score": 0.1535, "inbatch_pos_score": 0.7056, "learning_rate": 4.536842105263158e-05, "loss": 3.8605, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.8412, "num_token_overlap": 11.727, "num_token_query": 31.4916, "num_token_union": 65.2122, "num_word_context": 202.2804, "num_word_doc": 49.8674, "num_word_query": 23.3994, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3140.7603, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1536, "query_norm": 1.3177, "queue_k_norm": 1.3912, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4916, "sent_len_1": 66.8412, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.52, "stdk": 0.0473, "stdq": 0.0423, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 27600 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.8866, "doc_norm": 1.3945, "encoder_q-embeddings": 2078.0605, "encoder_q-layer.0": 1410.0527, "encoder_q-layer.1": 1388.5116, "encoder_q-layer.10": 1976.7252, "encoder_q-layer.11": 5169.0454, "encoder_q-layer.2": 1351.1012, "encoder_q-layer.3": 1332.5603, "encoder_q-layer.4": 1248.1378, "encoder_q-layer.5": 1278.2098, "encoder_q-layer.6": 1379.4252, "encoder_q-layer.7": 1570.2527, "encoder_q-layer.8": 1841.4919, "encoder_q-layer.9": 1687.7131, "epoch": 0.12, "inbatch_neg_score": 0.147, "inbatch_pos_score": 0.7158, "learning_rate": 4.53421052631579e-05, "loss": 3.8866, "norm_diff": 0.0371, "norm_loss": 0.0, "num_token_doc": 66.799, "num_token_overlap": 11.6566, "num_token_query": 31.3243, "num_token_union": 65.1364, "num_word_context": 202.2019, "num_word_doc": 49.8246, "num_word_query": 23.2654, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3280.7054, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.147, "query_norm": 1.3574, "queue_k_norm": 1.3903, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3243, "sent_len_1": 66.799, "sent_len_max_0": 127.965, "sent_len_max_1": 210.4988, "stdk": 0.0475, "stdq": 0.0439, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 27700 }, { "accuracy": 45.459, "active_queue_size": 16384.0, "cl_loss": 3.8871, "doc_norm": 1.3925, "encoder_q-embeddings": 1465.0385, "encoder_q-layer.0": 985.2951, "encoder_q-layer.1": 1031.5726, "encoder_q-layer.10": 1849.8997, "encoder_q-layer.11": 4782.377, "encoder_q-layer.2": 1162.5848, "encoder_q-layer.3": 1193.4059, "encoder_q-layer.4": 1221.2593, "encoder_q-layer.5": 1267.5704, "encoder_q-layer.6": 1363.0658, "encoder_q-layer.7": 1501.7682, "encoder_q-layer.8": 1845.139, "encoder_q-layer.9": 1635.3073, "epoch": 0.12, "inbatch_neg_score": 0.1546, "inbatch_pos_score": 0.708, "learning_rate": 4.531578947368421e-05, "loss": 3.8871, "norm_diff": 0.0724, "norm_loss": 0.0, "num_token_doc": 66.5822, "num_token_overlap": 11.6187, "num_token_query": 31.2883, "num_token_union": 64.9966, "num_word_context": 202.3014, "num_word_doc": 49.6708, "num_word_query": 23.2224, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2937.4537, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1548, "query_norm": 1.3202, "queue_k_norm": 1.3929, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2883, "sent_len_1": 66.5822, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.6738, "stdk": 0.0474, "stdq": 0.0428, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 27800 }, { "accuracy": 45.752, "active_queue_size": 16384.0, "cl_loss": 3.8672, "doc_norm": 1.395, "encoder_q-embeddings": 1685.4045, "encoder_q-layer.0": 1130.0028, "encoder_q-layer.1": 1202.0791, "encoder_q-layer.10": 1753.2338, "encoder_q-layer.11": 4863.4932, "encoder_q-layer.2": 1383.4491, "encoder_q-layer.3": 1494.4966, "encoder_q-layer.4": 1637.2415, "encoder_q-layer.5": 1672.5063, "encoder_q-layer.6": 1804.4513, "encoder_q-layer.7": 2076.8979, "encoder_q-layer.8": 2181.9429, "encoder_q-layer.9": 1673.7119, "epoch": 0.12, "inbatch_neg_score": 0.1426, "inbatch_pos_score": 0.6924, "learning_rate": 4.528947368421053e-05, "loss": 3.8672, "norm_diff": 0.0917, "norm_loss": 0.0, "num_token_doc": 66.7488, "num_token_overlap": 11.6535, "num_token_query": 31.3018, "num_token_union": 65.093, "num_word_context": 202.1203, "num_word_doc": 49.7982, "num_word_query": 23.2365, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3216.7091, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1426, "query_norm": 1.3033, "queue_k_norm": 1.3922, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3018, "sent_len_1": 66.7488, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.9863, "stdk": 0.0475, "stdq": 0.0426, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 27900 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.8604, "doc_norm": 1.3901, "encoder_q-embeddings": 1911.885, "encoder_q-layer.0": 1190.2462, "encoder_q-layer.1": 1328.2698, "encoder_q-layer.10": 2109.5, "encoder_q-layer.11": 5192.6479, "encoder_q-layer.2": 1507.176, "encoder_q-layer.3": 1594.3972, "encoder_q-layer.4": 1709.2842, "encoder_q-layer.5": 1720.13, "encoder_q-layer.6": 1839.4449, "encoder_q-layer.7": 1882.6938, "encoder_q-layer.8": 2252.1812, "encoder_q-layer.9": 1927.16, "epoch": 0.12, "inbatch_neg_score": 0.1463, "inbatch_pos_score": 0.6982, "learning_rate": 4.5263157894736846e-05, "loss": 3.8604, "norm_diff": 0.0869, "norm_loss": 0.0, "num_token_doc": 66.6541, "num_token_overlap": 11.6698, "num_token_query": 31.367, "num_token_union": 65.0569, "num_word_context": 202.1657, "num_word_doc": 49.7292, "num_word_query": 23.3231, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3375.0281, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1465, "query_norm": 1.3032, "queue_k_norm": 1.3901, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.367, "sent_len_1": 66.6541, "sent_len_max_0": 128.0, "sent_len_max_1": 208.5175, "stdk": 0.0474, "stdq": 0.0427, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28000 }, { "accuracy": 46.3867, "active_queue_size": 16384.0, "cl_loss": 3.8589, "doc_norm": 1.3838, "encoder_q-embeddings": 1630.5358, "encoder_q-layer.0": 1086.3024, "encoder_q-layer.1": 1119.9762, "encoder_q-layer.10": 1906.3433, "encoder_q-layer.11": 4926.5771, "encoder_q-layer.2": 1240.1522, "encoder_q-layer.3": 1330.2124, "encoder_q-layer.4": 1413.6934, "encoder_q-layer.5": 1421.7925, "encoder_q-layer.6": 1552.3286, "encoder_q-layer.7": 1801.1802, "encoder_q-layer.8": 2049.9275, "encoder_q-layer.9": 1805.4451, "epoch": 0.12, "inbatch_neg_score": 0.146, "inbatch_pos_score": 0.7026, "learning_rate": 4.523684210526316e-05, "loss": 3.8589, "norm_diff": 0.0852, "norm_loss": 0.0, "num_token_doc": 66.7862, "num_token_overlap": 11.6688, "num_token_query": 31.4346, "num_token_union": 65.1915, "num_word_context": 202.3547, "num_word_doc": 49.8244, "num_word_query": 23.3575, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3204.8615, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1466, "query_norm": 1.2986, "queue_k_norm": 1.3897, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4346, "sent_len_1": 66.7862, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.2088, "stdk": 0.0472, "stdq": 0.043, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28100 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.8538, "doc_norm": 1.3884, "encoder_q-embeddings": 3184.2727, "encoder_q-layer.0": 2368.8152, "encoder_q-layer.1": 2406.5964, "encoder_q-layer.10": 1791.2233, "encoder_q-layer.11": 4578.0708, "encoder_q-layer.2": 2858.418, "encoder_q-layer.3": 2830.4585, "encoder_q-layer.4": 2996.8667, "encoder_q-layer.5": 2352.8113, "encoder_q-layer.6": 2474.5688, "encoder_q-layer.7": 2320.0317, "encoder_q-layer.8": 2031.6737, "encoder_q-layer.9": 1741.4254, "epoch": 0.12, "inbatch_neg_score": 0.1418, "inbatch_pos_score": 0.7139, "learning_rate": 4.5210526315789476e-05, "loss": 3.8538, "norm_diff": 0.0902, "norm_loss": 0.0, "num_token_doc": 66.7796, "num_token_overlap": 11.7089, "num_token_query": 31.446, "num_token_union": 65.1421, "num_word_context": 202.6559, "num_word_doc": 49.8372, "num_word_query": 23.3518, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4187.1689, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1422, "query_norm": 1.2982, "queue_k_norm": 1.387, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.446, "sent_len_1": 66.7796, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.2562, "stdk": 0.0475, "stdq": 0.0431, "stdqueue_k": 0.0474, "stdqueue_q": 0.0, "step": 28200 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.8657, "doc_norm": 1.3865, "encoder_q-embeddings": 4114.4492, "encoder_q-layer.0": 2579.1917, "encoder_q-layer.1": 2761.2485, "encoder_q-layer.10": 3598.8076, "encoder_q-layer.11": 9538.1748, "encoder_q-layer.2": 3195.8411, "encoder_q-layer.3": 3336.2961, "encoder_q-layer.4": 3576.7493, "encoder_q-layer.5": 3580.2637, "encoder_q-layer.6": 3462.6111, "encoder_q-layer.7": 3891.4958, "encoder_q-layer.8": 4235.1123, "encoder_q-layer.9": 3792.5542, "epoch": 0.12, "inbatch_neg_score": 0.1415, "inbatch_pos_score": 0.6924, "learning_rate": 4.518421052631579e-05, "loss": 3.8657, "norm_diff": 0.1247, "norm_loss": 0.0, "num_token_doc": 66.7602, "num_token_overlap": 11.7154, "num_token_query": 31.4644, "num_token_union": 65.1315, "num_word_context": 202.4977, "num_word_doc": 49.8185, "num_word_query": 23.3737, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6680.4927, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1416, "query_norm": 1.2618, "queue_k_norm": 1.3873, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4644, "sent_len_1": 66.7602, "sent_len_max_0": 127.995, "sent_len_max_1": 211.7713, "stdk": 0.0474, "stdq": 0.042, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28300 }, { "accuracy": 46.7773, "active_queue_size": 16384.0, "cl_loss": 3.8617, "doc_norm": 1.3875, "encoder_q-embeddings": 3240.7603, "encoder_q-layer.0": 2081.6011, "encoder_q-layer.1": 2098.7268, "encoder_q-layer.10": 3828.5054, "encoder_q-layer.11": 9702.6543, "encoder_q-layer.2": 2342.8977, "encoder_q-layer.3": 2495.4109, "encoder_q-layer.4": 2636.5837, "encoder_q-layer.5": 2605.7046, "encoder_q-layer.6": 2967.6606, "encoder_q-layer.7": 3159.6777, "encoder_q-layer.8": 3806.7385, "encoder_q-layer.9": 3490.6919, "epoch": 0.12, "inbatch_neg_score": 0.1436, "inbatch_pos_score": 0.7163, "learning_rate": 4.515789473684211e-05, "loss": 3.8617, "norm_diff": 0.0946, "norm_loss": 0.0, "num_token_doc": 66.6909, "num_token_overlap": 11.6548, "num_token_query": 31.3875, "num_token_union": 65.1192, "num_word_context": 202.2367, "num_word_doc": 49.7764, "num_word_query": 23.3025, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6098.7942, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1445, "query_norm": 1.2929, "queue_k_norm": 1.386, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3875, "sent_len_1": 66.6909, "sent_len_max_0": 128.0, "sent_len_max_1": 207.6987, "stdk": 0.0475, "stdq": 0.0433, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28400 }, { "accuracy": 46.582, "active_queue_size": 16384.0, "cl_loss": 3.8546, "doc_norm": 1.3851, "encoder_q-embeddings": 3010.5388, "encoder_q-layer.0": 1856.3087, "encoder_q-layer.1": 1923.9424, "encoder_q-layer.10": 3648.9033, "encoder_q-layer.11": 9593.1904, "encoder_q-layer.2": 2125.9758, "encoder_q-layer.3": 2202.0986, "encoder_q-layer.4": 2340.5691, "encoder_q-layer.5": 2524.5107, "encoder_q-layer.6": 2701.0042, "encoder_q-layer.7": 2954.3069, "encoder_q-layer.8": 3678.5828, "encoder_q-layer.9": 3303.0442, "epoch": 0.12, "inbatch_neg_score": 0.1408, "inbatch_pos_score": 0.6909, "learning_rate": 4.5131578947368425e-05, "loss": 3.8546, "norm_diff": 0.1189, "norm_loss": 0.0, "num_token_doc": 66.6986, "num_token_overlap": 11.7107, "num_token_query": 31.5152, "num_token_union": 65.1557, "num_word_context": 202.256, "num_word_doc": 49.7896, "num_word_query": 23.4401, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5790.8951, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1403, "query_norm": 1.2661, "queue_k_norm": 1.3836, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5152, "sent_len_1": 66.6986, "sent_len_max_0": 128.0, "sent_len_max_1": 208.4675, "stdk": 0.0475, "stdq": 0.0422, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28500 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.8498, "doc_norm": 1.3818, "encoder_q-embeddings": 4967.3398, "encoder_q-layer.0": 3331.4453, "encoder_q-layer.1": 3956.8684, "encoder_q-layer.10": 3735.4055, "encoder_q-layer.11": 10013.3223, "encoder_q-layer.2": 4274.7764, "encoder_q-layer.3": 4580.1694, "encoder_q-layer.4": 4649.1787, "encoder_q-layer.5": 4900.626, "encoder_q-layer.6": 5279.3438, "encoder_q-layer.7": 5739.8213, "encoder_q-layer.8": 5023.7246, "encoder_q-layer.9": 3558.0913, "epoch": 0.12, "inbatch_neg_score": 0.1413, "inbatch_pos_score": 0.6958, "learning_rate": 4.5105263157894736e-05, "loss": 3.8498, "norm_diff": 0.0933, "norm_loss": 0.0, "num_token_doc": 66.7589, "num_token_overlap": 11.6923, "num_token_query": 31.3514, "num_token_union": 65.0932, "num_word_context": 202.1646, "num_word_doc": 49.7875, "num_word_query": 23.2833, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7938.0828, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1414, "query_norm": 1.2885, "queue_k_norm": 1.3834, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3514, "sent_len_1": 66.7589, "sent_len_max_0": 127.995, "sent_len_max_1": 209.5425, "stdk": 0.0474, "stdq": 0.0428, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 28600 }, { "accuracy": 46.2402, "active_queue_size": 16384.0, "cl_loss": 3.8586, "doc_norm": 1.3781, "encoder_q-embeddings": 3511.9641, "encoder_q-layer.0": 2226.2598, "encoder_q-layer.1": 2393.5508, "encoder_q-layer.10": 4153.5454, "encoder_q-layer.11": 9996.4834, "encoder_q-layer.2": 2699.1907, "encoder_q-layer.3": 2858.3347, "encoder_q-layer.4": 2995.3579, "encoder_q-layer.5": 2858.6033, "encoder_q-layer.6": 3405.6567, "encoder_q-layer.7": 3584.5791, "encoder_q-layer.8": 4181.4756, "encoder_q-layer.9": 3723.1189, "epoch": 0.12, "inbatch_neg_score": 0.1436, "inbatch_pos_score": 0.6885, "learning_rate": 4.5078947368421055e-05, "loss": 3.8586, "norm_diff": 0.1127, "norm_loss": 0.0, "num_token_doc": 66.7141, "num_token_overlap": 11.6545, "num_token_query": 31.3293, "num_token_union": 65.0765, "num_word_context": 202.138, "num_word_doc": 49.7604, "num_word_query": 23.2721, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6407.3944, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.144, "query_norm": 1.2654, "queue_k_norm": 1.3827, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3293, "sent_len_1": 66.7141, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.2363, "stdk": 0.0474, "stdq": 0.0421, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 28700 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.863, "doc_norm": 1.3771, "encoder_q-embeddings": 14686.2305, "encoder_q-layer.0": 10462.2754, "encoder_q-layer.1": 11689.6865, "encoder_q-layer.10": 3702.8716, "encoder_q-layer.11": 9345.8662, "encoder_q-layer.2": 15552.084, "encoder_q-layer.3": 15697.8682, "encoder_q-layer.4": 12439.4814, "encoder_q-layer.5": 8453.9561, "encoder_q-layer.6": 8845.6123, "encoder_q-layer.7": 7208.0854, "encoder_q-layer.8": 4807.2246, "encoder_q-layer.9": 3555.2925, "epoch": 0.12, "inbatch_neg_score": 0.1443, "inbatch_pos_score": 0.7051, "learning_rate": 4.5052631578947366e-05, "loss": 3.863, "norm_diff": 0.0894, "norm_loss": 0.0, "num_token_doc": 66.7494, "num_token_overlap": 11.7025, "num_token_query": 31.4276, "num_token_union": 65.1222, "num_word_context": 202.1709, "num_word_doc": 49.8279, "num_word_query": 23.3522, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16793.7216, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1436, "query_norm": 1.2877, "queue_k_norm": 1.3791, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4276, "sent_len_1": 66.7494, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.9863, "stdk": 0.0473, "stdq": 0.0426, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 28800 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.8577, "doc_norm": 1.3823, "encoder_q-embeddings": 3155.3923, "encoder_q-layer.0": 2012.5469, "encoder_q-layer.1": 2118.8621, "encoder_q-layer.10": 3817.668, "encoder_q-layer.11": 9313.749, "encoder_q-layer.2": 2442.3562, "encoder_q-layer.3": 2466.6438, "encoder_q-layer.4": 2583.0911, "encoder_q-layer.5": 2696.4756, "encoder_q-layer.6": 2874.6826, "encoder_q-layer.7": 3238.3801, "encoder_q-layer.8": 3725.4143, "encoder_q-layer.9": 3485.407, "epoch": 0.13, "inbatch_neg_score": 0.1443, "inbatch_pos_score": 0.7158, "learning_rate": 4.502631578947369e-05, "loss": 3.8577, "norm_diff": 0.0789, "norm_loss": 0.0, "num_token_doc": 66.5819, "num_token_overlap": 11.653, "num_token_query": 31.2534, "num_token_union": 64.966, "num_word_context": 202.0485, "num_word_doc": 49.6852, "num_word_query": 23.2063, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6010.318, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1447, "query_norm": 1.3034, "queue_k_norm": 1.3802, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2534, "sent_len_1": 66.5819, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.6937, "stdk": 0.0476, "stdq": 0.0433, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 28900 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.8479, "doc_norm": 1.3811, "encoder_q-embeddings": 3491.5676, "encoder_q-layer.0": 2457.0842, "encoder_q-layer.1": 2568.1345, "encoder_q-layer.10": 4157.0669, "encoder_q-layer.11": 10112.6475, "encoder_q-layer.2": 2882.2368, "encoder_q-layer.3": 3106.2288, "encoder_q-layer.4": 3502.7092, "encoder_q-layer.5": 3303.9177, "encoder_q-layer.6": 3527.7107, "encoder_q-layer.7": 3656.1743, "encoder_q-layer.8": 4243.8008, "encoder_q-layer.9": 3660.8562, "epoch": 0.13, "inbatch_neg_score": 0.1439, "inbatch_pos_score": 0.7031, "learning_rate": 4.5e-05, "loss": 3.8479, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.844, "num_token_overlap": 11.7092, "num_token_query": 31.3968, "num_token_union": 65.1594, "num_word_context": 202.379, "num_word_doc": 49.8711, "num_word_query": 23.3054, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6619.7069, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1428, "query_norm": 1.3106, "queue_k_norm": 1.3777, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3968, "sent_len_1": 66.844, "sent_len_max_0": 127.97, "sent_len_max_1": 208.4175, "stdk": 0.0476, "stdq": 0.0433, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 29000 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.8522, "doc_norm": 1.377, "encoder_q-embeddings": 3224.6528, "encoder_q-layer.0": 2162.4612, "encoder_q-layer.1": 2279.439, "encoder_q-layer.10": 3691.7935, "encoder_q-layer.11": 9426.6816, "encoder_q-layer.2": 2550.4949, "encoder_q-layer.3": 2672.1182, "encoder_q-layer.4": 2894.5618, "encoder_q-layer.5": 2835.3218, "encoder_q-layer.6": 2946.4741, "encoder_q-layer.7": 3174.7974, "encoder_q-layer.8": 3765.7651, "encoder_q-layer.9": 3312.1038, "epoch": 0.13, "inbatch_neg_score": 0.1486, "inbatch_pos_score": 0.7051, "learning_rate": 4.497368421052632e-05, "loss": 3.8522, "norm_diff": 0.0897, "norm_loss": 0.0, "num_token_doc": 66.7852, "num_token_overlap": 11.6366, "num_token_query": 31.3246, "num_token_union": 65.1491, "num_word_context": 202.4402, "num_word_doc": 49.8402, "num_word_query": 23.2625, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6067.2357, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1484, "query_norm": 1.2873, "queue_k_norm": 1.3779, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3246, "sent_len_1": 66.7852, "sent_len_max_0": 127.99, "sent_len_max_1": 208.065, "stdk": 0.0475, "stdq": 0.0426, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 29100 }, { "accuracy": 47.0215, "active_queue_size": 16384.0, "cl_loss": 3.8351, "doc_norm": 1.3779, "encoder_q-embeddings": 3112.9336, "encoder_q-layer.0": 1962.0664, "encoder_q-layer.1": 2056.2734, "encoder_q-layer.10": 3389.7734, "encoder_q-layer.11": 9053.3555, "encoder_q-layer.2": 2368.4541, "encoder_q-layer.3": 2491.4182, "encoder_q-layer.4": 2591.0022, "encoder_q-layer.5": 2610.2883, "encoder_q-layer.6": 3114.9749, "encoder_q-layer.7": 3454.4482, "encoder_q-layer.8": 3947.3818, "encoder_q-layer.9": 3289.8091, "epoch": 0.13, "inbatch_neg_score": 0.1516, "inbatch_pos_score": 0.708, "learning_rate": 4.494736842105263e-05, "loss": 3.8351, "norm_diff": 0.0998, "norm_loss": 0.0, "num_token_doc": 66.8936, "num_token_overlap": 11.6951, "num_token_query": 31.3811, "num_token_union": 65.1777, "num_word_context": 202.4036, "num_word_doc": 49.9235, "num_word_query": 23.3206, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5861.8995, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1509, "query_norm": 1.2781, "queue_k_norm": 1.3749, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3811, "sent_len_1": 66.8936, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.8825, "stdk": 0.0475, "stdq": 0.0425, "stdqueue_k": 0.0475, "stdqueue_q": 0.0, "step": 29200 }, { "accuracy": 44.9219, "active_queue_size": 16384.0, "cl_loss": 3.844, "doc_norm": 1.3771, "encoder_q-embeddings": 3232.4124, "encoder_q-layer.0": 2098.823, "encoder_q-layer.1": 2193.1484, "encoder_q-layer.10": 3516.5195, "encoder_q-layer.11": 9688.8301, "encoder_q-layer.2": 2477.8152, "encoder_q-layer.3": 2535.4263, "encoder_q-layer.4": 2730.8784, "encoder_q-layer.5": 2799.2444, "encoder_q-layer.6": 2971.6919, "encoder_q-layer.7": 3398.0303, "encoder_q-layer.8": 4269.2012, "encoder_q-layer.9": 3585.2219, "epoch": 0.13, "inbatch_neg_score": 0.1513, "inbatch_pos_score": 0.6982, "learning_rate": 4.4921052631578944e-05, "loss": 3.844, "norm_diff": 0.1124, "norm_loss": 0.0, "num_token_doc": 66.9286, "num_token_overlap": 11.6832, "num_token_query": 31.3734, "num_token_union": 65.2059, "num_word_context": 202.5932, "num_word_doc": 49.9346, "num_word_query": 23.303, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6194.0542, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1515, "query_norm": 1.2647, "queue_k_norm": 1.377, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3734, "sent_len_1": 66.9286, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.83, "stdk": 0.0476, "stdq": 0.0421, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 29300 }, { "accuracy": 45.9961, "active_queue_size": 16384.0, "cl_loss": 3.8608, "doc_norm": 1.3755, "encoder_q-embeddings": 3093.6716, "encoder_q-layer.0": 2055.95, "encoder_q-layer.1": 2161.2881, "encoder_q-layer.10": 3705.3374, "encoder_q-layer.11": 9384.6836, "encoder_q-layer.2": 2427.8274, "encoder_q-layer.3": 2499.2903, "encoder_q-layer.4": 2569.3696, "encoder_q-layer.5": 2595.5344, "encoder_q-layer.6": 2939.593, "encoder_q-layer.7": 3417.281, "encoder_q-layer.8": 4079.6914, "encoder_q-layer.9": 3406.3367, "epoch": 0.13, "inbatch_neg_score": 0.1503, "inbatch_pos_score": 0.7119, "learning_rate": 4.489473684210527e-05, "loss": 3.8608, "norm_diff": 0.0857, "norm_loss": 0.0, "num_token_doc": 66.6823, "num_token_overlap": 11.6329, "num_token_query": 31.2521, "num_token_union": 65.0282, "num_word_context": 202.0437, "num_word_doc": 49.7601, "num_word_query": 23.2056, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5891.9763, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.151, "query_norm": 1.2898, "queue_k_norm": 1.3753, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2521, "sent_len_1": 66.6823, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.45, "stdk": 0.0475, "stdq": 0.0433, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 29400 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.8326, "doc_norm": 1.3759, "encoder_q-embeddings": 3583.6206, "encoder_q-layer.0": 2380.26, "encoder_q-layer.1": 2571.7017, "encoder_q-layer.10": 3456.4705, "encoder_q-layer.11": 8880.1631, "encoder_q-layer.2": 3047.2932, "encoder_q-layer.3": 3179.208, "encoder_q-layer.4": 3207.9175, "encoder_q-layer.5": 2858.9648, "encoder_q-layer.6": 3041.0183, "encoder_q-layer.7": 3218.2251, "encoder_q-layer.8": 3581.9822, "encoder_q-layer.9": 3256.3994, "epoch": 0.13, "inbatch_neg_score": 0.1489, "inbatch_pos_score": 0.7339, "learning_rate": 4.486842105263158e-05, "loss": 3.8326, "norm_diff": 0.0864, "norm_loss": 0.0, "num_token_doc": 66.7463, "num_token_overlap": 11.7198, "num_token_query": 31.5213, "num_token_union": 65.1323, "num_word_context": 202.0148, "num_word_doc": 49.8044, "num_word_query": 23.4056, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6038.4974, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1492, "query_norm": 1.2895, "queue_k_norm": 1.3784, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5213, "sent_len_1": 66.7463, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.0613, "stdk": 0.0475, "stdq": 0.0433, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 29500 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.8474, "doc_norm": 1.3698, "encoder_q-embeddings": 3362.6167, "encoder_q-layer.0": 2236.1089, "encoder_q-layer.1": 2258.0652, "encoder_q-layer.10": 4013.9917, "encoder_q-layer.11": 9644.6182, "encoder_q-layer.2": 2489.9851, "encoder_q-layer.3": 2597.0376, "encoder_q-layer.4": 2704.1741, "encoder_q-layer.5": 2656.2476, "encoder_q-layer.6": 3030.3684, "encoder_q-layer.7": 3547.2461, "encoder_q-layer.8": 3942.5173, "encoder_q-layer.9": 3560.3904, "epoch": 0.13, "inbatch_neg_score": 0.1473, "inbatch_pos_score": 0.707, "learning_rate": 4.48421052631579e-05, "loss": 3.8474, "norm_diff": 0.0979, "norm_loss": 0.0, "num_token_doc": 66.5938, "num_token_overlap": 11.6387, "num_token_query": 31.3805, "num_token_union": 65.0606, "num_word_context": 202.3821, "num_word_doc": 49.6922, "num_word_query": 23.3089, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6121.7265, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1476, "query_norm": 1.2719, "queue_k_norm": 1.3774, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3805, "sent_len_1": 66.5938, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.3063, "stdk": 0.0474, "stdq": 0.0429, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 29600 }, { "accuracy": 46.2891, "active_queue_size": 16384.0, "cl_loss": 3.8499, "doc_norm": 1.3757, "encoder_q-embeddings": 3110.5305, "encoder_q-layer.0": 1914.2887, "encoder_q-layer.1": 2007.1819, "encoder_q-layer.10": 3915.0447, "encoder_q-layer.11": 10028.4551, "encoder_q-layer.2": 2263.613, "encoder_q-layer.3": 2486.1392, "encoder_q-layer.4": 2642.5544, "encoder_q-layer.5": 2620.1062, "encoder_q-layer.6": 3018.709, "encoder_q-layer.7": 3377.0591, "encoder_q-layer.8": 4039.6572, "encoder_q-layer.9": 3522.3057, "epoch": 0.13, "inbatch_neg_score": 0.1427, "inbatch_pos_score": 0.7051, "learning_rate": 4.481578947368421e-05, "loss": 3.8499, "norm_diff": 0.1107, "norm_loss": 0.0, "num_token_doc": 66.7958, "num_token_overlap": 11.6539, "num_token_query": 31.3206, "num_token_union": 65.1208, "num_word_context": 202.4491, "num_word_doc": 49.8094, "num_word_query": 23.2414, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6168.5001, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1428, "query_norm": 1.265, "queue_k_norm": 1.3757, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3206, "sent_len_1": 66.7958, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.555, "stdk": 0.0476, "stdq": 0.0427, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 29700 }, { "accuracy": 46.4355, "active_queue_size": 16384.0, "cl_loss": 3.8423, "doc_norm": 1.3743, "encoder_q-embeddings": 3279.0955, "encoder_q-layer.0": 2156.2302, "encoder_q-layer.1": 2262.791, "encoder_q-layer.10": 3549.5339, "encoder_q-layer.11": 9433.207, "encoder_q-layer.2": 2528.3103, "encoder_q-layer.3": 2612.9692, "encoder_q-layer.4": 2742.6885, "encoder_q-layer.5": 2857.8652, "encoder_q-layer.6": 3073.4666, "encoder_q-layer.7": 3300.9048, "encoder_q-layer.8": 3889.374, "encoder_q-layer.9": 3374.5679, "epoch": 0.13, "inbatch_neg_score": 0.1403, "inbatch_pos_score": 0.707, "learning_rate": 4.478947368421053e-05, "loss": 3.8423, "norm_diff": 0.1068, "norm_loss": 0.0, "num_token_doc": 66.6548, "num_token_overlap": 11.6505, "num_token_query": 31.3526, "num_token_union": 65.0737, "num_word_context": 202.5195, "num_word_doc": 49.7547, "num_word_query": 23.2951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6016.7227, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1392, "query_norm": 1.2675, "queue_k_norm": 1.3742, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3526, "sent_len_1": 66.6548, "sent_len_max_0": 127.99, "sent_len_max_1": 206.875, "stdk": 0.0475, "stdq": 0.0429, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 29800 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.8289, "doc_norm": 1.3702, "encoder_q-embeddings": 3314.7964, "encoder_q-layer.0": 2119.5078, "encoder_q-layer.1": 2243.0671, "encoder_q-layer.10": 4023.5266, "encoder_q-layer.11": 9692.1504, "encoder_q-layer.2": 2554.5691, "encoder_q-layer.3": 2577.0308, "encoder_q-layer.4": 2786.3127, "encoder_q-layer.5": 2711.9958, "encoder_q-layer.6": 3066.79, "encoder_q-layer.7": 3488.5225, "encoder_q-layer.8": 4195.1133, "encoder_q-layer.9": 3628.1677, "epoch": 0.13, "inbatch_neg_score": 0.1387, "inbatch_pos_score": 0.6997, "learning_rate": 4.476315789473685e-05, "loss": 3.8289, "norm_diff": 0.1133, "norm_loss": 0.0, "num_token_doc": 66.912, "num_token_overlap": 11.6812, "num_token_query": 31.3859, "num_token_union": 65.2202, "num_word_context": 202.5344, "num_word_doc": 49.9043, "num_word_query": 23.3223, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6100.9096, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1387, "query_norm": 1.2569, "queue_k_norm": 1.3748, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3859, "sent_len_1": 66.912, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.6238, "stdk": 0.0474, "stdq": 0.0426, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 29900 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.8305, "doc_norm": 1.3768, "encoder_q-embeddings": 3650.4663, "encoder_q-layer.0": 2330.9846, "encoder_q-layer.1": 2447.4487, "encoder_q-layer.10": 4431.8794, "encoder_q-layer.11": 10476.375, "encoder_q-layer.2": 2734.551, "encoder_q-layer.3": 2909.5115, "encoder_q-layer.4": 2988.8159, "encoder_q-layer.5": 2862.1538, "encoder_q-layer.6": 3274.5676, "encoder_q-layer.7": 3528.1416, "encoder_q-layer.8": 4323.6533, "encoder_q-layer.9": 3984.9417, "epoch": 0.13, "inbatch_neg_score": 0.1414, "inbatch_pos_score": 0.6904, "learning_rate": 4.473684210526316e-05, "loss": 3.8305, "norm_diff": 0.1358, "norm_loss": 0.0, "num_token_doc": 66.6552, "num_token_overlap": 11.6511, "num_token_query": 31.2969, "num_token_union": 65.031, "num_word_context": 202.0184, "num_word_doc": 49.7418, "num_word_query": 23.2614, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6648.4994, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1423, "query_norm": 1.241, "queue_k_norm": 1.3721, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2969, "sent_len_1": 66.6552, "sent_len_max_0": 127.9688, "sent_len_max_1": 208.3487, "stdk": 0.0477, "stdq": 0.0418, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 30000 }, { "dev_runtime": 30.5009, "dev_samples_per_second": 1.049, "dev_steps_per_second": 0.033, "epoch": 0.13, "step": 30000, "test_accuracy": 92.6513671875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.4281162619590759, "test_doc_norm": 1.323779821395874, "test_inbatch_neg_score": 0.3768174946308136, "test_inbatch_pos_score": 1.2955660820007324, "test_loss": 0.4281162619590759, "test_loss_align": 1.0047688484191895, "test_loss_unif": 3.949068784713745, "test_loss_unif_q@queue": 3.949068784713745, "test_norm_diff": 0.022104386240243912, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.12664198875427246, "test_query_norm": 1.3458842039108276, "test_queue_k_norm": 1.3720059394836426, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.042037054896354675, "test_stdq": 0.04164276272058487, "test_stdqueue_k": 0.047580599784851074, "test_stdqueue_q": 0.0 }, { "dev_runtime": 30.5009, "dev_samples_per_second": 1.049, "dev_steps_per_second": 0.033, "epoch": 0.13, "eval_beir-arguana_ndcg@10": 0.34677, "eval_beir-arguana_recall@10": 0.59175, "eval_beir-arguana_recall@100": 0.90612, "eval_beir-arguana_recall@20": 0.73684, "eval_beir-avg_ndcg@10": 0.36659616666666667, "eval_beir-avg_recall@10": 0.43351533333333325, "eval_beir-avg_recall@100": 0.6169480833333334, "eval_beir-avg_recall@20": 0.4968115833333333, "eval_beir-cqadupstack_ndcg@10": 0.2528916666666667, "eval_beir-cqadupstack_recall@10": 0.34464333333333336, "eval_beir-cqadupstack_recall@100": 0.5789508333333333, "eval_beir-cqadupstack_recall@20": 0.41378583333333335, "eval_beir-fiqa_ndcg@10": 0.22972, "eval_beir-fiqa_recall@10": 0.28676, "eval_beir-fiqa_recall@100": 0.55097, "eval_beir-fiqa_recall@20": 0.36051, "eval_beir-nfcorpus_ndcg@10": 0.28418, "eval_beir-nfcorpus_recall@10": 0.14078, "eval_beir-nfcorpus_recall@100": 0.27487, "eval_beir-nfcorpus_recall@20": 0.17132, "eval_beir-nq_ndcg@10": 0.27921, "eval_beir-nq_recall@10": 0.44998, "eval_beir-nq_recall@100": 0.79244, "eval_beir-nq_recall@20": 0.5793, "eval_beir-quora_ndcg@10": 0.76611, "eval_beir-quora_recall@10": 0.87818, "eval_beir-quora_recall@100": 0.97465, "eval_beir-quora_recall@20": 0.92141, "eval_beir-scidocs_ndcg@10": 0.14412, "eval_beir-scidocs_recall@10": 0.15292, "eval_beir-scidocs_recall@100": 0.34538, "eval_beir-scidocs_recall@20": 0.20847, "eval_beir-scifact_ndcg@10": 0.62371, "eval_beir-scifact_recall@10": 0.76806, "eval_beir-scifact_recall@100": 0.90656, "eval_beir-scifact_recall@20": 0.82078, "eval_beir-trec-covid_ndcg@10": 0.54524, "eval_beir-trec-covid_recall@10": 0.586, "eval_beir-trec-covid_recall@100": 0.4216, "eval_beir-trec-covid_recall@20": 0.559, "eval_beir-webis-touche2020_ndcg@10": 0.19401, "eval_beir-webis-touche2020_recall@10": 0.13608, "eval_beir-webis-touche2020_recall@100": 0.41794, "eval_beir-webis-touche2020_recall@20": 0.1967, "eval_senteval-avg_sts": 0.765800862837222, "eval_senteval-sickr_spearman": 0.7287842115010321, "eval_senteval-stsb_spearman": 0.802817514173412, "step": 30000, "test_accuracy": 92.6513671875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.4281162619590759, "test_doc_norm": 1.323779821395874, "test_inbatch_neg_score": 0.3768174946308136, "test_inbatch_pos_score": 1.2955660820007324, "test_loss": 0.4281162619590759, "test_loss_align": 1.0047688484191895, "test_loss_unif": 3.949068784713745, "test_loss_unif_q@queue": 3.949068784713745, "test_norm_diff": 0.022104386240243912, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.12664198875427246, "test_query_norm": 1.3458842039108276, "test_queue_k_norm": 1.3720059394836426, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.042037054896354675, "test_stdq": 0.04164276272058487, "test_stdqueue_k": 0.047580599784851074, "test_stdqueue_q": 0.0 }, { "accuracy": 45.9961, "active_queue_size": 16384.0, "cl_loss": 3.8202, "doc_norm": 1.3694, "encoder_q-embeddings": 2999.7493, "encoder_q-layer.0": 2063.4072, "encoder_q-layer.1": 2138.6887, "encoder_q-layer.10": 3659.4795, "encoder_q-layer.11": 8966.0459, "encoder_q-layer.2": 2429.4509, "encoder_q-layer.3": 2478.8203, "encoder_q-layer.4": 2619.5317, "encoder_q-layer.5": 2660.6836, "encoder_q-layer.6": 3068.0352, "encoder_q-layer.7": 3377.1689, "encoder_q-layer.8": 3790.6919, "encoder_q-layer.9": 3435.6399, "epoch": 0.13, "inbatch_neg_score": 0.1445, "inbatch_pos_score": 0.6919, "learning_rate": 4.471052631578948e-05, "loss": 3.8202, "norm_diff": 0.1104, "norm_loss": 0.0, "num_token_doc": 66.8032, "num_token_overlap": 11.6749, "num_token_query": 31.4203, "num_token_union": 65.2148, "num_word_context": 202.5285, "num_word_doc": 49.8764, "num_word_query": 23.353, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5844.7859, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1438, "query_norm": 1.259, "queue_k_norm": 1.3746, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4203, "sent_len_1": 66.8032, "sent_len_max_0": 127.9737, "sent_len_max_1": 205.8575, "stdk": 0.0474, "stdq": 0.0424, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30100 }, { "accuracy": 45.9473, "active_queue_size": 16384.0, "cl_loss": 3.8217, "doc_norm": 1.3798, "encoder_q-embeddings": 3291.1553, "encoder_q-layer.0": 2091.4453, "encoder_q-layer.1": 2206.3159, "encoder_q-layer.10": 3675.1829, "encoder_q-layer.11": 9358.8643, "encoder_q-layer.2": 2564.6028, "encoder_q-layer.3": 2568.0088, "encoder_q-layer.4": 2731.0103, "encoder_q-layer.5": 2762.0381, "encoder_q-layer.6": 2895.0403, "encoder_q-layer.7": 3206.5132, "encoder_q-layer.8": 3680.3608, "encoder_q-layer.9": 3305.051, "epoch": 0.13, "inbatch_neg_score": 0.1435, "inbatch_pos_score": 0.7056, "learning_rate": 4.468421052631579e-05, "loss": 3.8217, "norm_diff": 0.1091, "norm_loss": 0.0, "num_token_doc": 66.6069, "num_token_overlap": 11.675, "num_token_query": 31.3216, "num_token_union": 65.0033, "num_word_context": 202.2757, "num_word_doc": 49.7203, "num_word_query": 23.2539, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5964.1578, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1431, "query_norm": 1.2707, "queue_k_norm": 1.3732, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3216, "sent_len_1": 66.6069, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.5925, "stdk": 0.0478, "stdq": 0.0428, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 30200 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.8382, "doc_norm": 1.3712, "encoder_q-embeddings": 6132.0479, "encoder_q-layer.0": 4019.3301, "encoder_q-layer.1": 4190.2256, "encoder_q-layer.10": 9410.2305, "encoder_q-layer.11": 20146.5449, "encoder_q-layer.2": 4727.6689, "encoder_q-layer.3": 5031.1299, "encoder_q-layer.4": 5334.4375, "encoder_q-layer.5": 5193.583, "encoder_q-layer.6": 6030.2939, "encoder_q-layer.7": 7055.875, "encoder_q-layer.8": 8368.7109, "encoder_q-layer.9": 8298.1123, "epoch": 0.13, "inbatch_neg_score": 0.1415, "inbatch_pos_score": 0.7065, "learning_rate": 4.465789473684211e-05, "loss": 3.8382, "norm_diff": 0.0684, "norm_loss": 0.0, "num_token_doc": 66.6969, "num_token_overlap": 11.6422, "num_token_query": 31.2866, "num_token_union": 65.0475, "num_word_context": 202.1681, "num_word_doc": 49.7813, "num_word_query": 23.2241, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12210.8878, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1415, "query_norm": 1.3028, "queue_k_norm": 1.374, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2866, "sent_len_1": 66.6969, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.3925, "stdk": 0.0475, "stdq": 0.0438, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30300 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.8257, "doc_norm": 1.3744, "encoder_q-embeddings": 7148.8115, "encoder_q-layer.0": 4977.7114, "encoder_q-layer.1": 4964.1084, "encoder_q-layer.10": 8933.9561, "encoder_q-layer.11": 19836.166, "encoder_q-layer.2": 5356.4434, "encoder_q-layer.3": 5384.5015, "encoder_q-layer.4": 5643.8252, "encoder_q-layer.5": 6072.2368, "encoder_q-layer.6": 6533.1465, "encoder_q-layer.7": 7443.8628, "encoder_q-layer.8": 8767.085, "encoder_q-layer.9": 8058.252, "epoch": 0.13, "inbatch_neg_score": 0.1437, "inbatch_pos_score": 0.7158, "learning_rate": 4.4631578947368426e-05, "loss": 3.8257, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.8214, "num_token_overlap": 11.6933, "num_token_query": 31.4556, "num_token_union": 65.2184, "num_word_context": 202.312, "num_word_doc": 49.8461, "num_word_query": 23.364, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12527.7659, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1434, "query_norm": 1.2963, "queue_k_norm": 1.373, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4556, "sent_len_1": 66.8214, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.0275, "stdk": 0.0477, "stdq": 0.0436, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30400 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 3.8289, "doc_norm": 1.3739, "encoder_q-embeddings": 8236.5391, "encoder_q-layer.0": 5509.7598, "encoder_q-layer.1": 5563.3501, "encoder_q-layer.10": 8083.6475, "encoder_q-layer.11": 18541.5703, "encoder_q-layer.2": 6377.0483, "encoder_q-layer.3": 6846.5566, "encoder_q-layer.4": 7182.0894, "encoder_q-layer.5": 7961.7622, "encoder_q-layer.6": 8447.2471, "encoder_q-layer.7": 8200.832, "encoder_q-layer.8": 9034.1172, "encoder_q-layer.9": 7121.2412, "epoch": 0.13, "inbatch_neg_score": 0.1476, "inbatch_pos_score": 0.7183, "learning_rate": 4.460526315789474e-05, "loss": 3.8289, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.7582, "num_token_overlap": 11.703, "num_token_query": 31.4189, "num_token_union": 65.1106, "num_word_context": 202.1524, "num_word_doc": 49.7785, "num_word_query": 23.3309, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13111.0698, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1476, "query_norm": 1.2982, "queue_k_norm": 1.3731, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4189, "sent_len_1": 66.7582, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.7188, "stdk": 0.0477, "stdq": 0.0434, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30500 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.8455, "doc_norm": 1.3769, "encoder_q-embeddings": 8774.5127, "encoder_q-layer.0": 6177.5571, "encoder_q-layer.1": 7051.666, "encoder_q-layer.10": 7604.4155, "encoder_q-layer.11": 19490.0938, "encoder_q-layer.2": 8326.0957, "encoder_q-layer.3": 8349.752, "encoder_q-layer.4": 8167.5825, "encoder_q-layer.5": 7583.7212, "encoder_q-layer.6": 7284.6226, "encoder_q-layer.7": 7433.5166, "encoder_q-layer.8": 8472.2949, "encoder_q-layer.9": 7284.1509, "epoch": 0.13, "inbatch_neg_score": 0.1552, "inbatch_pos_score": 0.7021, "learning_rate": 4.4578947368421056e-05, "loss": 3.8455, "norm_diff": 0.0838, "norm_loss": 0.0, "num_token_doc": 66.7908, "num_token_overlap": 11.6873, "num_token_query": 31.4007, "num_token_union": 65.1533, "num_word_context": 202.4646, "num_word_doc": 49.8458, "num_word_query": 23.3208, "postclip_grad_norm": 1.0, "preclip_grad_norm": 14083.1176, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1544, "query_norm": 1.293, "queue_k_norm": 1.3725, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4007, "sent_len_1": 66.7908, "sent_len_max_0": 127.96, "sent_len_max_1": 207.62, "stdk": 0.0478, "stdq": 0.0424, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30600 }, { "accuracy": 45.2148, "active_queue_size": 16384.0, "cl_loss": 3.8304, "doc_norm": 1.3728, "encoder_q-embeddings": 6116.3286, "encoder_q-layer.0": 4079.8303, "encoder_q-layer.1": 4233.8936, "encoder_q-layer.10": 7330.8555, "encoder_q-layer.11": 17530.7676, "encoder_q-layer.2": 4685.5063, "encoder_q-layer.3": 4891.4917, "encoder_q-layer.4": 5239.7246, "encoder_q-layer.5": 5241.3535, "encoder_q-layer.6": 5733.7246, "encoder_q-layer.7": 6265.79, "encoder_q-layer.8": 7566.9771, "encoder_q-layer.9": 6801.5552, "epoch": 0.13, "inbatch_neg_score": 0.1515, "inbatch_pos_score": 0.7007, "learning_rate": 4.455263157894737e-05, "loss": 3.8304, "norm_diff": 0.1008, "norm_loss": 0.0, "num_token_doc": 66.6781, "num_token_overlap": 11.6786, "num_token_query": 31.3788, "num_token_union": 65.0594, "num_word_context": 202.2735, "num_word_doc": 49.7508, "num_word_query": 23.3022, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11220.0307, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1512, "query_norm": 1.2721, "queue_k_norm": 1.3726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3788, "sent_len_1": 66.6781, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.54, "stdk": 0.0476, "stdq": 0.0422, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30700 }, { "accuracy": 47.3145, "active_queue_size": 16384.0, "cl_loss": 3.793, "doc_norm": 1.3788, "encoder_q-embeddings": 7536.9937, "encoder_q-layer.0": 5123.7856, "encoder_q-layer.1": 5354.0342, "encoder_q-layer.10": 7373.4551, "encoder_q-layer.11": 18326.1523, "encoder_q-layer.2": 6308.6792, "encoder_q-layer.3": 6423.6201, "encoder_q-layer.4": 6833.5996, "encoder_q-layer.5": 6730.8276, "encoder_q-layer.6": 7127.064, "encoder_q-layer.7": 7370.9199, "encoder_q-layer.8": 7705.3154, "encoder_q-layer.9": 7103.709, "epoch": 0.13, "inbatch_neg_score": 0.1526, "inbatch_pos_score": 0.7217, "learning_rate": 4.4526315789473686e-05, "loss": 3.793, "norm_diff": 0.094, "norm_loss": 0.0, "num_token_doc": 66.9383, "num_token_overlap": 11.76, "num_token_query": 31.5231, "num_token_union": 65.2522, "num_word_context": 202.3563, "num_word_doc": 49.9622, "num_word_query": 23.4279, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12603.5861, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1526, "query_norm": 1.2848, "queue_k_norm": 1.3737, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5231, "sent_len_1": 66.9383, "sent_len_max_0": 127.97, "sent_len_max_1": 208.3913, "stdk": 0.0478, "stdq": 0.0429, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30800 }, { "accuracy": 46.8262, "active_queue_size": 16384.0, "cl_loss": 3.8185, "doc_norm": 1.3735, "encoder_q-embeddings": 7906.9756, "encoder_q-layer.0": 5589.0205, "encoder_q-layer.1": 5258.2832, "encoder_q-layer.10": 6887.5645, "encoder_q-layer.11": 17743.3789, "encoder_q-layer.2": 5822.1162, "encoder_q-layer.3": 6172.0, "encoder_q-layer.4": 6423.1484, "encoder_q-layer.5": 6022.4531, "encoder_q-layer.6": 6969.0723, "encoder_q-layer.7": 7334.9038, "encoder_q-layer.8": 8012.5586, "encoder_q-layer.9": 6571.9565, "epoch": 0.13, "inbatch_neg_score": 0.1487, "inbatch_pos_score": 0.709, "learning_rate": 4.4500000000000004e-05, "loss": 3.8185, "norm_diff": 0.0998, "norm_loss": 0.0, "num_token_doc": 66.728, "num_token_overlap": 11.681, "num_token_query": 31.3624, "num_token_union": 65.0818, "num_word_context": 202.1129, "num_word_doc": 49.7911, "num_word_query": 23.295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12431.8265, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1481, "query_norm": 1.2737, "queue_k_norm": 1.3737, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3624, "sent_len_1": 66.728, "sent_len_max_0": 127.99, "sent_len_max_1": 207.4863, "stdk": 0.0476, "stdq": 0.0427, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 30900 }, { "accuracy": 44.6289, "active_queue_size": 16384.0, "cl_loss": 3.8137, "doc_norm": 1.3655, "encoder_q-embeddings": 14848.3613, "encoder_q-layer.0": 9328.4521, "encoder_q-layer.1": 9387.8398, "encoder_q-layer.10": 7627.2271, "encoder_q-layer.11": 19070.4961, "encoder_q-layer.2": 10777.792, "encoder_q-layer.3": 10139.6045, "encoder_q-layer.4": 9951.3555, "encoder_q-layer.5": 8884.6885, "encoder_q-layer.6": 9452.7324, "encoder_q-layer.7": 10387.5869, "encoder_q-layer.8": 8618.2334, "encoder_q-layer.9": 7294.2012, "epoch": 0.13, "inbatch_neg_score": 0.1451, "inbatch_pos_score": 0.6855, "learning_rate": 4.4473684210526316e-05, "loss": 3.8137, "norm_diff": 0.1077, "norm_loss": 0.0, "num_token_doc": 66.621, "num_token_overlap": 11.6691, "num_token_query": 31.3455, "num_token_union": 65.0095, "num_word_context": 202.1802, "num_word_doc": 49.7249, "num_word_query": 23.2719, "postclip_grad_norm": 1.0, "preclip_grad_norm": 17112.0529, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1451, "query_norm": 1.2578, "queue_k_norm": 1.375, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3455, "sent_len_1": 66.621, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.9275, "stdk": 0.0473, "stdq": 0.0423, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31000 }, { "accuracy": 45.5078, "active_queue_size": 16384.0, "cl_loss": 3.8099, "doc_norm": 1.372, "encoder_q-embeddings": 6016.5249, "encoder_q-layer.0": 3853.3579, "encoder_q-layer.1": 4007.8301, "encoder_q-layer.10": 7770.7773, "encoder_q-layer.11": 19898.1484, "encoder_q-layer.2": 4467.978, "encoder_q-layer.3": 4636.8281, "encoder_q-layer.4": 4906.7256, "encoder_q-layer.5": 5112.9771, "encoder_q-layer.6": 5671.292, "encoder_q-layer.7": 6453.5498, "encoder_q-layer.8": 7748.6797, "encoder_q-layer.9": 6857.7759, "epoch": 0.13, "inbatch_neg_score": 0.1468, "inbatch_pos_score": 0.709, "learning_rate": 4.4447368421052634e-05, "loss": 3.8099, "norm_diff": 0.1049, "norm_loss": 0.0, "num_token_doc": 66.8291, "num_token_overlap": 11.6876, "num_token_query": 31.3801, "num_token_union": 65.1356, "num_word_context": 202.3247, "num_word_doc": 49.8861, "num_word_query": 23.3006, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12198.5368, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1471, "query_norm": 1.2671, "queue_k_norm": 1.374, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3801, "sent_len_1": 66.8291, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.9975, "stdk": 0.0476, "stdq": 0.0428, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31100 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.8263, "doc_norm": 1.3737, "encoder_q-embeddings": 9958.6396, "encoder_q-layer.0": 7094.8916, "encoder_q-layer.1": 6704.7422, "encoder_q-layer.10": 7002.2603, "encoder_q-layer.11": 17783.3105, "encoder_q-layer.2": 7270.0264, "encoder_q-layer.3": 7234.147, "encoder_q-layer.4": 7135.1626, "encoder_q-layer.5": 6285.1055, "encoder_q-layer.6": 6777.6411, "encoder_q-layer.7": 7510.2222, "encoder_q-layer.8": 7562.6289, "encoder_q-layer.9": 6540.4893, "epoch": 0.14, "inbatch_neg_score": 0.1446, "inbatch_pos_score": 0.7051, "learning_rate": 4.442105263157895e-05, "loss": 3.8263, "norm_diff": 0.1163, "norm_loss": 0.0, "num_token_doc": 66.7593, "num_token_overlap": 11.6802, "num_token_query": 31.3924, "num_token_union": 65.1198, "num_word_context": 202.3349, "num_word_doc": 49.8109, "num_word_query": 23.318, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13523.186, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1448, "query_norm": 1.2574, "queue_k_norm": 1.3737, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3924, "sent_len_1": 66.7593, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.9963, "stdk": 0.0476, "stdq": 0.0426, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31200 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.8284, "doc_norm": 1.3706, "encoder_q-embeddings": 62017.3086, "encoder_q-layer.0": 40966.1289, "encoder_q-layer.1": 40551.8047, "encoder_q-layer.10": 7428.7495, "encoder_q-layer.11": 17825.8242, "encoder_q-layer.2": 43258.9336, "encoder_q-layer.3": 41799.4688, "encoder_q-layer.4": 37783.6875, "encoder_q-layer.5": 36010.7617, "encoder_q-layer.6": 38854.2969, "encoder_q-layer.7": 25876.0762, "encoder_q-layer.8": 11540.7803, "encoder_q-layer.9": 7301.1045, "epoch": 0.14, "inbatch_neg_score": 0.1439, "inbatch_pos_score": 0.7178, "learning_rate": 4.4394736842105264e-05, "loss": 3.8284, "norm_diff": 0.0928, "norm_loss": 0.0, "num_token_doc": 66.6922, "num_token_overlap": 11.6361, "num_token_query": 31.3232, "num_token_union": 65.0879, "num_word_context": 202.2357, "num_word_doc": 49.7635, "num_word_query": 23.264, "postclip_grad_norm": 1.0, "preclip_grad_norm": 56461.1138, "preclip_grad_norm_avg": 0.0005, "q@queue_neg_score": 0.144, "query_norm": 1.2777, "queue_k_norm": 1.3717, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3232, "sent_len_1": 66.6922, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.3613, "stdk": 0.0476, "stdq": 0.0433, "stdqueue_k": 0.0476, "stdqueue_q": 0.0, "step": 31300 }, { "accuracy": 44.2871, "active_queue_size": 16384.0, "cl_loss": 3.8499, "doc_norm": 1.371, "encoder_q-embeddings": 6409.9185, "encoder_q-layer.0": 4190.8662, "encoder_q-layer.1": 4376.8608, "encoder_q-layer.10": 7515.168, "encoder_q-layer.11": 19562.5879, "encoder_q-layer.2": 4897.4829, "encoder_q-layer.3": 4954.7969, "encoder_q-layer.4": 5291.3623, "encoder_q-layer.5": 5363.0342, "encoder_q-layer.6": 5995.3345, "encoder_q-layer.7": 6827.9731, "encoder_q-layer.8": 8127.647, "encoder_q-layer.9": 6992.165, "epoch": 0.14, "inbatch_neg_score": 0.1411, "inbatch_pos_score": 0.6787, "learning_rate": 4.436842105263158e-05, "loss": 3.8499, "norm_diff": 0.1132, "norm_loss": 0.0, "num_token_doc": 66.6303, "num_token_overlap": 11.6519, "num_token_query": 31.3561, "num_token_union": 65.034, "num_word_context": 202.3835, "num_word_doc": 49.7207, "num_word_query": 23.299, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12316.408, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1412, "query_norm": 1.2578, "queue_k_norm": 1.3748, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3561, "sent_len_1": 66.6303, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.68, "stdk": 0.0476, "stdq": 0.0425, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 31400 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.8095, "doc_norm": 1.3715, "encoder_q-embeddings": 7143.8843, "encoder_q-layer.0": 4822.8193, "encoder_q-layer.1": 5144.6997, "encoder_q-layer.10": 8324.1367, "encoder_q-layer.11": 19409.5918, "encoder_q-layer.2": 5990.064, "encoder_q-layer.3": 6191.2954, "encoder_q-layer.4": 6465.2988, "encoder_q-layer.5": 6239.8892, "encoder_q-layer.6": 6469.8706, "encoder_q-layer.7": 7164.6987, "encoder_q-layer.8": 7786.0107, "encoder_q-layer.9": 7544.127, "epoch": 0.14, "inbatch_neg_score": 0.1384, "inbatch_pos_score": 0.6743, "learning_rate": 4.4342105263157894e-05, "loss": 3.8095, "norm_diff": 0.1289, "norm_loss": 0.0, "num_token_doc": 66.7582, "num_token_overlap": 11.6906, "num_token_query": 31.4133, "num_token_union": 65.1351, "num_word_context": 202.2535, "num_word_doc": 49.7957, "num_word_query": 23.3329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12400.8513, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1384, "query_norm": 1.2426, "queue_k_norm": 1.3731, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4133, "sent_len_1": 66.7582, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.9512, "stdk": 0.0476, "stdq": 0.042, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31500 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.8113, "doc_norm": 1.3744, "encoder_q-embeddings": 7544.3008, "encoder_q-layer.0": 5035.4341, "encoder_q-layer.1": 5139.0278, "encoder_q-layer.10": 7042.3467, "encoder_q-layer.11": 17860.1172, "encoder_q-layer.2": 5556.7349, "encoder_q-layer.3": 5526.7153, "encoder_q-layer.4": 5407.3594, "encoder_q-layer.5": 5398.8076, "encoder_q-layer.6": 5752.6509, "encoder_q-layer.7": 6097.3271, "encoder_q-layer.8": 7238.0537, "encoder_q-layer.9": 6630.7915, "epoch": 0.14, "inbatch_neg_score": 0.134, "inbatch_pos_score": 0.6963, "learning_rate": 4.431578947368421e-05, "loss": 3.8113, "norm_diff": 0.1185, "norm_loss": 0.0, "num_token_doc": 66.8055, "num_token_overlap": 11.6799, "num_token_query": 31.3914, "num_token_union": 65.1417, "num_word_context": 202.3018, "num_word_doc": 49.8414, "num_word_query": 23.2942, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11817.8873, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1343, "query_norm": 1.2558, "queue_k_norm": 1.3743, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3914, "sent_len_1": 66.8055, "sent_len_max_0": 127.99, "sent_len_max_1": 208.7512, "stdk": 0.0477, "stdq": 0.0425, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 31600 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.8029, "doc_norm": 1.3733, "encoder_q-embeddings": 6417.9824, "encoder_q-layer.0": 4102.5562, "encoder_q-layer.1": 4504.2646, "encoder_q-layer.10": 7125.4995, "encoder_q-layer.11": 16523.168, "encoder_q-layer.2": 5141.02, "encoder_q-layer.3": 5676.5396, "encoder_q-layer.4": 6012.7549, "encoder_q-layer.5": 6065.5215, "encoder_q-layer.6": 6396.4941, "encoder_q-layer.7": 6544.3628, "encoder_q-layer.8": 7708.9761, "encoder_q-layer.9": 6753.7427, "epoch": 0.14, "inbatch_neg_score": 0.1295, "inbatch_pos_score": 0.7065, "learning_rate": 4.428947368421053e-05, "loss": 3.8029, "norm_diff": 0.0989, "norm_loss": 0.0, "num_token_doc": 67.0106, "num_token_overlap": 11.7068, "num_token_query": 31.3829, "num_token_union": 65.2171, "num_word_context": 202.4282, "num_word_doc": 49.9937, "num_word_query": 23.3091, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11290.4965, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1305, "query_norm": 1.2744, "queue_k_norm": 1.3726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3829, "sent_len_1": 67.0106, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.1962, "stdk": 0.0477, "stdq": 0.0432, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31700 }, { "accuracy": 44.1895, "active_queue_size": 16384.0, "cl_loss": 3.799, "doc_norm": 1.3716, "encoder_q-embeddings": 7299.4492, "encoder_q-layer.0": 4892.7095, "encoder_q-layer.1": 4930.2031, "encoder_q-layer.10": 8917.7227, "encoder_q-layer.11": 20066.7422, "encoder_q-layer.2": 5377.7954, "encoder_q-layer.3": 5684.6299, "encoder_q-layer.4": 5753.8462, "encoder_q-layer.5": 5833.4658, "encoder_q-layer.6": 6654.7998, "encoder_q-layer.7": 7611.0493, "encoder_q-layer.8": 9112.3652, "encoder_q-layer.9": 8452.5918, "epoch": 0.14, "inbatch_neg_score": 0.1326, "inbatch_pos_score": 0.686, "learning_rate": 4.426315789473684e-05, "loss": 3.799, "norm_diff": 0.1005, "norm_loss": 0.0, "num_token_doc": 66.8282, "num_token_overlap": 11.6932, "num_token_query": 31.4071, "num_token_union": 65.1721, "num_word_context": 202.3678, "num_word_doc": 49.8536, "num_word_query": 23.314, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13015.0776, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.132, "query_norm": 1.2711, "queue_k_norm": 1.3734, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4071, "sent_len_1": 66.8282, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.1825, "stdk": 0.0476, "stdq": 0.0431, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 31800 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.8111, "doc_norm": 1.3753, "encoder_q-embeddings": 6492.104, "encoder_q-layer.0": 4415.0127, "encoder_q-layer.1": 4545.2231, "encoder_q-layer.10": 7305.7212, "encoder_q-layer.11": 17421.7461, "encoder_q-layer.2": 5061.6543, "encoder_q-layer.3": 5187.1069, "encoder_q-layer.4": 5437.6934, "encoder_q-layer.5": 5333.4541, "encoder_q-layer.6": 5969.8408, "encoder_q-layer.7": 6191.188, "encoder_q-layer.8": 7385.4531, "encoder_q-layer.9": 6622.5645, "epoch": 0.14, "inbatch_neg_score": 0.1304, "inbatch_pos_score": 0.7192, "learning_rate": 4.423684210526316e-05, "loss": 3.8111, "norm_diff": 0.0994, "norm_loss": 0.0, "num_token_doc": 66.7659, "num_token_overlap": 11.7049, "num_token_query": 31.4866, "num_token_union": 65.1323, "num_word_context": 202.3836, "num_word_doc": 49.8243, "num_word_query": 23.3922, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11283.527, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1306, "query_norm": 1.2758, "queue_k_norm": 1.372, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4866, "sent_len_1": 66.7659, "sent_len_max_0": 127.985, "sent_len_max_1": 208.835, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 31900 }, { "accuracy": 44.6777, "active_queue_size": 16384.0, "cl_loss": 3.8116, "doc_norm": 1.3689, "encoder_q-embeddings": 8191.6655, "encoder_q-layer.0": 5550.0483, "encoder_q-layer.1": 6301.9873, "encoder_q-layer.10": 7338.7983, "encoder_q-layer.11": 18196.0977, "encoder_q-layer.2": 6796.2285, "encoder_q-layer.3": 6506.7266, "encoder_q-layer.4": 6381.2104, "encoder_q-layer.5": 6331.8599, "encoder_q-layer.6": 7194.8682, "encoder_q-layer.7": 7248.5986, "encoder_q-layer.8": 8151.9033, "encoder_q-layer.9": 6937.9272, "epoch": 0.14, "inbatch_neg_score": 0.1265, "inbatch_pos_score": 0.6807, "learning_rate": 4.421052631578947e-05, "loss": 3.8116, "norm_diff": 0.0987, "norm_loss": 0.0, "num_token_doc": 66.644, "num_token_overlap": 11.6829, "num_token_query": 31.4096, "num_token_union": 65.0425, "num_word_context": 202.0772, "num_word_doc": 49.7265, "num_word_query": 23.3336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12759.9253, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1265, "query_norm": 1.2702, "queue_k_norm": 1.3707, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4096, "sent_len_1": 66.644, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.6625, "stdk": 0.0476, "stdq": 0.0428, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32000 }, { "accuracy": 46.7773, "active_queue_size": 16384.0, "cl_loss": 3.8058, "doc_norm": 1.3797, "encoder_q-embeddings": 6679.6411, "encoder_q-layer.0": 4346.6255, "encoder_q-layer.1": 4438.0664, "encoder_q-layer.10": 7252.2153, "encoder_q-layer.11": 17790.8438, "encoder_q-layer.2": 5006.3813, "encoder_q-layer.3": 5297.8828, "encoder_q-layer.4": 5657.3184, "encoder_q-layer.5": 5664.4819, "encoder_q-layer.6": 6559.979, "encoder_q-layer.7": 7222.439, "encoder_q-layer.8": 7956.9722, "encoder_q-layer.9": 6798.043, "epoch": 0.14, "inbatch_neg_score": 0.1267, "inbatch_pos_score": 0.6865, "learning_rate": 4.418421052631579e-05, "loss": 3.8058, "norm_diff": 0.1082, "norm_loss": 0.0, "num_token_doc": 66.7943, "num_token_overlap": 11.6774, "num_token_query": 31.3502, "num_token_union": 65.092, "num_word_context": 202.3289, "num_word_doc": 49.8439, "num_word_query": 23.2862, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11655.8734, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1265, "query_norm": 1.2715, "queue_k_norm": 1.3709, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3502, "sent_len_1": 66.7943, "sent_len_max_0": 127.9562, "sent_len_max_1": 207.0788, "stdk": 0.048, "stdq": 0.0426, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32100 }, { "accuracy": 44.9707, "active_queue_size": 16384.0, "cl_loss": 3.8116, "doc_norm": 1.3716, "encoder_q-embeddings": 6233.0156, "encoder_q-layer.0": 4123.5859, "encoder_q-layer.1": 4346.8589, "encoder_q-layer.10": 7497.4746, "encoder_q-layer.11": 17077.6289, "encoder_q-layer.2": 4922.8022, "encoder_q-layer.3": 5065.9194, "encoder_q-layer.4": 5180.479, "encoder_q-layer.5": 5244.1533, "encoder_q-layer.6": 5838.4922, "encoder_q-layer.7": 6549.4043, "encoder_q-layer.8": 7648.3882, "encoder_q-layer.9": 6993.3643, "epoch": 0.14, "inbatch_neg_score": 0.1254, "inbatch_pos_score": 0.6772, "learning_rate": 4.415789473684211e-05, "loss": 3.8116, "norm_diff": 0.0902, "norm_loss": 0.0, "num_token_doc": 66.7042, "num_token_overlap": 11.6526, "num_token_query": 31.3868, "num_token_union": 65.0882, "num_word_context": 202.1632, "num_word_doc": 49.7593, "num_word_query": 23.3065, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11133.7265, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1255, "query_norm": 1.2814, "queue_k_norm": 1.3711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3868, "sent_len_1": 66.7042, "sent_len_max_0": 127.99, "sent_len_max_1": 208.3487, "stdk": 0.0477, "stdq": 0.0429, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32200 }, { "accuracy": 45.5566, "active_queue_size": 16384.0, "cl_loss": 3.8142, "doc_norm": 1.3695, "encoder_q-embeddings": 13553.6475, "encoder_q-layer.0": 8496.4209, "encoder_q-layer.1": 8884.5195, "encoder_q-layer.10": 14187.5127, "encoder_q-layer.11": 35912.8438, "encoder_q-layer.2": 9816.9932, "encoder_q-layer.3": 10700.3896, "encoder_q-layer.4": 11356.7949, "encoder_q-layer.5": 11269.5664, "encoder_q-layer.6": 12945.1582, "encoder_q-layer.7": 14256.0332, "encoder_q-layer.8": 16706.1348, "encoder_q-layer.9": 14277.21, "epoch": 0.14, "inbatch_neg_score": 0.1275, "inbatch_pos_score": 0.6816, "learning_rate": 4.413157894736842e-05, "loss": 3.8142, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.6134, "num_token_overlap": 11.7189, "num_token_query": 31.5218, "num_token_union": 65.0786, "num_word_context": 202.0412, "num_word_doc": 49.7174, "num_word_query": 23.4288, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24135.28, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1277, "query_norm": 1.295, "queue_k_norm": 1.3701, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5218, "sent_len_1": 66.6134, "sent_len_max_0": 127.99, "sent_len_max_1": 208.9275, "stdk": 0.0476, "stdq": 0.0433, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32300 }, { "accuracy": 46.2402, "active_queue_size": 16384.0, "cl_loss": 3.7913, "doc_norm": 1.3692, "encoder_q-embeddings": 11962.5303, "encoder_q-layer.0": 8167.8306, "encoder_q-layer.1": 8739.082, "encoder_q-layer.10": 14269.7402, "encoder_q-layer.11": 35136.0938, "encoder_q-layer.2": 9528.3457, "encoder_q-layer.3": 9860.7715, "encoder_q-layer.4": 10372.9004, "encoder_q-layer.5": 10973.8262, "encoder_q-layer.6": 11898.1396, "encoder_q-layer.7": 14262.6035, "encoder_q-layer.8": 15895.0127, "encoder_q-layer.9": 13692.6104, "epoch": 0.14, "inbatch_neg_score": 0.1293, "inbatch_pos_score": 0.6826, "learning_rate": 4.410526315789474e-05, "loss": 3.7913, "norm_diff": 0.093, "norm_loss": 0.0, "num_token_doc": 66.7073, "num_token_overlap": 11.6883, "num_token_query": 31.4009, "num_token_union": 65.0961, "num_word_context": 202.3426, "num_word_doc": 49.7706, "num_word_query": 23.3137, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22508.1107, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1296, "query_norm": 1.2762, "queue_k_norm": 1.3705, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4009, "sent_len_1": 66.7073, "sent_len_max_0": 128.0, "sent_len_max_1": 209.2262, "stdk": 0.0477, "stdq": 0.0424, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 32400 }, { "accuracy": 45.752, "active_queue_size": 16384.0, "cl_loss": 3.8015, "doc_norm": 1.3676, "encoder_q-embeddings": 8032.7568, "encoder_q-layer.0": 5385.6831, "encoder_q-layer.1": 5932.252, "encoder_q-layer.10": 7898.0132, "encoder_q-layer.11": 18259.7734, "encoder_q-layer.2": 6480.3706, "encoder_q-layer.3": 6218.353, "encoder_q-layer.4": 6347.7827, "encoder_q-layer.5": 5988.1689, "encoder_q-layer.6": 6504.6284, "encoder_q-layer.7": 6583.7524, "encoder_q-layer.8": 7786.8608, "encoder_q-layer.9": 7206.6885, "epoch": 0.14, "inbatch_neg_score": 0.129, "inbatch_pos_score": 0.6875, "learning_rate": 4.407894736842105e-05, "loss": 3.8015, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.8747, "num_token_overlap": 11.6677, "num_token_query": 31.3622, "num_token_union": 65.1536, "num_word_context": 202.4118, "num_word_doc": 49.8784, "num_word_query": 23.2992, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12338.7534, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1292, "query_norm": 1.2846, "queue_k_norm": 1.368, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3622, "sent_len_1": 66.8747, "sent_len_max_0": 127.975, "sent_len_max_1": 209.0075, "stdk": 0.0476, "stdq": 0.0431, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32500 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.793, "doc_norm": 1.3736, "encoder_q-embeddings": 5937.085, "encoder_q-layer.0": 3777.7227, "encoder_q-layer.1": 3929.5977, "encoder_q-layer.10": 7300.9995, "encoder_q-layer.11": 17981.6699, "encoder_q-layer.2": 4539.833, "encoder_q-layer.3": 4824.8525, "encoder_q-layer.4": 5068.501, "encoder_q-layer.5": 5169.8594, "encoder_q-layer.6": 5823.8643, "encoder_q-layer.7": 6411.3506, "encoder_q-layer.8": 7625.1323, "encoder_q-layer.9": 7053.041, "epoch": 0.14, "inbatch_neg_score": 0.1305, "inbatch_pos_score": 0.6948, "learning_rate": 4.4052631578947376e-05, "loss": 3.793, "norm_diff": 0.0989, "norm_loss": 0.0, "num_token_doc": 66.8332, "num_token_overlap": 11.6978, "num_token_query": 31.3695, "num_token_union": 65.1046, "num_word_context": 201.9597, "num_word_doc": 49.8482, "num_word_query": 23.28, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11449.6629, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1309, "query_norm": 1.2748, "queue_k_norm": 1.3709, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3695, "sent_len_1": 66.8332, "sent_len_max_0": 128.0, "sent_len_max_1": 208.385, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 32600 }, { "accuracy": 45.8008, "active_queue_size": 16384.0, "cl_loss": 3.7961, "doc_norm": 1.3724, "encoder_q-embeddings": 6876.8472, "encoder_q-layer.0": 4591.4492, "encoder_q-layer.1": 4949.4736, "encoder_q-layer.10": 7287.7446, "encoder_q-layer.11": 18220.3828, "encoder_q-layer.2": 5502.2017, "encoder_q-layer.3": 5697.9795, "encoder_q-layer.4": 5946.5312, "encoder_q-layer.5": 5911.5908, "encoder_q-layer.6": 6141.8071, "encoder_q-layer.7": 6596.4048, "encoder_q-layer.8": 7801.0703, "encoder_q-layer.9": 6883.4438, "epoch": 0.14, "inbatch_neg_score": 0.1381, "inbatch_pos_score": 0.6855, "learning_rate": 4.402631578947369e-05, "loss": 3.7961, "norm_diff": 0.1107, "norm_loss": 0.0, "num_token_doc": 66.7786, "num_token_overlap": 11.6727, "num_token_query": 31.3791, "num_token_union": 65.1251, "num_word_context": 202.4969, "num_word_doc": 49.8413, "num_word_query": 23.3156, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12214.2046, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1381, "query_norm": 1.2617, "queue_k_norm": 1.3691, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3791, "sent_len_1": 66.7786, "sent_len_max_0": 127.9537, "sent_len_max_1": 209.1575, "stdk": 0.0478, "stdq": 0.0425, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32700 }, { "accuracy": 46.3867, "active_queue_size": 16384.0, "cl_loss": 3.8283, "doc_norm": 1.3682, "encoder_q-embeddings": 6252.1333, "encoder_q-layer.0": 3940.5825, "encoder_q-layer.1": 4053.1802, "encoder_q-layer.10": 7495.917, "encoder_q-layer.11": 18514.3066, "encoder_q-layer.2": 4483.3374, "encoder_q-layer.3": 4813.1055, "encoder_q-layer.4": 5147.7017, "encoder_q-layer.5": 5281.9585, "encoder_q-layer.6": 5839.439, "encoder_q-layer.7": 6914.0176, "encoder_q-layer.8": 7799.5435, "encoder_q-layer.9": 6818.9238, "epoch": 0.14, "inbatch_neg_score": 0.1369, "inbatch_pos_score": 0.7041, "learning_rate": 4.4000000000000006e-05, "loss": 3.8283, "norm_diff": 0.0881, "norm_loss": 0.0, "num_token_doc": 66.8044, "num_token_overlap": 11.6491, "num_token_query": 31.3126, "num_token_union": 65.1472, "num_word_context": 202.4458, "num_word_doc": 49.8203, "num_word_query": 23.2478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11623.0644, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1378, "query_norm": 1.2801, "queue_k_norm": 1.3695, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3126, "sent_len_1": 66.8044, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.6513, "stdk": 0.0477, "stdq": 0.0434, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 32800 }, { "accuracy": 45.459, "active_queue_size": 16384.0, "cl_loss": 3.7914, "doc_norm": 1.3701, "encoder_q-embeddings": 7144.0264, "encoder_q-layer.0": 4615.9287, "encoder_q-layer.1": 4700.8091, "encoder_q-layer.10": 6891.3066, "encoder_q-layer.11": 17027.3398, "encoder_q-layer.2": 5269.3159, "encoder_q-layer.3": 5637.0542, "encoder_q-layer.4": 5775.9263, "encoder_q-layer.5": 5917.0605, "encoder_q-layer.6": 6415.8623, "encoder_q-layer.7": 6957.7544, "encoder_q-layer.8": 8119.625, "encoder_q-layer.9": 6914.623, "epoch": 0.14, "inbatch_neg_score": 0.1388, "inbatch_pos_score": 0.6895, "learning_rate": 4.397368421052632e-05, "loss": 3.7914, "norm_diff": 0.1018, "norm_loss": 0.0, "num_token_doc": 66.8072, "num_token_overlap": 11.6867, "num_token_query": 31.3307, "num_token_union": 65.0598, "num_word_context": 202.2467, "num_word_doc": 49.8348, "num_word_query": 23.2561, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11562.3802, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1395, "query_norm": 1.2683, "queue_k_norm": 1.3688, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3307, "sent_len_1": 66.8072, "sent_len_max_0": 127.9537, "sent_len_max_1": 207.46, "stdk": 0.0477, "stdq": 0.043, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 32900 }, { "accuracy": 45.6055, "active_queue_size": 16384.0, "cl_loss": 3.7959, "doc_norm": 1.3681, "encoder_q-embeddings": 7361.7383, "encoder_q-layer.0": 4885.6343, "encoder_q-layer.1": 5123.48, "encoder_q-layer.10": 7456.3208, "encoder_q-layer.11": 18175.5977, "encoder_q-layer.2": 5753.3179, "encoder_q-layer.3": 6098.2021, "encoder_q-layer.4": 6538.4214, "encoder_q-layer.5": 6615.6182, "encoder_q-layer.6": 7031.7607, "encoder_q-layer.7": 7196.585, "encoder_q-layer.8": 8544.1729, "encoder_q-layer.9": 7179.1562, "epoch": 0.14, "inbatch_neg_score": 0.1367, "inbatch_pos_score": 0.6963, "learning_rate": 4.394736842105263e-05, "loss": 3.7959, "norm_diff": 0.0955, "norm_loss": 0.0, "num_token_doc": 66.7492, "num_token_overlap": 11.6197, "num_token_query": 31.2419, "num_token_union": 65.0687, "num_word_context": 202.3911, "num_word_doc": 49.8325, "num_word_query": 23.1977, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12521.0463, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1356, "query_norm": 1.2726, "queue_k_norm": 1.3709, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2419, "sent_len_1": 66.7492, "sent_len_max_0": 127.9475, "sent_len_max_1": 207.4025, "stdk": 0.0476, "stdq": 0.0431, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33000 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.7879, "doc_norm": 1.3685, "encoder_q-embeddings": 5784.3384, "encoder_q-layer.0": 3686.4541, "encoder_q-layer.1": 3916.6472, "encoder_q-layer.10": 6979.939, "encoder_q-layer.11": 16829.1699, "encoder_q-layer.2": 4313.395, "encoder_q-layer.3": 4458.0029, "encoder_q-layer.4": 4863.832, "encoder_q-layer.5": 4883.4683, "encoder_q-layer.6": 5562.0801, "encoder_q-layer.7": 6258.165, "encoder_q-layer.8": 7225.7808, "encoder_q-layer.9": 6592.3643, "epoch": 0.14, "inbatch_neg_score": 0.1333, "inbatch_pos_score": 0.6885, "learning_rate": 4.3921052631578954e-05, "loss": 3.7879, "norm_diff": 0.1279, "norm_loss": 0.0, "num_token_doc": 66.8202, "num_token_overlap": 11.6966, "num_token_query": 31.4806, "num_token_union": 65.205, "num_word_context": 202.6473, "num_word_doc": 49.86, "num_word_query": 23.3873, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10731.3217, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1323, "query_norm": 1.2406, "queue_k_norm": 1.3719, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4806, "sent_len_1": 66.8202, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.4975, "stdk": 0.0476, "stdq": 0.0421, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33100 }, { "accuracy": 47.0703, "active_queue_size": 16384.0, "cl_loss": 3.788, "doc_norm": 1.3691, "encoder_q-embeddings": 6898.6118, "encoder_q-layer.0": 4406.5171, "encoder_q-layer.1": 4919.6016, "encoder_q-layer.10": 8290.8594, "encoder_q-layer.11": 18742.4375, "encoder_q-layer.2": 5652.2583, "encoder_q-layer.3": 5825.3735, "encoder_q-layer.4": 6243.3452, "encoder_q-layer.5": 6260.1465, "encoder_q-layer.6": 6956.9351, "encoder_q-layer.7": 7205.3154, "encoder_q-layer.8": 8273.1953, "encoder_q-layer.9": 7757.0488, "epoch": 0.14, "inbatch_neg_score": 0.1327, "inbatch_pos_score": 0.7148, "learning_rate": 4.3894736842105266e-05, "loss": 3.788, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.6952, "num_token_overlap": 11.682, "num_token_query": 31.3077, "num_token_union": 65.0134, "num_word_context": 202.3278, "num_word_doc": 49.7735, "num_word_query": 23.2565, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12044.4577, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1331, "query_norm": 1.2889, "queue_k_norm": 1.3715, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3077, "sent_len_1": 66.6952, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.645, "stdk": 0.0477, "stdq": 0.0438, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33200 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.8055, "doc_norm": 1.3698, "encoder_q-embeddings": 5828.3657, "encoder_q-layer.0": 3902.0903, "encoder_q-layer.1": 3981.1301, "encoder_q-layer.10": 8808.3027, "encoder_q-layer.11": 18799.6348, "encoder_q-layer.2": 4453.3311, "encoder_q-layer.3": 4637.5801, "encoder_q-layer.4": 5073.9653, "encoder_q-layer.5": 5034.6621, "encoder_q-layer.6": 5617.3096, "encoder_q-layer.7": 6325.0938, "encoder_q-layer.8": 8377.3379, "encoder_q-layer.9": 7817.1699, "epoch": 0.14, "inbatch_neg_score": 0.1347, "inbatch_pos_score": 0.6953, "learning_rate": 4.3868421052631584e-05, "loss": 3.8055, "norm_diff": 0.1238, "norm_loss": 0.0, "num_token_doc": 66.5817, "num_token_overlap": 11.6325, "num_token_query": 31.2992, "num_token_union": 65.0265, "num_word_context": 202.1754, "num_word_doc": 49.6793, "num_word_query": 23.2373, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11404.2115, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1349, "query_norm": 1.246, "queue_k_norm": 1.3704, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2992, "sent_len_1": 66.5817, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.8862, "stdk": 0.0477, "stdq": 0.0422, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33300 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.7868, "doc_norm": 1.373, "encoder_q-embeddings": 5966.0576, "encoder_q-layer.0": 3806.2527, "encoder_q-layer.1": 3937.7893, "encoder_q-layer.10": 8223.9834, "encoder_q-layer.11": 17240.5684, "encoder_q-layer.2": 4426.4019, "encoder_q-layer.3": 4713.043, "encoder_q-layer.4": 5188.8462, "encoder_q-layer.5": 5262.2339, "encoder_q-layer.6": 5852.8032, "encoder_q-layer.7": 6512.8262, "encoder_q-layer.8": 8453.2803, "encoder_q-layer.9": 7847.5708, "epoch": 0.14, "inbatch_neg_score": 0.1297, "inbatch_pos_score": 0.7061, "learning_rate": 4.3842105263157895e-05, "loss": 3.7868, "norm_diff": 0.1014, "norm_loss": 0.0, "num_token_doc": 66.7871, "num_token_overlap": 11.7085, "num_token_query": 31.3726, "num_token_union": 65.0885, "num_word_context": 202.1805, "num_word_doc": 49.8251, "num_word_query": 23.3111, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11032.3695, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1299, "query_norm": 1.2716, "queue_k_norm": 1.3706, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3726, "sent_len_1": 66.7871, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.6513, "stdk": 0.0478, "stdq": 0.0432, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33400 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.7735, "doc_norm": 1.3704, "encoder_q-embeddings": 5932.7339, "encoder_q-layer.0": 3909.3274, "encoder_q-layer.1": 4196.0742, "encoder_q-layer.10": 7716.6006, "encoder_q-layer.11": 17832.1484, "encoder_q-layer.2": 4659.2231, "encoder_q-layer.3": 4911.7178, "encoder_q-layer.4": 4831.8374, "encoder_q-layer.5": 4964.1768, "encoder_q-layer.6": 5682.5371, "encoder_q-layer.7": 6063.6538, "encoder_q-layer.8": 7508.8643, "encoder_q-layer.9": 6820.7822, "epoch": 0.15, "inbatch_neg_score": 0.1314, "inbatch_pos_score": 0.7007, "learning_rate": 4.381578947368421e-05, "loss": 3.7735, "norm_diff": 0.112, "norm_loss": 0.0, "num_token_doc": 66.7995, "num_token_overlap": 11.6864, "num_token_query": 31.3803, "num_token_union": 65.1239, "num_word_context": 202.2932, "num_word_doc": 49.8624, "num_word_query": 23.3082, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11180.0285, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1315, "query_norm": 1.2584, "queue_k_norm": 1.3708, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3803, "sent_len_1": 66.7995, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.9837, "stdk": 0.0477, "stdq": 0.0426, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33500 }, { "accuracy": 45.5078, "active_queue_size": 16384.0, "cl_loss": 3.7827, "doc_norm": 1.3685, "encoder_q-embeddings": 6730.7119, "encoder_q-layer.0": 4537.5493, "encoder_q-layer.1": 5139.957, "encoder_q-layer.10": 6751.1636, "encoder_q-layer.11": 16459.1699, "encoder_q-layer.2": 5694.7402, "encoder_q-layer.3": 5781.2539, "encoder_q-layer.4": 5846.0127, "encoder_q-layer.5": 5877.2573, "encoder_q-layer.6": 6591.6851, "encoder_q-layer.7": 6872.5918, "encoder_q-layer.8": 7746.4355, "encoder_q-layer.9": 6525.1426, "epoch": 0.15, "inbatch_neg_score": 0.1267, "inbatch_pos_score": 0.6743, "learning_rate": 4.378947368421053e-05, "loss": 3.7827, "norm_diff": 0.1189, "norm_loss": 0.0, "num_token_doc": 66.6368, "num_token_overlap": 11.6472, "num_token_query": 31.3443, "num_token_union": 65.0487, "num_word_context": 202.1885, "num_word_doc": 49.7049, "num_word_query": 23.2683, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11515.1895, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1262, "query_norm": 1.2497, "queue_k_norm": 1.3703, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3443, "sent_len_1": 66.6368, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.0975, "stdk": 0.0476, "stdq": 0.0424, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33600 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.8016, "doc_norm": 1.3686, "encoder_q-embeddings": 7438.4624, "encoder_q-layer.0": 4760.3281, "encoder_q-layer.1": 5352.8857, "encoder_q-layer.10": 7268.2485, "encoder_q-layer.11": 16418.793, "encoder_q-layer.2": 5962.6846, "encoder_q-layer.3": 6277.2378, "encoder_q-layer.4": 6009.7085, "encoder_q-layer.5": 5809.2563, "encoder_q-layer.6": 6054.9814, "encoder_q-layer.7": 6734.417, "encoder_q-layer.8": 7644.7305, "encoder_q-layer.9": 6684.6172, "epoch": 0.15, "inbatch_neg_score": 0.1272, "inbatch_pos_score": 0.6855, "learning_rate": 4.3763157894736844e-05, "loss": 3.8016, "norm_diff": 0.1162, "norm_loss": 0.0, "num_token_doc": 66.4435, "num_token_overlap": 11.6199, "num_token_query": 31.3325, "num_token_union": 64.9418, "num_word_context": 202.1681, "num_word_doc": 49.5845, "num_word_query": 23.2791, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11690.2095, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1276, "query_norm": 1.2524, "queue_k_norm": 1.3689, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3325, "sent_len_1": 66.4435, "sent_len_max_0": 127.97, "sent_len_max_1": 208.9712, "stdk": 0.0477, "stdq": 0.0425, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 33700 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.7885, "doc_norm": 1.3696, "encoder_q-embeddings": 5888.5713, "encoder_q-layer.0": 3878.3708, "encoder_q-layer.1": 4203.4653, "encoder_q-layer.10": 7772.9141, "encoder_q-layer.11": 17033.8516, "encoder_q-layer.2": 4579.0645, "encoder_q-layer.3": 4716.4136, "encoder_q-layer.4": 5110.0171, "encoder_q-layer.5": 4949.8052, "encoder_q-layer.6": 5544.7988, "encoder_q-layer.7": 6415.9351, "encoder_q-layer.8": 7652.9756, "encoder_q-layer.9": 6932.8037, "epoch": 0.15, "inbatch_neg_score": 0.1322, "inbatch_pos_score": 0.7109, "learning_rate": 4.373684210526316e-05, "loss": 3.7885, "norm_diff": 0.1082, "norm_loss": 0.0, "num_token_doc": 66.8746, "num_token_overlap": 11.6982, "num_token_query": 31.425, "num_token_union": 65.1955, "num_word_context": 202.3587, "num_word_doc": 49.9162, "num_word_query": 23.3424, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10918.6761, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1321, "query_norm": 1.2613, "queue_k_norm": 1.3701, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.425, "sent_len_1": 66.8746, "sent_len_max_0": 127.9513, "sent_len_max_1": 206.7425, "stdk": 0.0477, "stdq": 0.0427, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33800 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.7767, "doc_norm": 1.3704, "encoder_q-embeddings": 6516.395, "encoder_q-layer.0": 4330.6875, "encoder_q-layer.1": 4536.8276, "encoder_q-layer.10": 6758.644, "encoder_q-layer.11": 16415.7812, "encoder_q-layer.2": 5244.4517, "encoder_q-layer.3": 5368.1255, "encoder_q-layer.4": 5794.9897, "encoder_q-layer.5": 5932.8418, "encoder_q-layer.6": 6134.9639, "encoder_q-layer.7": 6454.6787, "encoder_q-layer.8": 7829.6792, "encoder_q-layer.9": 6883.1377, "epoch": 0.15, "inbatch_neg_score": 0.1286, "inbatch_pos_score": 0.6978, "learning_rate": 4.3710526315789474e-05, "loss": 3.7767, "norm_diff": 0.0995, "norm_loss": 0.0, "num_token_doc": 66.8593, "num_token_overlap": 11.7087, "num_token_query": 31.443, "num_token_union": 65.1715, "num_word_context": 202.2099, "num_word_doc": 49.8822, "num_word_query": 23.3579, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11190.233, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1292, "query_norm": 1.2709, "queue_k_norm": 1.3696, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.443, "sent_len_1": 66.8593, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.7337, "stdk": 0.0478, "stdq": 0.043, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 33900 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.772, "doc_norm": 1.3713, "encoder_q-embeddings": 6556.1646, "encoder_q-layer.0": 4375.769, "encoder_q-layer.1": 4752.8813, "encoder_q-layer.10": 7011.7715, "encoder_q-layer.11": 16910.0625, "encoder_q-layer.2": 5304.918, "encoder_q-layer.3": 5425.4868, "encoder_q-layer.4": 5603.1953, "encoder_q-layer.5": 5666.9233, "encoder_q-layer.6": 6704.3403, "encoder_q-layer.7": 7068.0298, "encoder_q-layer.8": 8015.1245, "encoder_q-layer.9": 6790.4214, "epoch": 0.15, "inbatch_neg_score": 0.1283, "inbatch_pos_score": 0.6929, "learning_rate": 4.368421052631579e-05, "loss": 3.772, "norm_diff": 0.094, "norm_loss": 0.0, "num_token_doc": 66.6739, "num_token_overlap": 11.6741, "num_token_query": 31.3758, "num_token_union": 65.073, "num_word_context": 202.2702, "num_word_doc": 49.7565, "num_word_query": 23.3058, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11410.4166, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1282, "query_norm": 1.2773, "queue_k_norm": 1.3695, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3758, "sent_len_1": 66.6739, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6575, "stdk": 0.0478, "stdq": 0.0429, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34000 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.7843, "doc_norm": 1.3755, "encoder_q-embeddings": 6321.3022, "encoder_q-layer.0": 4175.2373, "encoder_q-layer.1": 4263.3188, "encoder_q-layer.10": 7119.4126, "encoder_q-layer.11": 16144.7627, "encoder_q-layer.2": 4602.6104, "encoder_q-layer.3": 4793.8945, "encoder_q-layer.4": 5084.3394, "encoder_q-layer.5": 5065.5439, "encoder_q-layer.6": 5594.6841, "encoder_q-layer.7": 6247.8638, "encoder_q-layer.8": 7598.9072, "encoder_q-layer.9": 6679.7476, "epoch": 0.15, "inbatch_neg_score": 0.1332, "inbatch_pos_score": 0.7021, "learning_rate": 4.365789473684211e-05, "loss": 3.7843, "norm_diff": 0.1054, "norm_loss": 0.0, "num_token_doc": 66.6544, "num_token_overlap": 11.6475, "num_token_query": 31.3333, "num_token_union": 65.0488, "num_word_context": 202.0747, "num_word_doc": 49.7502, "num_word_query": 23.2629, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10578.9555, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1333, "query_norm": 1.2702, "queue_k_norm": 1.3709, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3333, "sent_len_1": 66.6544, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.8025, "stdk": 0.0479, "stdq": 0.0424, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34100 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.7813, "doc_norm": 1.3674, "encoder_q-embeddings": 6998.9878, "encoder_q-layer.0": 4926.2915, "encoder_q-layer.1": 4913.9258, "encoder_q-layer.10": 6949.3057, "encoder_q-layer.11": 16367.3164, "encoder_q-layer.2": 5475.0591, "encoder_q-layer.3": 5698.6909, "encoder_q-layer.4": 5941.7358, "encoder_q-layer.5": 6153.2622, "encoder_q-layer.6": 6729.3398, "encoder_q-layer.7": 6944.6475, "encoder_q-layer.8": 7990.686, "encoder_q-layer.9": 6728.585, "epoch": 0.15, "inbatch_neg_score": 0.1345, "inbatch_pos_score": 0.7007, "learning_rate": 4.363157894736842e-05, "loss": 3.7813, "norm_diff": 0.0812, "norm_loss": 0.0, "num_token_doc": 66.6394, "num_token_overlap": 11.6692, "num_token_query": 31.3578, "num_token_union": 65.0702, "num_word_context": 202.1803, "num_word_doc": 49.7322, "num_word_query": 23.2937, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11442.4595, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1339, "query_norm": 1.2861, "queue_k_norm": 1.3718, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3578, "sent_len_1": 66.6394, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.1, "stdk": 0.0476, "stdq": 0.0427, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34200 }, { "accuracy": 47.0703, "active_queue_size": 16384.0, "cl_loss": 3.7771, "doc_norm": 1.3713, "encoder_q-embeddings": 6678.0405, "encoder_q-layer.0": 4520.3164, "encoder_q-layer.1": 4876.5967, "encoder_q-layer.10": 6895.8057, "encoder_q-layer.11": 15776.1992, "encoder_q-layer.2": 5725.1655, "encoder_q-layer.3": 5552.4072, "encoder_q-layer.4": 5748.0005, "encoder_q-layer.5": 6164.7075, "encoder_q-layer.6": 6116.0093, "encoder_q-layer.7": 6880.2495, "encoder_q-layer.8": 7636.8193, "encoder_q-layer.9": 6570.7632, "epoch": 0.15, "inbatch_neg_score": 0.1392, "inbatch_pos_score": 0.7075, "learning_rate": 4.360526315789474e-05, "loss": 3.7771, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.7605, "num_token_overlap": 11.6842, "num_token_query": 31.4238, "num_token_union": 65.1311, "num_word_context": 202.208, "num_word_doc": 49.8147, "num_word_query": 23.3352, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11048.5276, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1389, "query_norm": 1.2968, "queue_k_norm": 1.3696, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4238, "sent_len_1": 66.7605, "sent_len_max_0": 128.0, "sent_len_max_1": 208.3725, "stdk": 0.0478, "stdq": 0.0428, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34300 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.79, "doc_norm": 1.372, "encoder_q-embeddings": 26122.3418, "encoder_q-layer.0": 19606.6055, "encoder_q-layer.1": 19674.8828, "encoder_q-layer.10": 6841.6084, "encoder_q-layer.11": 15924.9414, "encoder_q-layer.2": 18831.8809, "encoder_q-layer.3": 17356.6387, "encoder_q-layer.4": 15678.5088, "encoder_q-layer.5": 15747.1943, "encoder_q-layer.6": 13474.043, "encoder_q-layer.7": 10610.4121, "encoder_q-layer.8": 8682.3291, "encoder_q-layer.9": 6795.2183, "epoch": 0.15, "inbatch_neg_score": 0.1399, "inbatch_pos_score": 0.7065, "learning_rate": 4.357894736842105e-05, "loss": 3.79, "norm_diff": 0.0871, "norm_loss": 0.0, "num_token_doc": 66.7774, "num_token_overlap": 11.6779, "num_token_query": 31.3764, "num_token_union": 65.1365, "num_word_context": 202.1417, "num_word_doc": 49.8179, "num_word_query": 23.3002, "postclip_grad_norm": 1.0, "preclip_grad_norm": 25016.587, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1398, "query_norm": 1.2849, "queue_k_norm": 1.3712, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3764, "sent_len_1": 66.7774, "sent_len_max_0": 128.0, "sent_len_max_1": 206.8438, "stdk": 0.0478, "stdq": 0.0428, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34400 }, { "accuracy": 45.9473, "active_queue_size": 16384.0, "cl_loss": 3.7747, "doc_norm": 1.3715, "encoder_q-embeddings": 42561.625, "encoder_q-layer.0": 32465.7754, "encoder_q-layer.1": 30618.2637, "encoder_q-layer.10": 14911.6602, "encoder_q-layer.11": 34723.7109, "encoder_q-layer.2": 32741.418, "encoder_q-layer.3": 36910.707, "encoder_q-layer.4": 34750.2734, "encoder_q-layer.5": 33876.668, "encoder_q-layer.6": 27244.5977, "encoder_q-layer.7": 23818.6406, "encoder_q-layer.8": 16742.1582, "encoder_q-layer.9": 14170.1162, "epoch": 0.15, "inbatch_neg_score": 0.142, "inbatch_pos_score": 0.7012, "learning_rate": 4.355263157894737e-05, "loss": 3.7747, "norm_diff": 0.0836, "norm_loss": 0.0, "num_token_doc": 66.7085, "num_token_overlap": 11.6929, "num_token_query": 31.4141, "num_token_union": 65.1097, "num_word_context": 202.1147, "num_word_doc": 49.7849, "num_word_query": 23.3326, "postclip_grad_norm": 1.0, "preclip_grad_norm": 45628.0878, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 0.1416, "query_norm": 1.2879, "queue_k_norm": 1.3687, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4141, "sent_len_1": 66.7085, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.695, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.0477, "stdqueue_q": 0.0, "step": 34500 }, { "accuracy": 47.0703, "active_queue_size": 16384.0, "cl_loss": 3.7884, "doc_norm": 1.3703, "encoder_q-embeddings": 13773.7139, "encoder_q-layer.0": 8752.6631, "encoder_q-layer.1": 9263.3613, "encoder_q-layer.10": 14712.335, "encoder_q-layer.11": 34902.9062, "encoder_q-layer.2": 10318.2744, "encoder_q-layer.3": 10399.541, "encoder_q-layer.4": 11653.8564, "encoder_q-layer.5": 11622.5791, "encoder_q-layer.6": 12866.667, "encoder_q-layer.7": 13604.7793, "encoder_q-layer.8": 16177.877, "encoder_q-layer.9": 14189.0781, "epoch": 0.15, "inbatch_neg_score": 0.1479, "inbatch_pos_score": 0.7178, "learning_rate": 4.352631578947369e-05, "loss": 3.7884, "norm_diff": 0.0885, "norm_loss": 0.0, "num_token_doc": 66.6681, "num_token_overlap": 11.6704, "num_token_query": 31.3896, "num_token_union": 65.0737, "num_word_context": 202.3002, "num_word_doc": 49.7717, "num_word_query": 23.3185, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23367.1593, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1484, "query_norm": 1.2818, "queue_k_norm": 1.3708, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3896, "sent_len_1": 66.6681, "sent_len_max_0": 127.99, "sent_len_max_1": 206.6725, "stdk": 0.0477, "stdq": 0.0431, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34600 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.7606, "doc_norm": 1.376, "encoder_q-embeddings": 11962.5723, "encoder_q-layer.0": 7800.4756, "encoder_q-layer.1": 8153.3643, "encoder_q-layer.10": 13886.5088, "encoder_q-layer.11": 33777.1133, "encoder_q-layer.2": 8882.7246, "encoder_q-layer.3": 9532.043, "encoder_q-layer.4": 9852.2666, "encoder_q-layer.5": 9862.5352, "encoder_q-layer.6": 11160.75, "encoder_q-layer.7": 12809.2539, "encoder_q-layer.8": 14819.6133, "encoder_q-layer.9": 13132.7363, "epoch": 0.15, "inbatch_neg_score": 0.1471, "inbatch_pos_score": 0.707, "learning_rate": 4.35e-05, "loss": 3.7606, "norm_diff": 0.1073, "norm_loss": 0.0, "num_token_doc": 66.8738, "num_token_overlap": 11.722, "num_token_query": 31.4048, "num_token_union": 65.165, "num_word_context": 202.4023, "num_word_doc": 49.8462, "num_word_query": 23.3143, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21796.444, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1469, "query_norm": 1.2688, "queue_k_norm": 1.3726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4048, "sent_len_1": 66.8738, "sent_len_max_0": 127.9675, "sent_len_max_1": 212.24, "stdk": 0.0479, "stdq": 0.0427, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34700 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.7748, "doc_norm": 1.3711, "encoder_q-embeddings": 11511.0723, "encoder_q-layer.0": 7606.2378, "encoder_q-layer.1": 7895.937, "encoder_q-layer.10": 16159.1777, "encoder_q-layer.11": 34912.043, "encoder_q-layer.2": 9013.6348, "encoder_q-layer.3": 9140.3438, "encoder_q-layer.4": 9677.2236, "encoder_q-layer.5": 10184.5762, "encoder_q-layer.6": 11341.708, "encoder_q-layer.7": 12155.5098, "encoder_q-layer.8": 15341.2598, "encoder_q-layer.9": 13808.9336, "epoch": 0.15, "inbatch_neg_score": 0.1441, "inbatch_pos_score": 0.7002, "learning_rate": 4.347368421052632e-05, "loss": 3.7748, "norm_diff": 0.1107, "norm_loss": 0.0, "num_token_doc": 66.8613, "num_token_overlap": 11.681, "num_token_query": 31.4129, "num_token_union": 65.1863, "num_word_context": 202.6167, "num_word_doc": 49.8738, "num_word_query": 23.3303, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21441.752, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1443, "query_norm": 1.2604, "queue_k_norm": 1.3727, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4129, "sent_len_1": 66.8613, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.255, "stdk": 0.0477, "stdq": 0.0426, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 34800 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.7568, "doc_norm": 1.3686, "encoder_q-embeddings": 11458.4893, "encoder_q-layer.0": 7540.7666, "encoder_q-layer.1": 7968.4116, "encoder_q-layer.10": 13803.7061, "encoder_q-layer.11": 32504.0117, "encoder_q-layer.2": 8869.4404, "encoder_q-layer.3": 9452.4219, "encoder_q-layer.4": 9952.4463, "encoder_q-layer.5": 9801.8174, "encoder_q-layer.6": 11196.9326, "encoder_q-layer.7": 12181.8613, "encoder_q-layer.8": 15558.5762, "encoder_q-layer.9": 13437.5225, "epoch": 0.15, "inbatch_neg_score": 0.1422, "inbatch_pos_score": 0.7148, "learning_rate": 4.344736842105263e-05, "loss": 3.7568, "norm_diff": 0.0992, "norm_loss": 0.0, "num_token_doc": 66.777, "num_token_overlap": 11.708, "num_token_query": 31.439, "num_token_union": 65.1068, "num_word_context": 202.2787, "num_word_doc": 49.8057, "num_word_query": 23.3536, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21254.8664, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1428, "query_norm": 1.2694, "queue_k_norm": 1.3735, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.439, "sent_len_1": 66.777, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.2763, "stdk": 0.0476, "stdq": 0.043, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 34900 }, { "accuracy": 44.0918, "active_queue_size": 16384.0, "cl_loss": 3.7677, "doc_norm": 1.3692, "encoder_q-embeddings": 11361.5391, "encoder_q-layer.0": 7439.8477, "encoder_q-layer.1": 7550.8296, "encoder_q-layer.10": 16951.2539, "encoder_q-layer.11": 37964.0703, "encoder_q-layer.2": 8376.4033, "encoder_q-layer.3": 8748.5996, "encoder_q-layer.4": 9388.7969, "encoder_q-layer.5": 9399.625, "encoder_q-layer.6": 10556.957, "encoder_q-layer.7": 12132.6992, "encoder_q-layer.8": 15131.1865, "encoder_q-layer.9": 14691.5664, "epoch": 0.15, "inbatch_neg_score": 0.1439, "inbatch_pos_score": 0.6738, "learning_rate": 4.342105263157895e-05, "loss": 3.7677, "norm_diff": 0.1431, "norm_loss": 0.0, "num_token_doc": 66.7942, "num_token_overlap": 11.6627, "num_token_query": 31.3717, "num_token_union": 65.119, "num_word_context": 202.3863, "num_word_doc": 49.812, "num_word_query": 23.296, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22219.1639, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1434, "query_norm": 1.2261, "queue_k_norm": 1.3735, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3717, "sent_len_1": 66.7942, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.6687, "stdk": 0.0476, "stdq": 0.0414, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 35000 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.757, "doc_norm": 1.3735, "encoder_q-embeddings": 12636.1455, "encoder_q-layer.0": 7730.8584, "encoder_q-layer.1": 8242.1201, "encoder_q-layer.10": 13632.29, "encoder_q-layer.11": 31651.873, "encoder_q-layer.2": 9363.0752, "encoder_q-layer.3": 9557.5293, "encoder_q-layer.4": 10137.2686, "encoder_q-layer.5": 10333.2441, "encoder_q-layer.6": 11681.584, "encoder_q-layer.7": 12431.0234, "encoder_q-layer.8": 14319.8066, "encoder_q-layer.9": 13142.2861, "epoch": 0.15, "inbatch_neg_score": 0.1365, "inbatch_pos_score": 0.7144, "learning_rate": 4.339473684210527e-05, "loss": 3.757, "norm_diff": 0.1066, "norm_loss": 0.0, "num_token_doc": 66.9164, "num_token_overlap": 11.7154, "num_token_query": 31.4526, "num_token_union": 65.1893, "num_word_context": 202.5676, "num_word_doc": 49.9661, "num_word_query": 23.36, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21062.6408, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1376, "query_norm": 1.2669, "queue_k_norm": 1.3745, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4526, "sent_len_1": 66.9164, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.1075, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35100 }, { "accuracy": 45.5078, "active_queue_size": 16384.0, "cl_loss": 3.7587, "doc_norm": 1.3766, "encoder_q-embeddings": 6339.395, "encoder_q-layer.0": 4006.7705, "encoder_q-layer.1": 4128.832, "encoder_q-layer.10": 7239.1147, "encoder_q-layer.11": 16687.0215, "encoder_q-layer.2": 4583.438, "encoder_q-layer.3": 4640.9023, "encoder_q-layer.4": 4915.1147, "encoder_q-layer.5": 5081.8369, "encoder_q-layer.6": 5682.0117, "encoder_q-layer.7": 6564.4683, "encoder_q-layer.8": 7619.3779, "encoder_q-layer.9": 6848.4175, "epoch": 0.15, "inbatch_neg_score": 0.1379, "inbatch_pos_score": 0.6895, "learning_rate": 4.336842105263158e-05, "loss": 3.7587, "norm_diff": 0.1255, "norm_loss": 0.0, "num_token_doc": 66.9581, "num_token_overlap": 11.6893, "num_token_query": 31.4156, "num_token_union": 65.2575, "num_word_context": 202.5886, "num_word_doc": 49.9449, "num_word_query": 23.3271, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11015.5928, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1367, "query_norm": 1.2511, "queue_k_norm": 1.3764, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4156, "sent_len_1": 66.9581, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.9475, "stdk": 0.0479, "stdq": 0.0426, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 35200 }, { "accuracy": 46.5332, "active_queue_size": 16384.0, "cl_loss": 3.7715, "doc_norm": 1.3733, "encoder_q-embeddings": 3115.5955, "encoder_q-layer.0": 2023.9497, "encoder_q-layer.1": 2166.8569, "encoder_q-layer.10": 3646.3269, "encoder_q-layer.11": 8205.3516, "encoder_q-layer.2": 2414.2393, "encoder_q-layer.3": 2389.9092, "encoder_q-layer.4": 2501.9971, "encoder_q-layer.5": 2577.5554, "encoder_q-layer.6": 2832.4368, "encoder_q-layer.7": 3038.5422, "encoder_q-layer.8": 3834.7097, "encoder_q-layer.9": 3507.5061, "epoch": 0.15, "inbatch_neg_score": 0.1393, "inbatch_pos_score": 0.7109, "learning_rate": 4.33421052631579e-05, "loss": 3.7715, "norm_diff": 0.1033, "norm_loss": 0.0, "num_token_doc": 66.7974, "num_token_overlap": 11.6846, "num_token_query": 31.3324, "num_token_union": 65.0688, "num_word_context": 202.2487, "num_word_doc": 49.8197, "num_word_query": 23.2643, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5393.6955, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1395, "query_norm": 1.27, "queue_k_norm": 1.3747, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3324, "sent_len_1": 66.7974, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.2463, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35300 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.7722, "doc_norm": 1.3729, "encoder_q-embeddings": 3304.9062, "encoder_q-layer.0": 2085.1228, "encoder_q-layer.1": 2166.752, "encoder_q-layer.10": 3329.562, "encoder_q-layer.11": 8031.0078, "encoder_q-layer.2": 2442.4753, "encoder_q-layer.3": 2499.3687, "encoder_q-layer.4": 2597.6824, "encoder_q-layer.5": 2769.1597, "encoder_q-layer.6": 3259.1938, "encoder_q-layer.7": 3684.1252, "encoder_q-layer.8": 3878.6963, "encoder_q-layer.9": 3101.47, "epoch": 0.15, "inbatch_neg_score": 0.1324, "inbatch_pos_score": 0.7305, "learning_rate": 4.3315789473684215e-05, "loss": 3.7722, "norm_diff": 0.0865, "norm_loss": 0.0, "num_token_doc": 66.7237, "num_token_overlap": 11.692, "num_token_query": 31.4094, "num_token_union": 65.1035, "num_word_context": 202.1107, "num_word_doc": 49.8267, "num_word_query": 23.3358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5382.2433, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1332, "query_norm": 1.2864, "queue_k_norm": 1.3751, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4094, "sent_len_1": 66.7237, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.4638, "stdk": 0.0478, "stdq": 0.0439, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35400 }, { "accuracy": 43.8477, "active_queue_size": 16384.0, "cl_loss": 3.7555, "doc_norm": 1.3719, "encoder_q-embeddings": 6713.0449, "encoder_q-layer.0": 4165.9463, "encoder_q-layer.1": 4467.4956, "encoder_q-layer.10": 3629.1523, "encoder_q-layer.11": 8804.9209, "encoder_q-layer.2": 4716.7158, "encoder_q-layer.3": 4451.4912, "encoder_q-layer.4": 4293.5405, "encoder_q-layer.5": 4080.168, "encoder_q-layer.6": 4587.0225, "encoder_q-layer.7": 5013.9629, "encoder_q-layer.8": 4526.6543, "encoder_q-layer.9": 3431.0618, "epoch": 0.15, "inbatch_neg_score": 0.1313, "inbatch_pos_score": 0.6685, "learning_rate": 4.328947368421053e-05, "loss": 3.7555, "norm_diff": 0.1269, "norm_loss": 0.0, "num_token_doc": 66.79, "num_token_overlap": 11.6846, "num_token_query": 31.4012, "num_token_union": 65.1405, "num_word_context": 202.2315, "num_word_doc": 49.8399, "num_word_query": 23.3112, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7802.7523, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1316, "query_norm": 1.245, "queue_k_norm": 1.3729, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4012, "sent_len_1": 66.79, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.1125, "stdk": 0.0477, "stdq": 0.0424, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 35500 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.7529, "doc_norm": 1.3708, "encoder_q-embeddings": 3015.4763, "encoder_q-layer.0": 1979.6841, "encoder_q-layer.1": 2062.438, "encoder_q-layer.10": 3700.0791, "encoder_q-layer.11": 8175.6162, "encoder_q-layer.2": 2252.498, "encoder_q-layer.3": 2366.6333, "encoder_q-layer.4": 2548.6077, "encoder_q-layer.5": 2505.5308, "encoder_q-layer.6": 2828.3457, "encoder_q-layer.7": 3187.7371, "encoder_q-layer.8": 3756.9265, "encoder_q-layer.9": 3429.207, "epoch": 0.15, "inbatch_neg_score": 0.1326, "inbatch_pos_score": 0.7061, "learning_rate": 4.3263157894736845e-05, "loss": 3.7529, "norm_diff": 0.1056, "norm_loss": 0.0, "num_token_doc": 66.9062, "num_token_overlap": 11.6652, "num_token_query": 31.2772, "num_token_union": 65.18, "num_word_context": 202.3495, "num_word_doc": 49.9586, "num_word_query": 23.2167, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5450.6699, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1331, "query_norm": 1.2652, "queue_k_norm": 1.3752, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2772, "sent_len_1": 66.9062, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.725, "stdk": 0.0477, "stdq": 0.0431, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35600 }, { "accuracy": 45.7031, "active_queue_size": 16384.0, "cl_loss": 3.7708, "doc_norm": 1.3744, "encoder_q-embeddings": 2954.46, "encoder_q-layer.0": 1871.4661, "encoder_q-layer.1": 2002.4912, "encoder_q-layer.10": 3720.8218, "encoder_q-layer.11": 8241.2686, "encoder_q-layer.2": 2231.9226, "encoder_q-layer.3": 2304.6804, "encoder_q-layer.4": 2420.1394, "encoder_q-layer.5": 2551.9219, "encoder_q-layer.6": 2970.1431, "encoder_q-layer.7": 3258.7183, "encoder_q-layer.8": 3626.7859, "encoder_q-layer.9": 3363.5515, "epoch": 0.15, "inbatch_neg_score": 0.1268, "inbatch_pos_score": 0.6836, "learning_rate": 4.323684210526316e-05, "loss": 3.7708, "norm_diff": 0.1159, "norm_loss": 0.0, "num_token_doc": 66.6056, "num_token_overlap": 11.6472, "num_token_query": 31.392, "num_token_union": 65.0206, "num_word_context": 201.9705, "num_word_doc": 49.7003, "num_word_query": 23.3192, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5342.8514, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1274, "query_norm": 1.2586, "queue_k_norm": 1.3745, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.392, "sent_len_1": 66.6056, "sent_len_max_0": 127.98, "sent_len_max_1": 208.3875, "stdk": 0.0479, "stdq": 0.043, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35700 }, { "accuracy": 45.2148, "active_queue_size": 16384.0, "cl_loss": 3.74, "doc_norm": 1.3707, "encoder_q-embeddings": 3124.8333, "encoder_q-layer.0": 2049.4001, "encoder_q-layer.1": 2202.8362, "encoder_q-layer.10": 3888.8755, "encoder_q-layer.11": 9006.5938, "encoder_q-layer.2": 2470.6729, "encoder_q-layer.3": 2555.8904, "encoder_q-layer.4": 2706.4863, "encoder_q-layer.5": 2845.7812, "encoder_q-layer.6": 3129.4075, "encoder_q-layer.7": 3338.3501, "encoder_q-layer.8": 4002.6145, "encoder_q-layer.9": 3503.1328, "epoch": 0.16, "inbatch_neg_score": 0.1291, "inbatch_pos_score": 0.6919, "learning_rate": 4.3210526315789475e-05, "loss": 3.74, "norm_diff": 0.1064, "norm_loss": 0.0, "num_token_doc": 66.8565, "num_token_overlap": 11.7072, "num_token_query": 31.4328, "num_token_union": 65.2113, "num_word_context": 202.3625, "num_word_doc": 49.9179, "num_word_query": 23.3463, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5787.6782, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1292, "query_norm": 1.2643, "queue_k_norm": 1.3745, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4328, "sent_len_1": 66.8565, "sent_len_max_0": 127.9788, "sent_len_max_1": 206.8638, "stdk": 0.0477, "stdq": 0.0432, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 35800 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.7552, "doc_norm": 1.3725, "encoder_q-embeddings": 6067.8882, "encoder_q-layer.0": 4023.2913, "encoder_q-layer.1": 4305.4526, "encoder_q-layer.10": 3520.27, "encoder_q-layer.11": 8619.9414, "encoder_q-layer.2": 4671.8047, "encoder_q-layer.3": 5182.0024, "encoder_q-layer.4": 4844.0425, "encoder_q-layer.5": 4667.6182, "encoder_q-layer.6": 4683.3027, "encoder_q-layer.7": 4527.1182, "encoder_q-layer.8": 4886.6323, "encoder_q-layer.9": 3669.6086, "epoch": 0.16, "inbatch_neg_score": 0.1294, "inbatch_pos_score": 0.6968, "learning_rate": 4.3184210526315793e-05, "loss": 3.7552, "norm_diff": 0.1075, "norm_loss": 0.0, "num_token_doc": 66.7411, "num_token_overlap": 11.6748, "num_token_query": 31.3853, "num_token_union": 65.1046, "num_word_context": 202.3277, "num_word_doc": 49.7983, "num_word_query": 23.3118, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7692.7419, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1298, "query_norm": 1.265, "queue_k_norm": 1.3759, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3853, "sent_len_1": 66.7411, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.4212, "stdk": 0.0478, "stdq": 0.0431, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 35900 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.769, "doc_norm": 1.3719, "encoder_q-embeddings": 3467.3223, "encoder_q-layer.0": 2302.7188, "encoder_q-layer.1": 2542.0667, "encoder_q-layer.10": 3568.3066, "encoder_q-layer.11": 8333.4756, "encoder_q-layer.2": 2797.3813, "encoder_q-layer.3": 2695.9705, "encoder_q-layer.4": 2779.3638, "encoder_q-layer.5": 2885.5449, "encoder_q-layer.6": 2994.8555, "encoder_q-layer.7": 3155.1865, "encoder_q-layer.8": 3923.2212, "encoder_q-layer.9": 3368.0415, "epoch": 0.16, "inbatch_neg_score": 0.1224, "inbatch_pos_score": 0.6899, "learning_rate": 4.3157894736842105e-05, "loss": 3.769, "norm_diff": 0.1143, "norm_loss": 0.0, "num_token_doc": 66.8266, "num_token_overlap": 11.679, "num_token_query": 31.3584, "num_token_union": 65.1274, "num_word_context": 202.4137, "num_word_doc": 49.8642, "num_word_query": 23.3002, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5539.4018, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1236, "query_norm": 1.2576, "queue_k_norm": 1.3737, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3584, "sent_len_1": 66.8266, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.3512, "stdk": 0.0478, "stdq": 0.043, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 36000 }, { "accuracy": 45.0684, "active_queue_size": 16384.0, "cl_loss": 3.7601, "doc_norm": 1.373, "encoder_q-embeddings": 3193.2256, "encoder_q-layer.0": 1983.2365, "encoder_q-layer.1": 2127.3914, "encoder_q-layer.10": 3768.6577, "encoder_q-layer.11": 8222.0586, "encoder_q-layer.2": 2354.166, "encoder_q-layer.3": 2435.2354, "encoder_q-layer.4": 2656.2048, "encoder_q-layer.5": 2556.8333, "encoder_q-layer.6": 2995.73, "encoder_q-layer.7": 3107.3376, "encoder_q-layer.8": 3727.9929, "encoder_q-layer.9": 3380.3821, "epoch": 0.16, "inbatch_neg_score": 0.1224, "inbatch_pos_score": 0.6875, "learning_rate": 4.3131578947368423e-05, "loss": 3.7601, "norm_diff": 0.0991, "norm_loss": 0.0, "num_token_doc": 66.7262, "num_token_overlap": 11.6665, "num_token_query": 31.3428, "num_token_union": 65.072, "num_word_context": 202.5263, "num_word_doc": 49.7832, "num_word_query": 23.2765, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5429.7588, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1224, "query_norm": 1.2738, "queue_k_norm": 1.3761, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3428, "sent_len_1": 66.7262, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.5838, "stdk": 0.0478, "stdq": 0.0435, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 36100 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.7455, "doc_norm": 1.3689, "encoder_q-embeddings": 3699.571, "encoder_q-layer.0": 2366.0139, "encoder_q-layer.1": 2649.9822, "encoder_q-layer.10": 3533.4084, "encoder_q-layer.11": 8412.4551, "encoder_q-layer.2": 2817.2739, "encoder_q-layer.3": 2951.2812, "encoder_q-layer.4": 3129.6321, "encoder_q-layer.5": 3057.55, "encoder_q-layer.6": 3361.436, "encoder_q-layer.7": 3490.7966, "encoder_q-layer.8": 4075.2651, "encoder_q-layer.9": 3600.9834, "epoch": 0.16, "inbatch_neg_score": 0.1239, "inbatch_pos_score": 0.708, "learning_rate": 4.3105263157894735e-05, "loss": 3.7455, "norm_diff": 0.0942, "norm_loss": 0.0, "num_token_doc": 66.7904, "num_token_overlap": 11.6734, "num_token_query": 31.3943, "num_token_union": 65.1431, "num_word_context": 202.4268, "num_word_doc": 49.8479, "num_word_query": 23.3104, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5958.2986, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1232, "query_norm": 1.2747, "queue_k_norm": 1.3741, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3943, "sent_len_1": 66.7904, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.9038, "stdk": 0.0477, "stdq": 0.0433, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 36200 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.7423, "doc_norm": 1.3782, "encoder_q-embeddings": 3432.1328, "encoder_q-layer.0": 2234.4016, "encoder_q-layer.1": 2423.0496, "encoder_q-layer.10": 3684.8818, "encoder_q-layer.11": 8208.9766, "encoder_q-layer.2": 2737.5933, "encoder_q-layer.3": 2818.7983, "encoder_q-layer.4": 2949.147, "encoder_q-layer.5": 2834.9456, "encoder_q-layer.6": 3225.6321, "encoder_q-layer.7": 3441.8711, "encoder_q-layer.8": 3604.2178, "encoder_q-layer.9": 3513.3418, "epoch": 0.16, "inbatch_neg_score": 0.1237, "inbatch_pos_score": 0.6855, "learning_rate": 4.307894736842105e-05, "loss": 3.7423, "norm_diff": 0.1032, "norm_loss": 0.0, "num_token_doc": 66.7894, "num_token_overlap": 11.6853, "num_token_query": 31.3703, "num_token_union": 65.1039, "num_word_context": 202.2312, "num_word_doc": 49.8695, "num_word_query": 23.299, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5597.5028, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1238, "query_norm": 1.275, "queue_k_norm": 1.3742, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3703, "sent_len_1": 66.7894, "sent_len_max_0": 127.9638, "sent_len_max_1": 207.405, "stdk": 0.048, "stdq": 0.0431, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 36300 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.749, "doc_norm": 1.3763, "encoder_q-embeddings": 3244.7739, "encoder_q-layer.0": 2150.3901, "encoder_q-layer.1": 2465.7625, "encoder_q-layer.10": 3463.5913, "encoder_q-layer.11": 8242.5801, "encoder_q-layer.2": 2735.0952, "encoder_q-layer.3": 2955.2705, "encoder_q-layer.4": 3354.4812, "encoder_q-layer.5": 3671.8701, "encoder_q-layer.6": 3678.479, "encoder_q-layer.7": 3698.103, "encoder_q-layer.8": 3786.5312, "encoder_q-layer.9": 3316.9358, "epoch": 0.16, "inbatch_neg_score": 0.1262, "inbatch_pos_score": 0.7129, "learning_rate": 4.305263157894737e-05, "loss": 3.749, "norm_diff": 0.0896, "norm_loss": 0.0, "num_token_doc": 66.8212, "num_token_overlap": 11.7024, "num_token_query": 31.3878, "num_token_union": 65.1161, "num_word_context": 202.2413, "num_word_doc": 49.8673, "num_word_query": 23.3091, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5737.5915, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.126, "query_norm": 1.2867, "queue_k_norm": 1.3714, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3878, "sent_len_1": 66.8212, "sent_len_max_0": 127.975, "sent_len_max_1": 209.07, "stdk": 0.048, "stdq": 0.0433, "stdqueue_k": 0.0478, "stdqueue_q": 0.0, "step": 36400 }, { "accuracy": 46.7773, "active_queue_size": 16384.0, "cl_loss": 3.7571, "doc_norm": 1.3729, "encoder_q-embeddings": 2979.4331, "encoder_q-layer.0": 1894.5526, "encoder_q-layer.1": 1978.4817, "encoder_q-layer.10": 3318.0771, "encoder_q-layer.11": 7851.0952, "encoder_q-layer.2": 2189.9072, "encoder_q-layer.3": 2245.2861, "encoder_q-layer.4": 2328.5122, "encoder_q-layer.5": 2351.5181, "encoder_q-layer.6": 2609.3467, "encoder_q-layer.7": 2824.0142, "encoder_q-layer.8": 3502.5789, "encoder_q-layer.9": 3155.1421, "epoch": 0.16, "inbatch_neg_score": 0.1268, "inbatch_pos_score": 0.688, "learning_rate": 4.302631578947369e-05, "loss": 3.7571, "norm_diff": 0.091, "norm_loss": 0.0, "num_token_doc": 66.83, "num_token_overlap": 11.6752, "num_token_query": 31.3454, "num_token_union": 65.1222, "num_word_context": 202.4909, "num_word_doc": 49.8612, "num_word_query": 23.2786, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5184.3991, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1276, "query_norm": 1.2818, "queue_k_norm": 1.3735, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3454, "sent_len_1": 66.83, "sent_len_max_0": 127.9688, "sent_len_max_1": 210.4975, "stdk": 0.0479, "stdq": 0.0427, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 36500 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.7416, "doc_norm": 1.374, "encoder_q-embeddings": 2854.9111, "encoder_q-layer.0": 1818.4928, "encoder_q-layer.1": 1943.3827, "encoder_q-layer.10": 3192.9141, "encoder_q-layer.11": 7472.751, "encoder_q-layer.2": 2188.5886, "encoder_q-layer.3": 2206.7749, "encoder_q-layer.4": 2361.6667, "encoder_q-layer.5": 2406.1829, "encoder_q-layer.6": 2710.8684, "encoder_q-layer.7": 2958.0557, "encoder_q-layer.8": 3481.7302, "encoder_q-layer.9": 3154.7241, "epoch": 0.16, "inbatch_neg_score": 0.1304, "inbatch_pos_score": 0.6895, "learning_rate": 4.3e-05, "loss": 3.7416, "norm_diff": 0.0948, "norm_loss": 0.0, "num_token_doc": 66.6737, "num_token_overlap": 11.7173, "num_token_query": 31.4779, "num_token_union": 65.0897, "num_word_context": 202.416, "num_word_doc": 49.7561, "num_word_query": 23.389, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4939.5255, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1296, "query_norm": 1.2793, "queue_k_norm": 1.3724, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4779, "sent_len_1": 66.6737, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.5962, "stdk": 0.0479, "stdq": 0.0422, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 36600 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.7604, "doc_norm": 1.3735, "encoder_q-embeddings": 3292.002, "encoder_q-layer.0": 2240.1001, "encoder_q-layer.1": 2414.5779, "encoder_q-layer.10": 3487.3359, "encoder_q-layer.11": 8047.1514, "encoder_q-layer.2": 2865.667, "encoder_q-layer.3": 2934.7998, "encoder_q-layer.4": 3033.8242, "encoder_q-layer.5": 3002.6094, "encoder_q-layer.6": 3207.3782, "encoder_q-layer.7": 3294.1802, "encoder_q-layer.8": 3786.0107, "encoder_q-layer.9": 3400.605, "epoch": 0.16, "inbatch_neg_score": 0.1347, "inbatch_pos_score": 0.7051, "learning_rate": 4.297368421052631e-05, "loss": 3.7604, "norm_diff": 0.0716, "norm_loss": 0.0, "num_token_doc": 66.6095, "num_token_overlap": 11.6568, "num_token_query": 31.4171, "num_token_union": 65.0803, "num_word_context": 202.2656, "num_word_doc": 49.7393, "num_word_query": 23.3288, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5651.54, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1343, "query_norm": 1.3018, "queue_k_norm": 1.3735, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4171, "sent_len_1": 66.6095, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.2425, "stdk": 0.0479, "stdq": 0.043, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 36700 }, { "accuracy": 45.0195, "active_queue_size": 16384.0, "cl_loss": 3.7417, "doc_norm": 1.3691, "encoder_q-embeddings": 5032.1089, "encoder_q-layer.0": 3566.9587, "encoder_q-layer.1": 4023.1851, "encoder_q-layer.10": 3597.9595, "encoder_q-layer.11": 8441.6572, "encoder_q-layer.2": 4140.2827, "encoder_q-layer.3": 3493.4792, "encoder_q-layer.4": 3381.1001, "encoder_q-layer.5": 3057.3354, "encoder_q-layer.6": 3001.7104, "encoder_q-layer.7": 3353.2861, "encoder_q-layer.8": 3891.7964, "encoder_q-layer.9": 3392.593, "epoch": 0.16, "inbatch_neg_score": 0.1387, "inbatch_pos_score": 0.6841, "learning_rate": 4.294736842105264e-05, "loss": 3.7417, "norm_diff": 0.0804, "norm_loss": 0.0, "num_token_doc": 66.701, "num_token_overlap": 11.7123, "num_token_query": 31.4589, "num_token_union": 65.0955, "num_word_context": 202.0015, "num_word_doc": 49.7709, "num_word_query": 23.3747, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6357.2019, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1375, "query_norm": 1.2887, "queue_k_norm": 1.3737, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4589, "sent_len_1": 66.701, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.405, "stdk": 0.0477, "stdq": 0.0426, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 36800 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.7442, "doc_norm": 1.3788, "encoder_q-embeddings": 3357.2434, "encoder_q-layer.0": 2305.8115, "encoder_q-layer.1": 2452.1465, "encoder_q-layer.10": 3298.3762, "encoder_q-layer.11": 7676.9116, "encoder_q-layer.2": 2837.9453, "encoder_q-layer.3": 2977.9148, "encoder_q-layer.4": 3108.2908, "encoder_q-layer.5": 3069.1873, "encoder_q-layer.6": 3242.2239, "encoder_q-layer.7": 3483.343, "encoder_q-layer.8": 3782.3613, "encoder_q-layer.9": 3186.6785, "epoch": 0.16, "inbatch_neg_score": 0.1425, "inbatch_pos_score": 0.7188, "learning_rate": 4.292105263157895e-05, "loss": 3.7442, "norm_diff": 0.0872, "norm_loss": 0.0, "num_token_doc": 66.7528, "num_token_overlap": 11.6876, "num_token_query": 31.3706, "num_token_union": 65.0856, "num_word_context": 202.2489, "num_word_doc": 49.7866, "num_word_query": 23.3048, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5510.4202, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1433, "query_norm": 1.2917, "queue_k_norm": 1.3757, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3706, "sent_len_1": 66.7528, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.485, "stdk": 0.0481, "stdq": 0.0431, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 36900 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.7447, "doc_norm": 1.3768, "encoder_q-embeddings": 2841.7664, "encoder_q-layer.0": 1922.3491, "encoder_q-layer.1": 1972.7506, "encoder_q-layer.10": 3611.3213, "encoder_q-layer.11": 8213.1318, "encoder_q-layer.2": 2220.6746, "encoder_q-layer.3": 2286.2432, "encoder_q-layer.4": 2388.707, "encoder_q-layer.5": 2390.4197, "encoder_q-layer.6": 2649.7261, "encoder_q-layer.7": 2947.0674, "encoder_q-layer.8": 3417.916, "encoder_q-layer.9": 3273.3665, "epoch": 0.16, "inbatch_neg_score": 0.1444, "inbatch_pos_score": 0.7036, "learning_rate": 4.289473684210527e-05, "loss": 3.7447, "norm_diff": 0.1061, "norm_loss": 0.0, "num_token_doc": 66.7388, "num_token_overlap": 11.6773, "num_token_query": 31.3575, "num_token_union": 65.0957, "num_word_context": 202.3427, "num_word_doc": 49.806, "num_word_query": 23.2897, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5198.5993, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1448, "query_norm": 1.2708, "queue_k_norm": 1.3746, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3575, "sent_len_1": 66.7388, "sent_len_max_0": 127.985, "sent_len_max_1": 210.4625, "stdk": 0.048, "stdq": 0.0427, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37000 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.7505, "doc_norm": 1.3745, "encoder_q-embeddings": 3031.1577, "encoder_q-layer.0": 2039.5629, "encoder_q-layer.1": 2105.7092, "encoder_q-layer.10": 3528.6982, "encoder_q-layer.11": 8292.2881, "encoder_q-layer.2": 2334.4275, "encoder_q-layer.3": 2326.2932, "encoder_q-layer.4": 2456.147, "encoder_q-layer.5": 2453.9529, "encoder_q-layer.6": 2763.7012, "encoder_q-layer.7": 3047.4023, "encoder_q-layer.8": 3558.5962, "encoder_q-layer.9": 3359.1521, "epoch": 0.16, "inbatch_neg_score": 0.145, "inbatch_pos_score": 0.707, "learning_rate": 4.286842105263158e-05, "loss": 3.7505, "norm_diff": 0.1052, "norm_loss": 0.0, "num_token_doc": 66.7785, "num_token_overlap": 11.6326, "num_token_query": 31.2958, "num_token_union": 65.1163, "num_word_context": 202.2364, "num_word_doc": 49.8247, "num_word_query": 23.2341, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5220.4033, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1451, "query_norm": 1.2692, "queue_k_norm": 1.3736, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2958, "sent_len_1": 66.7785, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.0337, "stdk": 0.0479, "stdq": 0.0429, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 37100 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.7326, "doc_norm": 1.3722, "encoder_q-embeddings": 3209.0916, "encoder_q-layer.0": 2028.6456, "encoder_q-layer.1": 2153.6135, "encoder_q-layer.10": 3723.7778, "encoder_q-layer.11": 7940.6099, "encoder_q-layer.2": 2425.416, "encoder_q-layer.3": 2578.9434, "encoder_q-layer.4": 2708.8364, "encoder_q-layer.5": 2780.6282, "encoder_q-layer.6": 3051.0969, "encoder_q-layer.7": 3361.9155, "encoder_q-layer.8": 3864.2109, "encoder_q-layer.9": 3404.4897, "epoch": 0.16, "inbatch_neg_score": 0.1419, "inbatch_pos_score": 0.709, "learning_rate": 4.284210526315789e-05, "loss": 3.7326, "norm_diff": 0.106, "norm_loss": 0.0, "num_token_doc": 66.961, "num_token_overlap": 11.7499, "num_token_query": 31.497, "num_token_union": 65.2292, "num_word_context": 202.4356, "num_word_doc": 49.9673, "num_word_query": 23.398, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5429.4806, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.142, "query_norm": 1.2662, "queue_k_norm": 1.3772, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.497, "sent_len_1": 66.961, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.6813, "stdk": 0.0478, "stdq": 0.043, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37200 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.7447, "doc_norm": 1.3727, "encoder_q-embeddings": 6633.917, "encoder_q-layer.0": 4348.9043, "encoder_q-layer.1": 4718.8169, "encoder_q-layer.10": 7010.7041, "encoder_q-layer.11": 15140.2012, "encoder_q-layer.2": 5565.3276, "encoder_q-layer.3": 5768.29, "encoder_q-layer.4": 5770.3955, "encoder_q-layer.5": 5831.1743, "encoder_q-layer.6": 6113.4189, "encoder_q-layer.7": 6567.2734, "encoder_q-layer.8": 7042.5142, "encoder_q-layer.9": 6388.5781, "epoch": 0.16, "inbatch_neg_score": 0.1415, "inbatch_pos_score": 0.7163, "learning_rate": 4.281578947368422e-05, "loss": 3.7447, "norm_diff": 0.1011, "norm_loss": 0.0, "num_token_doc": 66.6519, "num_token_overlap": 11.6519, "num_token_query": 31.3308, "num_token_union": 65.0682, "num_word_context": 202.2583, "num_word_doc": 49.751, "num_word_query": 23.2812, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10782.0159, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1422, "query_norm": 1.2717, "queue_k_norm": 1.3751, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3308, "sent_len_1": 66.6519, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.3038, "stdk": 0.0478, "stdq": 0.0432, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 37300 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.7422, "doc_norm": 1.379, "encoder_q-embeddings": 7737.9922, "encoder_q-layer.0": 5109.0107, "encoder_q-layer.1": 5207.481, "encoder_q-layer.10": 6990.7207, "encoder_q-layer.11": 16462.2344, "encoder_q-layer.2": 5916.2471, "encoder_q-layer.3": 6432.4829, "encoder_q-layer.4": 6673.3945, "encoder_q-layer.5": 6226.4556, "encoder_q-layer.6": 7176.7266, "encoder_q-layer.7": 6863.1069, "encoder_q-layer.8": 7217.9697, "encoder_q-layer.9": 6363.1479, "epoch": 0.16, "inbatch_neg_score": 0.1423, "inbatch_pos_score": 0.7236, "learning_rate": 4.278947368421053e-05, "loss": 3.7422, "norm_diff": 0.1151, "norm_loss": 0.0, "num_token_doc": 66.7937, "num_token_overlap": 11.6392, "num_token_query": 31.3279, "num_token_union": 65.1111, "num_word_context": 202.1727, "num_word_doc": 49.823, "num_word_query": 23.2644, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11774.5732, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1427, "query_norm": 1.2639, "queue_k_norm": 1.3767, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3279, "sent_len_1": 66.7937, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.3338, "stdk": 0.048, "stdq": 0.0429, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37400 }, { "accuracy": 47.2656, "active_queue_size": 16384.0, "cl_loss": 3.7508, "doc_norm": 1.3767, "encoder_q-embeddings": 5615.1587, "encoder_q-layer.0": 3770.4951, "encoder_q-layer.1": 4164.1338, "encoder_q-layer.10": 6413.5498, "encoder_q-layer.11": 15227.8447, "encoder_q-layer.2": 4485.4541, "encoder_q-layer.3": 4696.1436, "encoder_q-layer.4": 4837.7856, "encoder_q-layer.5": 4809.4863, "encoder_q-layer.6": 5469.3286, "encoder_q-layer.7": 5864.6851, "encoder_q-layer.8": 6983.3115, "encoder_q-layer.9": 6261.3672, "epoch": 0.16, "inbatch_neg_score": 0.143, "inbatch_pos_score": 0.7129, "learning_rate": 4.2763157894736847e-05, "loss": 3.7508, "norm_diff": 0.107, "norm_loss": 0.0, "num_token_doc": 66.8079, "num_token_overlap": 11.6517, "num_token_query": 31.273, "num_token_union": 65.0976, "num_word_context": 202.1717, "num_word_doc": 49.8602, "num_word_query": 23.2175, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10067.5445, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1432, "query_norm": 1.2696, "queue_k_norm": 1.376, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.273, "sent_len_1": 66.8079, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.79, "stdk": 0.0479, "stdq": 0.043, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37500 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.7309, "doc_norm": 1.3739, "encoder_q-embeddings": 5701.5303, "encoder_q-layer.0": 3717.8931, "encoder_q-layer.1": 3879.7332, "encoder_q-layer.10": 6497.8975, "encoder_q-layer.11": 15704.7881, "encoder_q-layer.2": 4285.647, "encoder_q-layer.3": 4360.8018, "encoder_q-layer.4": 4642.0361, "encoder_q-layer.5": 4801.3828, "encoder_q-layer.6": 5395.9692, "encoder_q-layer.7": 5677.4951, "encoder_q-layer.8": 6681.7778, "encoder_q-layer.9": 6157.2891, "epoch": 0.16, "inbatch_neg_score": 0.132, "inbatch_pos_score": 0.6963, "learning_rate": 4.273684210526316e-05, "loss": 3.7309, "norm_diff": 0.1259, "norm_loss": 0.0, "num_token_doc": 66.9181, "num_token_overlap": 11.6627, "num_token_query": 31.3009, "num_token_union": 65.1498, "num_word_context": 202.2395, "num_word_doc": 49.9401, "num_word_query": 23.2508, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9982.4938, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1317, "query_norm": 1.2479, "queue_k_norm": 1.3789, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3009, "sent_len_1": 66.9181, "sent_len_max_0": 128.0, "sent_len_max_1": 209.4375, "stdk": 0.0478, "stdq": 0.0425, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 37600 }, { "accuracy": 46.6797, "active_queue_size": 16384.0, "cl_loss": 3.747, "doc_norm": 1.3782, "encoder_q-embeddings": 5896.5508, "encoder_q-layer.0": 3829.5811, "encoder_q-layer.1": 4044.9807, "encoder_q-layer.10": 7036.8779, "encoder_q-layer.11": 16313.7324, "encoder_q-layer.2": 4505.3262, "encoder_q-layer.3": 4578.8784, "encoder_q-layer.4": 4819.7085, "encoder_q-layer.5": 4867.957, "encoder_q-layer.6": 5344.5645, "encoder_q-layer.7": 5664.0859, "encoder_q-layer.8": 6981.3706, "encoder_q-layer.9": 6254.7612, "epoch": 0.16, "inbatch_neg_score": 0.1351, "inbatch_pos_score": 0.6802, "learning_rate": 4.271052631578947e-05, "loss": 3.747, "norm_diff": 0.1472, "norm_loss": 0.0, "num_token_doc": 66.999, "num_token_overlap": 11.6879, "num_token_query": 31.3552, "num_token_union": 65.2366, "num_word_context": 202.6072, "num_word_doc": 49.9842, "num_word_query": 23.283, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10194.5136, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1354, "query_norm": 1.2311, "queue_k_norm": 1.3779, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3552, "sent_len_1": 66.999, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.5112, "stdk": 0.048, "stdq": 0.0417, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37700 }, { "accuracy": 45.3613, "active_queue_size": 16384.0, "cl_loss": 3.7567, "doc_norm": 1.3742, "encoder_q-embeddings": 5787.4951, "encoder_q-layer.0": 3728.3589, "encoder_q-layer.1": 3827.0146, "encoder_q-layer.10": 6597.0537, "encoder_q-layer.11": 15796.8955, "encoder_q-layer.2": 4163.4741, "encoder_q-layer.3": 4390.6328, "encoder_q-layer.4": 4613.728, "encoder_q-layer.5": 4681.9097, "encoder_q-layer.6": 5244.2256, "encoder_q-layer.7": 5915.2266, "encoder_q-layer.8": 7259.8916, "encoder_q-layer.9": 6424.918, "epoch": 0.16, "inbatch_neg_score": 0.1323, "inbatch_pos_score": 0.6948, "learning_rate": 4.2684210526315795e-05, "loss": 3.7567, "norm_diff": 0.122, "norm_loss": 0.0, "num_token_doc": 66.7036, "num_token_overlap": 11.6274, "num_token_query": 31.2554, "num_token_union": 65.0515, "num_word_context": 202.2537, "num_word_doc": 49.7661, "num_word_query": 23.2047, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10246.6727, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1322, "query_norm": 1.2522, "queue_k_norm": 1.3744, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2554, "sent_len_1": 66.7036, "sent_len_max_0": 127.9638, "sent_len_max_1": 211.1513, "stdk": 0.0478, "stdq": 0.0427, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 37800 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.7338, "doc_norm": 1.3812, "encoder_q-embeddings": 5746.1206, "encoder_q-layer.0": 3838.3928, "encoder_q-layer.1": 3955.6497, "encoder_q-layer.10": 6540.9277, "encoder_q-layer.11": 15538.5088, "encoder_q-layer.2": 4291.9399, "encoder_q-layer.3": 4405.4912, "encoder_q-layer.4": 4477.4277, "encoder_q-layer.5": 4573.8467, "encoder_q-layer.6": 5286.5723, "encoder_q-layer.7": 5749.6113, "encoder_q-layer.8": 7130.7109, "encoder_q-layer.9": 6132.0317, "epoch": 0.16, "inbatch_neg_score": 0.1328, "inbatch_pos_score": 0.7148, "learning_rate": 4.2657894736842106e-05, "loss": 3.7338, "norm_diff": 0.1178, "norm_loss": 0.0, "num_token_doc": 66.874, "num_token_overlap": 11.6722, "num_token_query": 31.4187, "num_token_union": 65.194, "num_word_context": 202.4319, "num_word_doc": 49.91, "num_word_query": 23.3306, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10143.4695, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1334, "query_norm": 1.2634, "queue_k_norm": 1.3772, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4187, "sent_len_1": 66.874, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.3462, "stdk": 0.0481, "stdq": 0.0429, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 37900 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 3.733, "doc_norm": 1.3772, "encoder_q-embeddings": 5818.3657, "encoder_q-layer.0": 3964.6704, "encoder_q-layer.1": 4042.6633, "encoder_q-layer.10": 6530.23, "encoder_q-layer.11": 15490.1445, "encoder_q-layer.2": 4472.0786, "encoder_q-layer.3": 4597.1914, "encoder_q-layer.4": 4864.7681, "encoder_q-layer.5": 4929.9814, "encoder_q-layer.6": 5367.1772, "encoder_q-layer.7": 5812.5967, "encoder_q-layer.8": 7087.1792, "encoder_q-layer.9": 6199.4287, "epoch": 0.16, "inbatch_neg_score": 0.132, "inbatch_pos_score": 0.6973, "learning_rate": 4.2631578947368425e-05, "loss": 3.733, "norm_diff": 0.1337, "norm_loss": 0.0, "num_token_doc": 66.8788, "num_token_overlap": 11.7322, "num_token_query": 31.403, "num_token_union": 65.1552, "num_word_context": 202.2496, "num_word_doc": 49.9172, "num_word_query": 23.3196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10205.1886, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1324, "query_norm": 1.2435, "queue_k_norm": 1.3772, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.403, "sent_len_1": 66.8788, "sent_len_max_0": 127.995, "sent_len_max_1": 209.945, "stdk": 0.048, "stdq": 0.0422, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38000 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.7311, "doc_norm": 1.3792, "encoder_q-embeddings": 5649.813, "encoder_q-layer.0": 3765.4419, "encoder_q-layer.1": 3974.3369, "encoder_q-layer.10": 7400.9873, "encoder_q-layer.11": 16813.7227, "encoder_q-layer.2": 4453.6304, "encoder_q-layer.3": 4524.1011, "encoder_q-layer.4": 4850.814, "encoder_q-layer.5": 5043.6768, "encoder_q-layer.6": 5721.1797, "encoder_q-layer.7": 6504.6284, "encoder_q-layer.8": 7877.478, "encoder_q-layer.9": 6999.4995, "epoch": 0.17, "inbatch_neg_score": 0.1337, "inbatch_pos_score": 0.707, "learning_rate": 4.2605263157894736e-05, "loss": 3.7311, "norm_diff": 0.095, "norm_loss": 0.0, "num_token_doc": 66.7767, "num_token_overlap": 11.6654, "num_token_query": 31.3339, "num_token_union": 65.1034, "num_word_context": 202.163, "num_word_doc": 49.8111, "num_word_query": 23.2583, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10742.2879, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1343, "query_norm": 1.2841, "queue_k_norm": 1.3779, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3339, "sent_len_1": 66.7767, "sent_len_max_0": 127.995, "sent_len_max_1": 208.0588, "stdk": 0.048, "stdq": 0.0436, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38100 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.7345, "doc_norm": 1.3784, "encoder_q-embeddings": 6306.7275, "encoder_q-layer.0": 4017.3354, "encoder_q-layer.1": 4315.6406, "encoder_q-layer.10": 7179.0557, "encoder_q-layer.11": 16435.3359, "encoder_q-layer.2": 4855.4302, "encoder_q-layer.3": 5124.4258, "encoder_q-layer.4": 5455.75, "encoder_q-layer.5": 5463.8823, "encoder_q-layer.6": 5935.4775, "encoder_q-layer.7": 7138.1943, "encoder_q-layer.8": 8277.5771, "encoder_q-layer.9": 7013.6104, "epoch": 0.17, "inbatch_neg_score": 0.1325, "inbatch_pos_score": 0.7065, "learning_rate": 4.2578947368421055e-05, "loss": 3.7345, "norm_diff": 0.0997, "norm_loss": 0.0, "num_token_doc": 66.7674, "num_token_overlap": 11.6732, "num_token_query": 31.3726, "num_token_union": 65.0806, "num_word_context": 202.4329, "num_word_doc": 49.8316, "num_word_query": 23.2978, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11042.8477, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1324, "query_norm": 1.2786, "queue_k_norm": 1.3784, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3726, "sent_len_1": 66.7674, "sent_len_max_0": 127.9475, "sent_len_max_1": 208.5475, "stdk": 0.048, "stdq": 0.0434, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 38200 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.7366, "doc_norm": 1.3849, "encoder_q-embeddings": 5911.9434, "encoder_q-layer.0": 3892.8452, "encoder_q-layer.1": 4057.4907, "encoder_q-layer.10": 6840.6421, "encoder_q-layer.11": 16535.5117, "encoder_q-layer.2": 4683.353, "encoder_q-layer.3": 4895.3447, "encoder_q-layer.4": 5062.1201, "encoder_q-layer.5": 5178.353, "encoder_q-layer.6": 5829.8882, "encoder_q-layer.7": 6469.1851, "encoder_q-layer.8": 7594.4312, "encoder_q-layer.9": 6712.6362, "epoch": 0.17, "inbatch_neg_score": 0.1307, "inbatch_pos_score": 0.7129, "learning_rate": 4.255263157894737e-05, "loss": 3.7366, "norm_diff": 0.1077, "norm_loss": 0.0, "num_token_doc": 66.789, "num_token_overlap": 11.6562, "num_token_query": 31.3868, "num_token_union": 65.1476, "num_word_context": 202.404, "num_word_doc": 49.8384, "num_word_query": 23.3096, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10718.4154, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1305, "query_norm": 1.2772, "queue_k_norm": 1.3777, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3868, "sent_len_1": 66.789, "sent_len_max_0": 128.0, "sent_len_max_1": 208.785, "stdk": 0.0482, "stdq": 0.0433, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38300 }, { "accuracy": 47.3633, "active_queue_size": 16384.0, "cl_loss": 3.7222, "doc_norm": 1.3736, "encoder_q-embeddings": 8089.5581, "encoder_q-layer.0": 5476.4692, "encoder_q-layer.1": 6393.4907, "encoder_q-layer.10": 7021.2524, "encoder_q-layer.11": 14938.0479, "encoder_q-layer.2": 7240.2817, "encoder_q-layer.3": 7677.7769, "encoder_q-layer.4": 8079.8208, "encoder_q-layer.5": 8434.7412, "encoder_q-layer.6": 8400.8633, "encoder_q-layer.7": 8290.8018, "encoder_q-layer.8": 8331.6113, "encoder_q-layer.9": 7049.832, "epoch": 0.17, "inbatch_neg_score": 0.1363, "inbatch_pos_score": 0.7012, "learning_rate": 4.2526315789473685e-05, "loss": 3.7222, "norm_diff": 0.1016, "norm_loss": 0.0, "num_token_doc": 66.6338, "num_token_overlap": 11.6856, "num_token_query": 31.4075, "num_token_union": 65.0489, "num_word_context": 202.1884, "num_word_doc": 49.6858, "num_word_query": 23.3113, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12400.7626, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1364, "query_norm": 1.2721, "queue_k_norm": 1.3771, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4075, "sent_len_1": 66.6338, "sent_len_max_0": 127.97, "sent_len_max_1": 209.2575, "stdk": 0.0478, "stdq": 0.0427, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38400 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.7169, "doc_norm": 1.3855, "encoder_q-embeddings": 6011.4521, "encoder_q-layer.0": 3932.1404, "encoder_q-layer.1": 4113.0239, "encoder_q-layer.10": 7439.1846, "encoder_q-layer.11": 16396.9473, "encoder_q-layer.2": 4513.2485, "encoder_q-layer.3": 4642.5518, "encoder_q-layer.4": 4997.7529, "encoder_q-layer.5": 5209.2329, "encoder_q-layer.6": 5928.0127, "encoder_q-layer.7": 6938.2612, "encoder_q-layer.8": 7981.0962, "encoder_q-layer.9": 7109.9478, "epoch": 0.17, "inbatch_neg_score": 0.1367, "inbatch_pos_score": 0.7007, "learning_rate": 4.25e-05, "loss": 3.7169, "norm_diff": 0.1105, "norm_loss": 0.0, "num_token_doc": 66.8642, "num_token_overlap": 11.732, "num_token_query": 31.5144, "num_token_union": 65.203, "num_word_context": 202.5307, "num_word_doc": 49.909, "num_word_query": 23.4199, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10680.4915, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.137, "query_norm": 1.275, "queue_k_norm": 1.3772, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5144, "sent_len_1": 66.8642, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.5775, "stdk": 0.0483, "stdq": 0.0426, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38500 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.7309, "doc_norm": 1.3793, "encoder_q-embeddings": 5724.8735, "encoder_q-layer.0": 3649.9121, "encoder_q-layer.1": 3987.6023, "encoder_q-layer.10": 6805.1572, "encoder_q-layer.11": 16230.4072, "encoder_q-layer.2": 4473.5132, "encoder_q-layer.3": 4746.1064, "encoder_q-layer.4": 5037.8301, "encoder_q-layer.5": 5089.4287, "encoder_q-layer.6": 5713.9248, "encoder_q-layer.7": 6352.6411, "encoder_q-layer.8": 7512.3633, "encoder_q-layer.9": 6812.042, "epoch": 0.17, "inbatch_neg_score": 0.1393, "inbatch_pos_score": 0.6943, "learning_rate": 4.2473684210526315e-05, "loss": 3.7309, "norm_diff": 0.0955, "norm_loss": 0.0, "num_token_doc": 66.7265, "num_token_overlap": 11.6855, "num_token_query": 31.2949, "num_token_union": 65.0239, "num_word_context": 202.3013, "num_word_doc": 49.842, "num_word_query": 23.2459, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10550.7712, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1396, "query_norm": 1.2838, "queue_k_norm": 1.3751, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2949, "sent_len_1": 66.7265, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.4275, "stdk": 0.048, "stdq": 0.0425, "stdqueue_k": 0.0479, "stdqueue_q": 0.0, "step": 38600 }, { "accuracy": 46.9727, "active_queue_size": 16384.0, "cl_loss": 3.7383, "doc_norm": 1.3765, "encoder_q-embeddings": 6417.2861, "encoder_q-layer.0": 4467.4253, "encoder_q-layer.1": 4689.2046, "encoder_q-layer.10": 6594.2354, "encoder_q-layer.11": 15493.2529, "encoder_q-layer.2": 5294.9829, "encoder_q-layer.3": 5424.0342, "encoder_q-layer.4": 5566.6538, "encoder_q-layer.5": 5660.8862, "encoder_q-layer.6": 6198.0415, "encoder_q-layer.7": 6782.126, "encoder_q-layer.8": 7215.3638, "encoder_q-layer.9": 6431.9985, "epoch": 0.17, "inbatch_neg_score": 0.1446, "inbatch_pos_score": 0.7124, "learning_rate": 4.244736842105263e-05, "loss": 3.7383, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.5823, "num_token_overlap": 11.6142, "num_token_query": 31.2782, "num_token_union": 65.0079, "num_word_context": 202.383, "num_word_doc": 49.7151, "num_word_query": 23.2225, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10814.7725, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1444, "query_norm": 1.2941, "queue_k_norm": 1.3785, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2782, "sent_len_1": 66.5823, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.8363, "stdk": 0.0479, "stdq": 0.0427, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 38700 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.7427, "doc_norm": 1.377, "encoder_q-embeddings": 11430.4365, "encoder_q-layer.0": 8104.5156, "encoder_q-layer.1": 7866.5996, "encoder_q-layer.10": 7502.4893, "encoder_q-layer.11": 16809.3398, "encoder_q-layer.2": 9129.8594, "encoder_q-layer.3": 8582.4463, "encoder_q-layer.4": 8468.1221, "encoder_q-layer.5": 8118.9448, "encoder_q-layer.6": 8850.0205, "encoder_q-layer.7": 8487.4766, "encoder_q-layer.8": 7905.2373, "encoder_q-layer.9": 6800.9683, "epoch": 0.17, "inbatch_neg_score": 0.1515, "inbatch_pos_score": 0.7217, "learning_rate": 4.242105263157895e-05, "loss": 3.7427, "norm_diff": 0.0535, "norm_loss": 0.0, "num_token_doc": 66.6598, "num_token_overlap": 11.6688, "num_token_query": 31.3704, "num_token_union": 65.0209, "num_word_context": 202.1065, "num_word_doc": 49.7105, "num_word_query": 23.3013, "postclip_grad_norm": 1.0, "preclip_grad_norm": 14248.3083, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1523, "query_norm": 1.3235, "queue_k_norm": 1.3798, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3704, "sent_len_1": 66.6598, "sent_len_max_0": 127.9838, "sent_len_max_1": 211.1863, "stdk": 0.048, "stdq": 0.0435, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 38800 }, { "accuracy": 46.5332, "active_queue_size": 16384.0, "cl_loss": 3.7257, "doc_norm": 1.3842, "encoder_q-embeddings": 5481.9824, "encoder_q-layer.0": 3634.3521, "encoder_q-layer.1": 3749.877, "encoder_q-layer.10": 6755.1938, "encoder_q-layer.11": 15710.8457, "encoder_q-layer.2": 4282.1533, "encoder_q-layer.3": 4424.5317, "encoder_q-layer.4": 4640.9121, "encoder_q-layer.5": 4660.8091, "encoder_q-layer.6": 5186.8691, "encoder_q-layer.7": 6023.3896, "encoder_q-layer.8": 6875.5884, "encoder_q-layer.9": 6332.7793, "epoch": 0.17, "inbatch_neg_score": 0.1625, "inbatch_pos_score": 0.7188, "learning_rate": 4.239473684210526e-05, "loss": 3.7257, "norm_diff": 0.0652, "norm_loss": 0.0, "num_token_doc": 66.8796, "num_token_overlap": 11.6696, "num_token_query": 31.2961, "num_token_union": 65.1115, "num_word_context": 202.268, "num_word_doc": 49.8818, "num_word_query": 23.2477, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10108.7373, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1614, "query_norm": 1.3191, "queue_k_norm": 1.3802, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2961, "sent_len_1": 66.8796, "sent_len_max_0": 127.935, "sent_len_max_1": 210.2875, "stdk": 0.0482, "stdq": 0.0427, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 38900 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.7339, "doc_norm": 1.3843, "encoder_q-embeddings": 7877.9927, "encoder_q-layer.0": 5483.9004, "encoder_q-layer.1": 5591.3535, "encoder_q-layer.10": 6808.7852, "encoder_q-layer.11": 16413.2051, "encoder_q-layer.2": 4811.6211, "encoder_q-layer.3": 4542.5933, "encoder_q-layer.4": 4951.0615, "encoder_q-layer.5": 4827.8691, "encoder_q-layer.6": 5263.6768, "encoder_q-layer.7": 6044.6431, "encoder_q-layer.8": 7143.2231, "encoder_q-layer.9": 6425.875, "epoch": 0.17, "inbatch_neg_score": 0.1654, "inbatch_pos_score": 0.7505, "learning_rate": 4.236842105263158e-05, "loss": 3.7339, "norm_diff": 0.047, "norm_loss": 0.0, "num_token_doc": 66.8902, "num_token_overlap": 11.6898, "num_token_query": 31.3466, "num_token_union": 65.1644, "num_word_context": 202.1504, "num_word_doc": 49.9246, "num_word_query": 23.2724, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10912.0766, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1655, "query_norm": 1.3373, "queue_k_norm": 1.3811, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3466, "sent_len_1": 66.8902, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.6163, "stdk": 0.0481, "stdq": 0.0437, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39000 }, { "accuracy": 45.5566, "active_queue_size": 16384.0, "cl_loss": 3.724, "doc_norm": 1.3832, "encoder_q-embeddings": 26699.1953, "encoder_q-layer.0": 18344.7871, "encoder_q-layer.1": 17703.7871, "encoder_q-layer.10": 6634.0591, "encoder_q-layer.11": 15849.6562, "encoder_q-layer.2": 16387.9238, "encoder_q-layer.3": 14732.5322, "encoder_q-layer.4": 14593.6348, "encoder_q-layer.5": 14017.0273, "encoder_q-layer.6": 14667.1094, "encoder_q-layer.7": 14903.834, "encoder_q-layer.8": 10316.6865, "encoder_q-layer.9": 6873.3174, "epoch": 0.17, "inbatch_neg_score": 0.1707, "inbatch_pos_score": 0.7378, "learning_rate": 4.234210526315789e-05, "loss": 3.724, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.5291, "num_token_overlap": 11.6449, "num_token_query": 31.3005, "num_token_union": 64.9438, "num_word_context": 202.1106, "num_word_doc": 49.6297, "num_word_query": 23.2484, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24757.3836, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1702, "query_norm": 1.3103, "queue_k_norm": 1.3798, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3005, "sent_len_1": 66.5291, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.025, "stdk": 0.0481, "stdq": 0.0432, "stdqueue_k": 0.048, "stdqueue_q": 0.0, "step": 39100 }, { "accuracy": 47.3633, "active_queue_size": 16384.0, "cl_loss": 3.7374, "doc_norm": 1.384, "encoder_q-embeddings": 13150.5762, "encoder_q-layer.0": 9932.1299, "encoder_q-layer.1": 11873.7471, "encoder_q-layer.10": 7088.3853, "encoder_q-layer.11": 16200.0967, "encoder_q-layer.2": 12655.7666, "encoder_q-layer.3": 14277.373, "encoder_q-layer.4": 15412.1396, "encoder_q-layer.5": 14419.0098, "encoder_q-layer.6": 13612.6641, "encoder_q-layer.7": 12987.0928, "encoder_q-layer.8": 9037.6816, "encoder_q-layer.9": 6694.2158, "epoch": 0.17, "inbatch_neg_score": 0.1649, "inbatch_pos_score": 0.7363, "learning_rate": 4.231578947368421e-05, "loss": 3.7374, "norm_diff": 0.0897, "norm_loss": 0.0, "num_token_doc": 66.9503, "num_token_overlap": 11.7085, "num_token_query": 31.4209, "num_token_union": 65.1935, "num_word_context": 202.3101, "num_word_doc": 49.9335, "num_word_query": 23.344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18541.433, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1653, "query_norm": 1.2944, "queue_k_norm": 1.3843, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4209, "sent_len_1": 66.9503, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7012, "stdk": 0.0481, "stdq": 0.0432, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39200 }, { "accuracy": 46.8262, "active_queue_size": 16384.0, "cl_loss": 3.7237, "doc_norm": 1.3889, "encoder_q-embeddings": 22742.0137, "encoder_q-layer.0": 15164.0225, "encoder_q-layer.1": 17775.4238, "encoder_q-layer.10": 13688.0889, "encoder_q-layer.11": 33814.6602, "encoder_q-layer.2": 20221.3203, "encoder_q-layer.3": 20381.5977, "encoder_q-layer.4": 21594.8828, "encoder_q-layer.5": 22466.791, "encoder_q-layer.6": 23653.6152, "encoder_q-layer.7": 22443.0469, "encoder_q-layer.8": 20157.1543, "encoder_q-layer.9": 14853.6035, "epoch": 0.17, "inbatch_neg_score": 0.1603, "inbatch_pos_score": 0.7227, "learning_rate": 4.228947368421053e-05, "loss": 3.7237, "norm_diff": 0.1241, "norm_loss": 0.0, "num_token_doc": 66.7213, "num_token_overlap": 11.683, "num_token_query": 31.3771, "num_token_union": 65.0961, "num_word_context": 202.3246, "num_word_doc": 49.7878, "num_word_query": 23.2971, "postclip_grad_norm": 1.0, "preclip_grad_norm": 31994.5474, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 0.1603, "query_norm": 1.2648, "queue_k_norm": 1.3831, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3771, "sent_len_1": 66.7213, "sent_len_max_0": 127.9525, "sent_len_max_1": 208.3038, "stdk": 0.0482, "stdq": 0.0424, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39300 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.7147, "doc_norm": 1.3835, "encoder_q-embeddings": 11026.3438, "encoder_q-layer.0": 7134.3062, "encoder_q-layer.1": 7554.5098, "encoder_q-layer.10": 14151.708, "encoder_q-layer.11": 30874.3203, "encoder_q-layer.2": 8220.666, "encoder_q-layer.3": 8711.7637, "encoder_q-layer.4": 9156.8818, "encoder_q-layer.5": 9430.8145, "encoder_q-layer.6": 10668.7471, "encoder_q-layer.7": 11947.043, "encoder_q-layer.8": 15292.8857, "encoder_q-layer.9": 13788.1094, "epoch": 0.17, "inbatch_neg_score": 0.1575, "inbatch_pos_score": 0.729, "learning_rate": 4.226315789473684e-05, "loss": 3.7147, "norm_diff": 0.1018, "norm_loss": 0.0, "num_token_doc": 66.8923, "num_token_overlap": 11.678, "num_token_query": 31.3624, "num_token_union": 65.1764, "num_word_context": 202.3248, "num_word_doc": 49.9233, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20079.1836, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1583, "query_norm": 1.2817, "queue_k_norm": 1.3853, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3624, "sent_len_1": 66.8923, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.6987, "stdk": 0.048, "stdq": 0.0433, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39400 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.7245, "doc_norm": 1.3846, "encoder_q-embeddings": 13320.8779, "encoder_q-layer.0": 8788.3018, "encoder_q-layer.1": 9409.8887, "encoder_q-layer.10": 16459.8457, "encoder_q-layer.11": 34201.7305, "encoder_q-layer.2": 10982.3672, "encoder_q-layer.3": 11690.7588, "encoder_q-layer.4": 12283.1309, "encoder_q-layer.5": 13648.4209, "encoder_q-layer.6": 14569.5947, "encoder_q-layer.7": 14303.3408, "encoder_q-layer.8": 17536.6641, "encoder_q-layer.9": 15112.9541, "epoch": 0.17, "inbatch_neg_score": 0.1527, "inbatch_pos_score": 0.7319, "learning_rate": 4.223684210526316e-05, "loss": 3.7245, "norm_diff": 0.132, "norm_loss": 0.0, "num_token_doc": 66.7596, "num_token_overlap": 11.6714, "num_token_query": 31.3899, "num_token_union": 65.1259, "num_word_context": 202.2064, "num_word_doc": 49.795, "num_word_query": 23.327, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23166.2871, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1531, "query_norm": 1.2526, "queue_k_norm": 1.3834, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3899, "sent_len_1": 66.7596, "sent_len_max_0": 127.97, "sent_len_max_1": 210.0375, "stdk": 0.048, "stdq": 0.0423, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39500 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.7162, "doc_norm": 1.3828, "encoder_q-embeddings": 12071.5371, "encoder_q-layer.0": 7710.791, "encoder_q-layer.1": 8172.6738, "encoder_q-layer.10": 14071.168, "encoder_q-layer.11": 32454.2695, "encoder_q-layer.2": 8989.625, "encoder_q-layer.3": 9233.0205, "encoder_q-layer.4": 9741.1621, "encoder_q-layer.5": 9778.8037, "encoder_q-layer.6": 10639.1475, "encoder_q-layer.7": 12154.6182, "encoder_q-layer.8": 14404.9844, "encoder_q-layer.9": 13022.1416, "epoch": 0.17, "inbatch_neg_score": 0.1478, "inbatch_pos_score": 0.7124, "learning_rate": 4.221052631578948e-05, "loss": 3.7162, "norm_diff": 0.1197, "norm_loss": 0.0, "num_token_doc": 66.629, "num_token_overlap": 11.6477, "num_token_query": 31.2751, "num_token_union": 65.0006, "num_word_context": 202.0239, "num_word_doc": 49.7458, "num_word_query": 23.2297, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21348.7318, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1482, "query_norm": 1.2631, "queue_k_norm": 1.3866, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2751, "sent_len_1": 66.629, "sent_len_max_0": 127.9875, "sent_len_max_1": 206.2125, "stdk": 0.048, "stdq": 0.0428, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 39600 }, { "accuracy": 47.7539, "active_queue_size": 16384.0, "cl_loss": 3.7227, "doc_norm": 1.387, "encoder_q-embeddings": 11866.1689, "encoder_q-layer.0": 7852.2642, "encoder_q-layer.1": 8299.9492, "encoder_q-layer.10": 12948.7354, "encoder_q-layer.11": 30182.0117, "encoder_q-layer.2": 9036.874, "encoder_q-layer.3": 9233.5381, "encoder_q-layer.4": 9579.1504, "encoder_q-layer.5": 9759.0869, "encoder_q-layer.6": 10786.9746, "encoder_q-layer.7": 11802.2217, "encoder_q-layer.8": 13915.9043, "encoder_q-layer.9": 12534.3838, "epoch": 0.17, "inbatch_neg_score": 0.1521, "inbatch_pos_score": 0.7319, "learning_rate": 4.218421052631579e-05, "loss": 3.7227, "norm_diff": 0.109, "norm_loss": 0.0, "num_token_doc": 66.8353, "num_token_overlap": 11.6941, "num_token_query": 31.3995, "num_token_union": 65.1802, "num_word_context": 202.4545, "num_word_doc": 49.9046, "num_word_query": 23.3236, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20293.0341, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1519, "query_norm": 1.278, "queue_k_norm": 1.3844, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3995, "sent_len_1": 66.8353, "sent_len_max_0": 127.985, "sent_len_max_1": 208.5087, "stdk": 0.0481, "stdq": 0.0433, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39700 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.7232, "doc_norm": 1.385, "encoder_q-embeddings": 10799.9365, "encoder_q-layer.0": 7168.1226, "encoder_q-layer.1": 7459.5557, "encoder_q-layer.10": 14576.9443, "encoder_q-layer.11": 33716.3789, "encoder_q-layer.2": 8356.0459, "encoder_q-layer.3": 8681.3584, "encoder_q-layer.4": 9212.4268, "encoder_q-layer.5": 9161.4541, "encoder_q-layer.6": 10256.1543, "encoder_q-layer.7": 11708.4316, "encoder_q-layer.8": 14213.3672, "encoder_q-layer.9": 12773.3125, "epoch": 0.17, "inbatch_neg_score": 0.15, "inbatch_pos_score": 0.7139, "learning_rate": 4.215789473684211e-05, "loss": 3.7232, "norm_diff": 0.1231, "norm_loss": 0.0, "num_token_doc": 66.6852, "num_token_overlap": 11.6787, "num_token_query": 31.3871, "num_token_union": 65.1012, "num_word_context": 202.2762, "num_word_doc": 49.7644, "num_word_query": 23.3111, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20982.4363, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1497, "query_norm": 1.262, "queue_k_norm": 1.3861, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3871, "sent_len_1": 66.6852, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.195, "stdk": 0.0481, "stdq": 0.0427, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39800 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.7172, "doc_norm": 1.3877, "encoder_q-embeddings": 10655.2803, "encoder_q-layer.0": 6889.6777, "encoder_q-layer.1": 7146.6538, "encoder_q-layer.10": 16758.9941, "encoder_q-layer.11": 34901.7422, "encoder_q-layer.2": 8424.6748, "encoder_q-layer.3": 8784.7432, "encoder_q-layer.4": 9463.709, "encoder_q-layer.5": 9360.5635, "encoder_q-layer.6": 10523.5225, "encoder_q-layer.7": 12383.8574, "encoder_q-layer.8": 15307.9043, "encoder_q-layer.9": 14301.8789, "epoch": 0.17, "inbatch_neg_score": 0.1457, "inbatch_pos_score": 0.7109, "learning_rate": 4.213157894736842e-05, "loss": 3.7172, "norm_diff": 0.1541, "norm_loss": 0.0, "num_token_doc": 66.7127, "num_token_overlap": 11.6604, "num_token_query": 31.3795, "num_token_union": 65.1144, "num_word_context": 202.0286, "num_word_doc": 49.8227, "num_word_query": 23.3091, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20889.614, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1461, "query_norm": 1.2336, "queue_k_norm": 1.384, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3795, "sent_len_1": 66.7127, "sent_len_max_0": 127.9838, "sent_len_max_1": 205.4387, "stdk": 0.0481, "stdq": 0.0417, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 39900 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.6992, "doc_norm": 1.3813, "encoder_q-embeddings": 10903.0898, "encoder_q-layer.0": 7180.5488, "encoder_q-layer.1": 7589.2886, "encoder_q-layer.10": 13526.8369, "encoder_q-layer.11": 31227.793, "encoder_q-layer.2": 8345.5518, "encoder_q-layer.3": 8760.251, "encoder_q-layer.4": 9083.2568, "encoder_q-layer.5": 9496.6152, "encoder_q-layer.6": 10920.709, "encoder_q-layer.7": 11925.3125, "encoder_q-layer.8": 14612.71, "encoder_q-layer.9": 12984.2949, "epoch": 0.17, "inbatch_neg_score": 0.1436, "inbatch_pos_score": 0.708, "learning_rate": 4.210526315789474e-05, "loss": 3.6992, "norm_diff": 0.1316, "norm_loss": 0.0, "num_token_doc": 66.8052, "num_token_overlap": 11.6999, "num_token_query": 31.3533, "num_token_union": 65.0978, "num_word_context": 202.2709, "num_word_doc": 49.8422, "num_word_query": 23.2926, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20100.8362, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1428, "query_norm": 1.2497, "queue_k_norm": 1.3861, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3533, "sent_len_1": 66.8052, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.3988, "stdk": 0.0479, "stdq": 0.0423, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 40000 }, { "dev_runtime": 28.3565, "dev_samples_per_second": 1.128, "dev_steps_per_second": 0.035, "epoch": 0.17, "step": 40000, "test_accuracy": 92.88330078125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.40745043754577637, "test_doc_norm": 1.3490679264068604, "test_inbatch_neg_score": 0.40154141187667847, "test_inbatch_pos_score": 1.35885488986969, "test_loss": 0.40745043754577637, "test_loss_align": 1.0345231294631958, "test_loss_unif": 3.942417860031128, "test_loss_unif_q@queue": 3.942417860031128, "test_norm_diff": 0.02755141258239746, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.13850966095924377, "test_query_norm": 1.3766192197799683, "test_queue_k_norm": 1.3847103118896484, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04287203401327133, "test_stdq": 0.04270744323730469, "test_stdqueue_k": 0.04811163991689682, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.3565, "dev_samples_per_second": 1.128, "dev_steps_per_second": 0.035, "epoch": 0.17, "eval_beir-arguana_ndcg@10": 0.35333, "eval_beir-arguana_recall@10": 0.61024, "eval_beir-arguana_recall@100": 0.91963, "eval_beir-arguana_recall@20": 0.75462, "eval_beir-avg_ndcg@10": 0.3814010833333333, "eval_beir-avg_recall@10": 0.44954333333333335, "eval_beir-avg_recall@100": 0.6334318333333333, "eval_beir-avg_recall@20": 0.5146191666666666, "eval_beir-cqadupstack_ndcg@10": 0.2555608333333333, "eval_beir-cqadupstack_recall@10": 0.34996333333333324, "eval_beir-cqadupstack_recall@100": 0.5855183333333333, "eval_beir-cqadupstack_recall@20": 0.41883166666666666, "eval_beir-fiqa_ndcg@10": 0.23532, "eval_beir-fiqa_recall@10": 0.29033, "eval_beir-fiqa_recall@100": 0.57249, "eval_beir-fiqa_recall@20": 0.37025, "eval_beir-nfcorpus_ndcg@10": 0.29777, "eval_beir-nfcorpus_recall@10": 0.15075, "eval_beir-nfcorpus_recall@100": 0.28188, "eval_beir-nfcorpus_recall@20": 0.17744, "eval_beir-nq_ndcg@10": 0.28833, "eval_beir-nq_recall@10": 0.46678, "eval_beir-nq_recall@100": 0.80234, "eval_beir-nq_recall@20": 0.58647, "eval_beir-quora_ndcg@10": 0.77387, "eval_beir-quora_recall@10": 0.88271, "eval_beir-quora_recall@100": 0.97775, "eval_beir-quora_recall@20": 0.92584, "eval_beir-scidocs_ndcg@10": 0.1465, "eval_beir-scidocs_recall@10": 0.15472, "eval_beir-scidocs_recall@100": 0.35852, "eval_beir-scidocs_recall@20": 0.21612, "eval_beir-scifact_ndcg@10": 0.64086, "eval_beir-scifact_recall@10": 0.78833, "eval_beir-scifact_recall@100": 0.926, "eval_beir-scifact_recall@20": 0.83744, "eval_beir-trec-covid_ndcg@10": 0.61239, "eval_beir-trec-covid_recall@10": 0.664, "eval_beir-trec-covid_recall@100": 0.4632, "eval_beir-trec-covid_recall@20": 0.633, "eval_beir-webis-touche2020_ndcg@10": 0.21008, "eval_beir-webis-touche2020_recall@10": 0.13761, "eval_beir-webis-touche2020_recall@100": 0.44699, "eval_beir-webis-touche2020_recall@20": 0.22618, "eval_senteval-avg_sts": 0.7667332946949093, "eval_senteval-sickr_spearman": 0.7313564210763474, "eval_senteval-stsb_spearman": 0.8021101683134713, "step": 40000, "test_accuracy": 92.88330078125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.40745043754577637, "test_doc_norm": 1.3490679264068604, "test_inbatch_neg_score": 0.40154141187667847, "test_inbatch_pos_score": 1.35885488986969, "test_loss": 0.40745043754577637, "test_loss_align": 1.0345231294631958, "test_loss_unif": 3.942417860031128, "test_loss_unif_q@queue": 3.942417860031128, "test_norm_diff": 0.02755141258239746, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.13850966095924377, "test_query_norm": 1.3766192197799683, "test_queue_k_norm": 1.3847103118896484, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04287203401327133, "test_stdq": 0.04270744323730469, "test_stdqueue_k": 0.04811163991689682, "test_stdqueue_q": 0.0 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.704, "doc_norm": 1.3838, "encoder_q-embeddings": 11872.3867, "encoder_q-layer.0": 7286.5146, "encoder_q-layer.1": 7755.2349, "encoder_q-layer.10": 13257.5449, "encoder_q-layer.11": 30708.2637, "encoder_q-layer.2": 8387.459, "encoder_q-layer.3": 8961.4443, "encoder_q-layer.4": 9266.6895, "encoder_q-layer.5": 9693.6592, "encoder_q-layer.6": 10529.585, "encoder_q-layer.7": 12194.0264, "encoder_q-layer.8": 14582.4668, "encoder_q-layer.9": 12818.0381, "epoch": 0.17, "inbatch_neg_score": 0.1375, "inbatch_pos_score": 0.71, "learning_rate": 4.2078947368421056e-05, "loss": 3.704, "norm_diff": 0.1175, "norm_loss": 0.0, "num_token_doc": 66.676, "num_token_overlap": 11.6736, "num_token_query": 31.39, "num_token_union": 65.0698, "num_word_context": 202.1961, "num_word_doc": 49.7377, "num_word_query": 23.3143, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20377.382, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1377, "query_norm": 1.2663, "queue_k_norm": 1.3838, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.39, "sent_len_1": 66.676, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.9613, "stdk": 0.048, "stdq": 0.043, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40100 }, { "accuracy": 46.1426, "active_queue_size": 16384.0, "cl_loss": 3.7122, "doc_norm": 1.3827, "encoder_q-embeddings": 10849.3135, "encoder_q-layer.0": 7047.9067, "encoder_q-layer.1": 7309.269, "encoder_q-layer.10": 12850.8516, "encoder_q-layer.11": 32314.8984, "encoder_q-layer.2": 8285.3867, "encoder_q-layer.3": 8459.6816, "encoder_q-layer.4": 8846.9326, "encoder_q-layer.5": 9078.4844, "encoder_q-layer.6": 10297.9746, "encoder_q-layer.7": 11568.9102, "encoder_q-layer.8": 14430.1309, "encoder_q-layer.9": 12388.2305, "epoch": 0.17, "inbatch_neg_score": 0.1368, "inbatch_pos_score": 0.6973, "learning_rate": 4.2052631578947375e-05, "loss": 3.7122, "norm_diff": 0.1322, "norm_loss": 0.0, "num_token_doc": 66.7802, "num_token_overlap": 11.6849, "num_token_query": 31.3888, "num_token_union": 65.1163, "num_word_context": 202.2149, "num_word_doc": 49.846, "num_word_query": 23.3183, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20254.296, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1355, "query_norm": 1.2505, "queue_k_norm": 1.3851, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3888, "sent_len_1": 66.7802, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.4837, "stdk": 0.048, "stdq": 0.0424, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40200 }, { "accuracy": 45.0195, "active_queue_size": 16384.0, "cl_loss": 3.716, "doc_norm": 1.381, "encoder_q-embeddings": 11391.665, "encoder_q-layer.0": 7350.9502, "encoder_q-layer.1": 7742.4531, "encoder_q-layer.10": 13893.4111, "encoder_q-layer.11": 32109.7168, "encoder_q-layer.2": 8571.6201, "encoder_q-layer.3": 8976.0537, "encoder_q-layer.4": 9494.085, "encoder_q-layer.5": 9757.0459, "encoder_q-layer.6": 10621.3623, "encoder_q-layer.7": 11950.0811, "encoder_q-layer.8": 14706.7988, "encoder_q-layer.9": 13052.0811, "epoch": 0.17, "inbatch_neg_score": 0.1338, "inbatch_pos_score": 0.6855, "learning_rate": 4.2026315789473686e-05, "loss": 3.716, "norm_diff": 0.1229, "norm_loss": 0.0, "num_token_doc": 66.8502, "num_token_overlap": 11.6775, "num_token_query": 31.3816, "num_token_union": 65.193, "num_word_context": 202.6329, "num_word_doc": 49.8724, "num_word_query": 23.326, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20476.215, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1331, "query_norm": 1.2581, "queue_k_norm": 1.3829, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3816, "sent_len_1": 66.8502, "sent_len_max_0": 127.995, "sent_len_max_1": 208.845, "stdk": 0.0479, "stdq": 0.0427, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40300 }, { "accuracy": 47.3145, "active_queue_size": 16384.0, "cl_loss": 3.7127, "doc_norm": 1.3855, "encoder_q-embeddings": 10804.7861, "encoder_q-layer.0": 7063.6704, "encoder_q-layer.1": 7389.0693, "encoder_q-layer.10": 12798.2441, "encoder_q-layer.11": 30627.6484, "encoder_q-layer.2": 8225.332, "encoder_q-layer.3": 8887.7109, "encoder_q-layer.4": 8896.999, "encoder_q-layer.5": 9006.3301, "encoder_q-layer.6": 10066.2891, "encoder_q-layer.7": 11464.7324, "encoder_q-layer.8": 13535.9307, "encoder_q-layer.9": 12266.7373, "epoch": 0.18, "inbatch_neg_score": 0.1365, "inbatch_pos_score": 0.7114, "learning_rate": 4.2e-05, "loss": 3.7127, "norm_diff": 0.1072, "norm_loss": 0.0, "num_token_doc": 66.7017, "num_token_overlap": 11.699, "num_token_query": 31.3921, "num_token_union": 65.0891, "num_word_context": 202.5109, "num_word_doc": 49.7914, "num_word_query": 23.3114, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19583.3003, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1362, "query_norm": 1.2784, "queue_k_norm": 1.3844, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3921, "sent_len_1": 66.7017, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.3288, "stdk": 0.0481, "stdq": 0.0432, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40400 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.7126, "doc_norm": 1.387, "encoder_q-embeddings": 11352.2676, "encoder_q-layer.0": 7184.0713, "encoder_q-layer.1": 7725.8413, "encoder_q-layer.10": 14540.7959, "encoder_q-layer.11": 32323.8105, "encoder_q-layer.2": 8689.957, "encoder_q-layer.3": 9186.4346, "encoder_q-layer.4": 9285.9355, "encoder_q-layer.5": 9355.4707, "encoder_q-layer.6": 10589.9639, "encoder_q-layer.7": 11904.459, "encoder_q-layer.8": 13757.8799, "encoder_q-layer.9": 13135.1465, "epoch": 0.18, "inbatch_neg_score": 0.138, "inbatch_pos_score": 0.7061, "learning_rate": 4.1973684210526316e-05, "loss": 3.7126, "norm_diff": 0.1245, "norm_loss": 0.0, "num_token_doc": 66.8055, "num_token_overlap": 11.6569, "num_token_query": 31.3579, "num_token_union": 65.1562, "num_word_context": 202.3396, "num_word_doc": 49.8083, "num_word_query": 23.2889, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20745.0174, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1379, "query_norm": 1.2625, "queue_k_norm": 1.3844, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3579, "sent_len_1": 66.8055, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.8113, "stdk": 0.0482, "stdq": 0.0424, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 40500 }, { "accuracy": 46.8262, "active_queue_size": 16384.0, "cl_loss": 3.718, "doc_norm": 1.3887, "encoder_q-embeddings": 11445.4775, "encoder_q-layer.0": 7503.3252, "encoder_q-layer.1": 7887.5068, "encoder_q-layer.10": 13487.3379, "encoder_q-layer.11": 32154.1113, "encoder_q-layer.2": 8829.0791, "encoder_q-layer.3": 9235.4727, "encoder_q-layer.4": 9605.7363, "encoder_q-layer.5": 10167.1738, "encoder_q-layer.6": 11406.8857, "encoder_q-layer.7": 13151.9727, "encoder_q-layer.8": 16267.6416, "encoder_q-layer.9": 13488.2217, "epoch": 0.18, "inbatch_neg_score": 0.1385, "inbatch_pos_score": 0.7095, "learning_rate": 4.1947368421052634e-05, "loss": 3.718, "norm_diff": 0.1056, "norm_loss": 0.0, "num_token_doc": 66.8403, "num_token_overlap": 11.6521, "num_token_query": 31.3784, "num_token_union": 65.1912, "num_word_context": 202.6572, "num_word_doc": 49.9075, "num_word_query": 23.2973, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21091.7727, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1383, "query_norm": 1.283, "queue_k_norm": 1.3849, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3784, "sent_len_1": 66.8403, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.2337, "stdk": 0.0483, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 40600 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.696, "doc_norm": 1.3843, "encoder_q-embeddings": 10891.041, "encoder_q-layer.0": 7130.0698, "encoder_q-layer.1": 7556.8604, "encoder_q-layer.10": 13455.7217, "encoder_q-layer.11": 32143.9629, "encoder_q-layer.2": 8326.7422, "encoder_q-layer.3": 8619.167, "encoder_q-layer.4": 9161.0488, "encoder_q-layer.5": 9390.7793, "encoder_q-layer.6": 10111.3447, "encoder_q-layer.7": 11194.8115, "encoder_q-layer.8": 13903.3418, "encoder_q-layer.9": 12530.8828, "epoch": 0.18, "inbatch_neg_score": 0.1394, "inbatch_pos_score": 0.7056, "learning_rate": 4.192105263157895e-05, "loss": 3.696, "norm_diff": 0.0998, "norm_loss": 0.0, "num_token_doc": 66.8805, "num_token_overlap": 11.732, "num_token_query": 31.5014, "num_token_union": 65.2324, "num_word_context": 202.386, "num_word_doc": 49.917, "num_word_query": 23.4028, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20105.2374, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1399, "query_norm": 1.2845, "queue_k_norm": 1.3832, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5014, "sent_len_1": 66.8805, "sent_len_max_0": 127.985, "sent_len_max_1": 208.5325, "stdk": 0.0481, "stdq": 0.0425, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40700 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.6987, "doc_norm": 1.382, "encoder_q-embeddings": 11521.5742, "encoder_q-layer.0": 7411.4956, "encoder_q-layer.1": 7956.9365, "encoder_q-layer.10": 14647.0137, "encoder_q-layer.11": 32086.0898, "encoder_q-layer.2": 8793.8242, "encoder_q-layer.3": 8979.7305, "encoder_q-layer.4": 9334.8799, "encoder_q-layer.5": 9145.834, "encoder_q-layer.6": 10315.9482, "encoder_q-layer.7": 11774.4004, "encoder_q-layer.8": 13887.2119, "encoder_q-layer.9": 13032.0244, "epoch": 0.18, "inbatch_neg_score": 0.149, "inbatch_pos_score": 0.729, "learning_rate": 4.1894736842105264e-05, "loss": 3.6987, "norm_diff": 0.083, "norm_loss": 0.0, "num_token_doc": 66.8309, "num_token_overlap": 11.7459, "num_token_query": 31.5205, "num_token_union": 65.1952, "num_word_context": 202.2957, "num_word_doc": 49.8943, "num_word_query": 23.4111, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20414.0619, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1488, "query_norm": 1.299, "queue_k_norm": 1.3856, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5205, "sent_len_1": 66.8309, "sent_len_max_0": 128.0, "sent_len_max_1": 208.3063, "stdk": 0.048, "stdq": 0.0428, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 40800 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 3.7088, "doc_norm": 1.3825, "encoder_q-embeddings": 12246.6562, "encoder_q-layer.0": 7741.1519, "encoder_q-layer.1": 8586.2686, "encoder_q-layer.10": 15288.2012, "encoder_q-layer.11": 33289.7031, "encoder_q-layer.2": 9183.501, "encoder_q-layer.3": 9508.7607, "encoder_q-layer.4": 10059.4248, "encoder_q-layer.5": 9891.0, "encoder_q-layer.6": 11223.0254, "encoder_q-layer.7": 12474.1387, "encoder_q-layer.8": 15795.0576, "encoder_q-layer.9": 13815.5117, "epoch": 0.18, "inbatch_neg_score": 0.1534, "inbatch_pos_score": 0.7349, "learning_rate": 4.1868421052631576e-05, "loss": 3.7088, "norm_diff": 0.0548, "norm_loss": 0.0, "num_token_doc": 66.7556, "num_token_overlap": 11.6837, "num_token_query": 31.4283, "num_token_union": 65.1319, "num_word_context": 202.2045, "num_word_doc": 49.8113, "num_word_query": 23.3523, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21801.2618, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1538, "query_norm": 1.3277, "queue_k_norm": 1.384, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4283, "sent_len_1": 66.7556, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.1625, "stdk": 0.048, "stdq": 0.0438, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 40900 }, { "accuracy": 46.582, "active_queue_size": 16384.0, "cl_loss": 3.7118, "doc_norm": 1.3865, "encoder_q-embeddings": 14280.5342, "encoder_q-layer.0": 9541.0879, "encoder_q-layer.1": 10006.2891, "encoder_q-layer.10": 14054.0557, "encoder_q-layer.11": 32031.4531, "encoder_q-layer.2": 11279.6562, "encoder_q-layer.3": 11337.959, "encoder_q-layer.4": 11163.7383, "encoder_q-layer.5": 10841.168, "encoder_q-layer.6": 11795.6113, "encoder_q-layer.7": 12777.8799, "encoder_q-layer.8": 14987.3184, "encoder_q-layer.9": 13244.4219, "epoch": 0.18, "inbatch_neg_score": 0.1615, "inbatch_pos_score": 0.7334, "learning_rate": 4.18421052631579e-05, "loss": 3.7118, "norm_diff": 0.0701, "norm_loss": 0.0, "num_token_doc": 66.8896, "num_token_overlap": 11.6768, "num_token_query": 31.3913, "num_token_union": 65.2024, "num_word_context": 202.5913, "num_word_doc": 49.9437, "num_word_query": 23.3241, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22435.0744, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1617, "query_norm": 1.3164, "queue_k_norm": 1.3859, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3913, "sent_len_1": 66.8896, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.9112, "stdk": 0.0482, "stdq": 0.0432, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41000 }, { "accuracy": 45.9473, "active_queue_size": 16384.0, "cl_loss": 3.6967, "doc_norm": 1.3874, "encoder_q-embeddings": 11685.3408, "encoder_q-layer.0": 7549.3413, "encoder_q-layer.1": 7972.1777, "encoder_q-layer.10": 13839.0537, "encoder_q-layer.11": 32700.4297, "encoder_q-layer.2": 8628.1309, "encoder_q-layer.3": 9153.2842, "encoder_q-layer.4": 9824.9453, "encoder_q-layer.5": 10064.374, "encoder_q-layer.6": 11212.8926, "encoder_q-layer.7": 13133.9639, "encoder_q-layer.8": 14736.5635, "encoder_q-layer.9": 13333.1982, "epoch": 0.18, "inbatch_neg_score": 0.1695, "inbatch_pos_score": 0.7354, "learning_rate": 4.181578947368421e-05, "loss": 3.6967, "norm_diff": 0.0864, "norm_loss": 0.0, "num_token_doc": 66.7385, "num_token_overlap": 11.6611, "num_token_query": 31.3084, "num_token_union": 65.0883, "num_word_context": 202.1307, "num_word_doc": 49.8018, "num_word_query": 23.2321, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20965.8903, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1686, "query_norm": 1.301, "queue_k_norm": 1.3859, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3084, "sent_len_1": 66.7385, "sent_len_max_0": 127.985, "sent_len_max_1": 209.1575, "stdk": 0.0482, "stdq": 0.0428, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41100 }, { "accuracy": 47.2656, "active_queue_size": 16384.0, "cl_loss": 3.705, "doc_norm": 1.3896, "encoder_q-embeddings": 11495.8867, "encoder_q-layer.0": 7695.1543, "encoder_q-layer.1": 8116.6387, "encoder_q-layer.10": 13318.7432, "encoder_q-layer.11": 31537.3789, "encoder_q-layer.2": 8917.8672, "encoder_q-layer.3": 9149.541, "encoder_q-layer.4": 9424.3887, "encoder_q-layer.5": 9596.7148, "encoder_q-layer.6": 9978.75, "encoder_q-layer.7": 11217.2812, "encoder_q-layer.8": 13583.167, "encoder_q-layer.9": 12464.374, "epoch": 0.18, "inbatch_neg_score": 0.1706, "inbatch_pos_score": 0.7446, "learning_rate": 4.178947368421053e-05, "loss": 3.705, "norm_diff": 0.0894, "norm_loss": 0.0, "num_token_doc": 66.9557, "num_token_overlap": 11.7069, "num_token_query": 31.46, "num_token_union": 65.2586, "num_word_context": 202.6569, "num_word_doc": 49.9602, "num_word_query": 23.3805, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20126.6552, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1699, "query_norm": 1.3002, "queue_k_norm": 1.386, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.46, "sent_len_1": 66.9557, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.895, "stdk": 0.0482, "stdq": 0.0432, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41200 }, { "accuracy": 47.2656, "active_queue_size": 16384.0, "cl_loss": 3.6985, "doc_norm": 1.3895, "encoder_q-embeddings": 5396.1963, "encoder_q-layer.0": 3516.585, "encoder_q-layer.1": 3771.0544, "encoder_q-layer.10": 6527.7891, "encoder_q-layer.11": 15350.166, "encoder_q-layer.2": 4206.3726, "encoder_q-layer.3": 4405.4722, "encoder_q-layer.4": 4526.9463, "encoder_q-layer.5": 4646.7007, "encoder_q-layer.6": 5310.6514, "encoder_q-layer.7": 5754.7393, "encoder_q-layer.8": 7083.6963, "encoder_q-layer.9": 6323.9316, "epoch": 0.18, "inbatch_neg_score": 0.1696, "inbatch_pos_score": 0.7417, "learning_rate": 4.176315789473684e-05, "loss": 3.6985, "norm_diff": 0.1046, "norm_loss": 0.0, "num_token_doc": 66.8318, "num_token_overlap": 11.6628, "num_token_query": 31.3615, "num_token_union": 65.1532, "num_word_context": 202.3365, "num_word_doc": 49.8899, "num_word_query": 23.2892, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10099.6097, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1691, "query_norm": 1.2848, "queue_k_norm": 1.3872, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3615, "sent_len_1": 66.8318, "sent_len_max_0": 127.975, "sent_len_max_1": 206.9487, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41300 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.7141, "doc_norm": 1.3901, "encoder_q-embeddings": 5750.7588, "encoder_q-layer.0": 3659.0547, "encoder_q-layer.1": 3908.3567, "encoder_q-layer.10": 7484.4487, "encoder_q-layer.11": 17237.8457, "encoder_q-layer.2": 4334.3003, "encoder_q-layer.3": 4463.5898, "encoder_q-layer.4": 4690.3794, "encoder_q-layer.5": 4945.8013, "encoder_q-layer.6": 5491.3501, "encoder_q-layer.7": 6194.0605, "encoder_q-layer.8": 7389.3711, "encoder_q-layer.9": 6601.8481, "epoch": 0.18, "inbatch_neg_score": 0.1669, "inbatch_pos_score": 0.7441, "learning_rate": 4.1736842105263154e-05, "loss": 3.7141, "norm_diff": 0.1091, "norm_loss": 0.0, "num_token_doc": 66.8353, "num_token_overlap": 11.6524, "num_token_query": 31.3299, "num_token_union": 65.1538, "num_word_context": 202.3604, "num_word_doc": 49.8818, "num_word_query": 23.2649, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10891.7599, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1676, "query_norm": 1.281, "queue_k_norm": 1.3895, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3299, "sent_len_1": 66.8353, "sent_len_max_0": 127.97, "sent_len_max_1": 209.8137, "stdk": 0.0482, "stdq": 0.043, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41400 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.6969, "doc_norm": 1.3926, "encoder_q-embeddings": 5684.8667, "encoder_q-layer.0": 3732.4685, "encoder_q-layer.1": 3972.7407, "encoder_q-layer.10": 6643.8364, "encoder_q-layer.11": 15655.041, "encoder_q-layer.2": 4488.2056, "encoder_q-layer.3": 4632.5195, "encoder_q-layer.4": 4904.5356, "encoder_q-layer.5": 4856.3662, "encoder_q-layer.6": 5468.2856, "encoder_q-layer.7": 5921.9365, "encoder_q-layer.8": 6741.9302, "encoder_q-layer.9": 6234.0146, "epoch": 0.18, "inbatch_neg_score": 0.1695, "inbatch_pos_score": 0.7578, "learning_rate": 4.171052631578948e-05, "loss": 3.6969, "norm_diff": 0.1038, "norm_loss": 0.0, "num_token_doc": 66.7623, "num_token_overlap": 11.6549, "num_token_query": 31.3718, "num_token_union": 65.1264, "num_word_context": 202.5009, "num_word_doc": 49.8366, "num_word_query": 23.3112, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10188.5237, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1694, "query_norm": 1.2888, "queue_k_norm": 1.3904, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3718, "sent_len_1": 66.7623, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.4988, "stdk": 0.0483, "stdq": 0.0432, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 41500 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.7051, "doc_norm": 1.3853, "encoder_q-embeddings": 6167.1768, "encoder_q-layer.0": 4254.5029, "encoder_q-layer.1": 4341.2783, "encoder_q-layer.10": 6708.417, "encoder_q-layer.11": 16180.6543, "encoder_q-layer.2": 4871.5508, "encoder_q-layer.3": 5041.8564, "encoder_q-layer.4": 5324.7788, "encoder_q-layer.5": 5479.7612, "encoder_q-layer.6": 6100.6357, "encoder_q-layer.7": 6976.2559, "encoder_q-layer.8": 7400.7515, "encoder_q-layer.9": 6480.6543, "epoch": 0.18, "inbatch_neg_score": 0.1693, "inbatch_pos_score": 0.7319, "learning_rate": 4.168421052631579e-05, "loss": 3.7051, "norm_diff": 0.0977, "norm_loss": 0.0, "num_token_doc": 66.8107, "num_token_overlap": 11.6675, "num_token_query": 31.395, "num_token_union": 65.1944, "num_word_context": 202.5113, "num_word_doc": 49.8421, "num_word_query": 23.3092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10642.7795, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1685, "query_norm": 1.2876, "queue_k_norm": 1.3912, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.395, "sent_len_1": 66.8107, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.4888, "stdk": 0.048, "stdq": 0.0432, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 41600 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.7124, "doc_norm": 1.3944, "encoder_q-embeddings": 6086.1738, "encoder_q-layer.0": 4525.6826, "encoder_q-layer.1": 4891.2261, "encoder_q-layer.10": 7142.9458, "encoder_q-layer.11": 15494.1592, "encoder_q-layer.2": 5385.4204, "encoder_q-layer.3": 5469.4927, "encoder_q-layer.4": 5623.1816, "encoder_q-layer.5": 5417.4595, "encoder_q-layer.6": 6348.1914, "encoder_q-layer.7": 6611.1875, "encoder_q-layer.8": 7699.7158, "encoder_q-layer.9": 6991.3843, "epoch": 0.18, "inbatch_neg_score": 0.1636, "inbatch_pos_score": 0.7476, "learning_rate": 4.165789473684211e-05, "loss": 3.7124, "norm_diff": 0.1221, "norm_loss": 0.0, "num_token_doc": 66.7321, "num_token_overlap": 11.6559, "num_token_query": 31.3587, "num_token_union": 65.0877, "num_word_context": 202.6308, "num_word_doc": 49.7821, "num_word_query": 23.303, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10872.7791, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1639, "query_norm": 1.2723, "queue_k_norm": 1.3887, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3587, "sent_len_1": 66.7321, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.1662, "stdk": 0.0483, "stdq": 0.0428, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 41700 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.7137, "doc_norm": 1.3839, "encoder_q-embeddings": 5959.4019, "encoder_q-layer.0": 3839.6465, "encoder_q-layer.1": 3993.978, "encoder_q-layer.10": 6607.9653, "encoder_q-layer.11": 15626.1279, "encoder_q-layer.2": 4608.0244, "encoder_q-layer.3": 4918.8804, "encoder_q-layer.4": 5205.0322, "encoder_q-layer.5": 5323.1333, "encoder_q-layer.6": 5775.6841, "encoder_q-layer.7": 6198.2393, "encoder_q-layer.8": 7215.7744, "encoder_q-layer.9": 6456.1484, "epoch": 0.18, "inbatch_neg_score": 0.1623, "inbatch_pos_score": 0.7466, "learning_rate": 4.163157894736842e-05, "loss": 3.7137, "norm_diff": 0.1111, "norm_loss": 0.0, "num_token_doc": 66.8454, "num_token_overlap": 11.6636, "num_token_query": 31.2978, "num_token_union": 65.1296, "num_word_context": 202.2866, "num_word_doc": 49.8741, "num_word_query": 23.2427, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10305.3007, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1626, "query_norm": 1.2728, "queue_k_norm": 1.3888, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2978, "sent_len_1": 66.8454, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.9175, "stdk": 0.0479, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41800 }, { "accuracy": 48.1445, "active_queue_size": 16384.0, "cl_loss": 3.6802, "doc_norm": 1.3944, "encoder_q-embeddings": 21558.8086, "encoder_q-layer.0": 14415.4277, "encoder_q-layer.1": 14659.3994, "encoder_q-layer.10": 6417.9438, "encoder_q-layer.11": 15609.5195, "encoder_q-layer.2": 15142.1533, "encoder_q-layer.3": 14998.9385, "encoder_q-layer.4": 14385.7217, "encoder_q-layer.5": 13660.9023, "encoder_q-layer.6": 12117.7832, "encoder_q-layer.7": 9128.498, "encoder_q-layer.8": 8465.4043, "encoder_q-layer.9": 7071.5464, "epoch": 0.18, "inbatch_neg_score": 0.1596, "inbatch_pos_score": 0.73, "learning_rate": 4.160526315789474e-05, "loss": 3.6802, "norm_diff": 0.1223, "norm_loss": 0.0, "num_token_doc": 66.8381, "num_token_overlap": 11.72, "num_token_query": 31.4583, "num_token_union": 65.1704, "num_word_context": 202.2621, "num_word_doc": 49.8596, "num_word_query": 23.37, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21141.4, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1594, "query_norm": 1.2721, "queue_k_norm": 1.3895, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4583, "sent_len_1": 66.8381, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.89, "stdk": 0.0483, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 41900 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.6957, "doc_norm": 1.3869, "encoder_q-embeddings": 5835.5718, "encoder_q-layer.0": 3847.4927, "encoder_q-layer.1": 4038.3872, "encoder_q-layer.10": 7430.8584, "encoder_q-layer.11": 16512.8574, "encoder_q-layer.2": 4479.8618, "encoder_q-layer.3": 4616.7856, "encoder_q-layer.4": 4903.4658, "encoder_q-layer.5": 4969.6538, "encoder_q-layer.6": 5454.6743, "encoder_q-layer.7": 6034.208, "encoder_q-layer.8": 7684.46, "encoder_q-layer.9": 6711.8354, "epoch": 0.18, "inbatch_neg_score": 0.1581, "inbatch_pos_score": 0.7363, "learning_rate": 4.157894736842106e-05, "loss": 3.6957, "norm_diff": 0.1244, "norm_loss": 0.0, "num_token_doc": 66.6476, "num_token_overlap": 11.6985, "num_token_query": 31.3852, "num_token_union": 65.0166, "num_word_context": 202.2488, "num_word_doc": 49.7285, "num_word_query": 23.3107, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10517.1057, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.158, "query_norm": 1.2624, "queue_k_norm": 1.3899, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3852, "sent_len_1": 66.6476, "sent_len_max_0": 127.98, "sent_len_max_1": 209.035, "stdk": 0.048, "stdq": 0.0425, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42000 }, { "accuracy": 45.7031, "active_queue_size": 16384.0, "cl_loss": 3.7033, "doc_norm": 1.3899, "encoder_q-embeddings": 5962.0654, "encoder_q-layer.0": 3881.3374, "encoder_q-layer.1": 4028.2209, "encoder_q-layer.10": 6853.8262, "encoder_q-layer.11": 16153.3281, "encoder_q-layer.2": 4451.8711, "encoder_q-layer.3": 4660.8408, "encoder_q-layer.4": 5018.1455, "encoder_q-layer.5": 4936.0981, "encoder_q-layer.6": 5443.8018, "encoder_q-layer.7": 6229.5391, "encoder_q-layer.8": 7703.6455, "encoder_q-layer.9": 6895.7295, "epoch": 0.18, "inbatch_neg_score": 0.1516, "inbatch_pos_score": 0.7144, "learning_rate": 4.155263157894737e-05, "loss": 3.7033, "norm_diff": 0.1245, "norm_loss": 0.0, "num_token_doc": 66.8251, "num_token_overlap": 11.6805, "num_token_query": 31.3454, "num_token_union": 65.1347, "num_word_context": 202.382, "num_word_doc": 49.8734, "num_word_query": 23.2759, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10635.359, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1519, "query_norm": 1.2654, "queue_k_norm": 1.3911, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3454, "sent_len_1": 66.8251, "sent_len_max_0": 127.9625, "sent_len_max_1": 210.2063, "stdk": 0.0481, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42100 }, { "accuracy": 46.4355, "active_queue_size": 16384.0, "cl_loss": 3.6965, "doc_norm": 1.3895, "encoder_q-embeddings": 5543.7544, "encoder_q-layer.0": 3664.5295, "encoder_q-layer.1": 3767.2078, "encoder_q-layer.10": 7012.2095, "encoder_q-layer.11": 15764.4004, "encoder_q-layer.2": 4155.0864, "encoder_q-layer.3": 4275.8823, "encoder_q-layer.4": 4625.0718, "encoder_q-layer.5": 4679.04, "encoder_q-layer.6": 5268.8164, "encoder_q-layer.7": 6054.0342, "encoder_q-layer.8": 7483.6543, "encoder_q-layer.9": 6653.4473, "epoch": 0.18, "inbatch_neg_score": 0.1514, "inbatch_pos_score": 0.7021, "learning_rate": 4.152631578947369e-05, "loss": 3.6965, "norm_diff": 0.1382, "norm_loss": 0.0, "num_token_doc": 66.921, "num_token_overlap": 11.6556, "num_token_query": 31.3248, "num_token_union": 65.1918, "num_word_context": 202.5414, "num_word_doc": 49.9272, "num_word_query": 23.2572, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10258.1248, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.152, "query_norm": 1.2513, "queue_k_norm": 1.3904, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3248, "sent_len_1": 66.921, "sent_len_max_0": 127.98, "sent_len_max_1": 209.7862, "stdk": 0.0481, "stdq": 0.0423, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42200 }, { "accuracy": 46.4355, "active_queue_size": 16384.0, "cl_loss": 3.705, "doc_norm": 1.3826, "encoder_q-embeddings": 5536.0923, "encoder_q-layer.0": 3572.3801, "encoder_q-layer.1": 3715.3274, "encoder_q-layer.10": 6397.9634, "encoder_q-layer.11": 15963.0928, "encoder_q-layer.2": 4109.7139, "encoder_q-layer.3": 4268.895, "encoder_q-layer.4": 4543.1035, "encoder_q-layer.5": 4550.0742, "encoder_q-layer.6": 5234.6196, "encoder_q-layer.7": 5639.4429, "encoder_q-layer.8": 6853.1021, "encoder_q-layer.9": 6067.856, "epoch": 0.18, "inbatch_neg_score": 0.1529, "inbatch_pos_score": 0.7148, "learning_rate": 4.15e-05, "loss": 3.705, "norm_diff": 0.1291, "norm_loss": 0.0, "num_token_doc": 66.7476, "num_token_overlap": 11.6244, "num_token_query": 31.2852, "num_token_union": 65.0722, "num_word_context": 202.1891, "num_word_doc": 49.8075, "num_word_query": 23.2343, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10125.189, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1528, "query_norm": 1.2536, "queue_k_norm": 1.3881, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2852, "sent_len_1": 66.7476, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.4137, "stdk": 0.0479, "stdq": 0.0423, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 42300 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.6915, "doc_norm": 1.3868, "encoder_q-embeddings": 5737.6909, "encoder_q-layer.0": 3778.115, "encoder_q-layer.1": 4010.8923, "encoder_q-layer.10": 6904.4917, "encoder_q-layer.11": 15379.4658, "encoder_q-layer.2": 4591.6982, "encoder_q-layer.3": 4897.1812, "encoder_q-layer.4": 5193.0112, "encoder_q-layer.5": 5337.4409, "encoder_q-layer.6": 5772.4082, "encoder_q-layer.7": 6784.4438, "encoder_q-layer.8": 7744.3721, "encoder_q-layer.9": 6670.332, "epoch": 0.18, "inbatch_neg_score": 0.1485, "inbatch_pos_score": 0.7246, "learning_rate": 4.147368421052632e-05, "loss": 3.6915, "norm_diff": 0.1124, "norm_loss": 0.0, "num_token_doc": 66.8163, "num_token_overlap": 11.687, "num_token_query": 31.2807, "num_token_union": 65.0971, "num_word_context": 202.0967, "num_word_doc": 49.8384, "num_word_query": 23.2112, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10260.5159, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1489, "query_norm": 1.2744, "queue_k_norm": 1.391, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2807, "sent_len_1": 66.8163, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.9837, "stdk": 0.0481, "stdq": 0.0432, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42400 }, { "accuracy": 47.3145, "active_queue_size": 16384.0, "cl_loss": 3.7046, "doc_norm": 1.3879, "encoder_q-embeddings": 5593.5718, "encoder_q-layer.0": 3627.4507, "encoder_q-layer.1": 3747.5701, "encoder_q-layer.10": 7323.4243, "encoder_q-layer.11": 15635.6025, "encoder_q-layer.2": 4025.2603, "encoder_q-layer.3": 4068.54, "encoder_q-layer.4": 4197.332, "encoder_q-layer.5": 4348.0674, "encoder_q-layer.6": 4807.8472, "encoder_q-layer.7": 5570.625, "encoder_q-layer.8": 7062.0371, "encoder_q-layer.9": 6159.4243, "epoch": 0.18, "inbatch_neg_score": 0.1481, "inbatch_pos_score": 0.7036, "learning_rate": 4.1447368421052636e-05, "loss": 3.7046, "norm_diff": 0.1286, "norm_loss": 0.0, "num_token_doc": 66.7376, "num_token_overlap": 11.6627, "num_token_query": 31.2539, "num_token_union": 65.0634, "num_word_context": 202.3652, "num_word_doc": 49.7941, "num_word_query": 23.204, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9912.3531, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1482, "query_norm": 1.2593, "queue_k_norm": 1.3885, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2539, "sent_len_1": 66.7376, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.1838, "stdk": 0.0481, "stdq": 0.0425, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42500 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.6871, "doc_norm": 1.3964, "encoder_q-embeddings": 5546.6821, "encoder_q-layer.0": 3683.5156, "encoder_q-layer.1": 3726.144, "encoder_q-layer.10": 7630.9321, "encoder_q-layer.11": 16676.9141, "encoder_q-layer.2": 4203.6812, "encoder_q-layer.3": 4545.8076, "encoder_q-layer.4": 4662.1167, "encoder_q-layer.5": 4666.0117, "encoder_q-layer.6": 5240.5742, "encoder_q-layer.7": 6038.354, "encoder_q-layer.8": 7977.4688, "encoder_q-layer.9": 6811.6924, "epoch": 0.18, "inbatch_neg_score": 0.1467, "inbatch_pos_score": 0.7153, "learning_rate": 4.142105263157895e-05, "loss": 3.6871, "norm_diff": 0.122, "norm_loss": 0.0, "num_token_doc": 66.8551, "num_token_overlap": 11.6487, "num_token_query": 31.2533, "num_token_union": 65.1253, "num_word_context": 202.2915, "num_word_doc": 49.8823, "num_word_query": 23.1919, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10356.4523, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1472, "query_norm": 1.2743, "queue_k_norm": 1.3887, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2533, "sent_len_1": 66.8551, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.2075, "stdk": 0.0484, "stdq": 0.043, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42600 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.6973, "doc_norm": 1.3832, "encoder_q-embeddings": 5452.4155, "encoder_q-layer.0": 3617.7053, "encoder_q-layer.1": 3823.0698, "encoder_q-layer.10": 7120.4556, "encoder_q-layer.11": 15698.1328, "encoder_q-layer.2": 4471.3999, "encoder_q-layer.3": 4642.7588, "encoder_q-layer.4": 4997.8867, "encoder_q-layer.5": 4813.3306, "encoder_q-layer.6": 5489.4873, "encoder_q-layer.7": 6050.7432, "encoder_q-layer.8": 7217.6226, "encoder_q-layer.9": 6454.3853, "epoch": 0.19, "inbatch_neg_score": 0.151, "inbatch_pos_score": 0.7153, "learning_rate": 4.1394736842105266e-05, "loss": 3.6973, "norm_diff": 0.1061, "norm_loss": 0.0, "num_token_doc": 66.8465, "num_token_overlap": 11.6801, "num_token_query": 31.3226, "num_token_union": 65.1149, "num_word_context": 202.3315, "num_word_doc": 49.8852, "num_word_query": 23.2491, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10155.1664, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.151, "query_norm": 1.277, "queue_k_norm": 1.3907, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3226, "sent_len_1": 66.8465, "sent_len_max_0": 127.98, "sent_len_max_1": 209.445, "stdk": 0.0479, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42700 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.6949, "doc_norm": 1.3868, "encoder_q-embeddings": 5702.0166, "encoder_q-layer.0": 3761.6084, "encoder_q-layer.1": 3997.3889, "encoder_q-layer.10": 6694.9438, "encoder_q-layer.11": 15319.6777, "encoder_q-layer.2": 4370.6362, "encoder_q-layer.3": 4582.2295, "encoder_q-layer.4": 4813.7695, "encoder_q-layer.5": 4984.6611, "encoder_q-layer.6": 5787.2788, "encoder_q-layer.7": 6266.8267, "encoder_q-layer.8": 7484.229, "encoder_q-layer.9": 6412.2275, "epoch": 0.19, "inbatch_neg_score": 0.1515, "inbatch_pos_score": 0.7222, "learning_rate": 4.136842105263158e-05, "loss": 3.6949, "norm_diff": 0.109, "norm_loss": 0.0, "num_token_doc": 66.9676, "num_token_overlap": 11.7051, "num_token_query": 31.4112, "num_token_union": 65.2392, "num_word_context": 202.5969, "num_word_doc": 49.9422, "num_word_query": 23.3217, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10172.4345, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1512, "query_norm": 1.2778, "queue_k_norm": 1.3883, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4112, "sent_len_1": 66.9676, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.5825, "stdk": 0.0481, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42800 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.6803, "doc_norm": 1.3893, "encoder_q-embeddings": 5493.2178, "encoder_q-layer.0": 3650.5789, "encoder_q-layer.1": 3825.302, "encoder_q-layer.10": 6737.9678, "encoder_q-layer.11": 15154.082, "encoder_q-layer.2": 4248.147, "encoder_q-layer.3": 4437.6504, "encoder_q-layer.4": 4764.7695, "encoder_q-layer.5": 4843.749, "encoder_q-layer.6": 5550.2275, "encoder_q-layer.7": 6025.1709, "encoder_q-layer.8": 7449.5068, "encoder_q-layer.9": 6400.0742, "epoch": 0.19, "inbatch_neg_score": 0.1534, "inbatch_pos_score": 0.7256, "learning_rate": 4.1342105263157896e-05, "loss": 3.6803, "norm_diff": 0.1009, "norm_loss": 0.0, "num_token_doc": 66.7217, "num_token_overlap": 11.6876, "num_token_query": 31.3258, "num_token_union": 65.0716, "num_word_context": 202.4376, "num_word_doc": 49.812, "num_word_query": 23.2599, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10032.9306, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1541, "query_norm": 1.2885, "queue_k_norm": 1.3903, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3258, "sent_len_1": 66.7217, "sent_len_max_0": 127.9575, "sent_len_max_1": 207.4062, "stdk": 0.0482, "stdq": 0.0428, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 42900 }, { "accuracy": 45.8984, "active_queue_size": 16384.0, "cl_loss": 3.6811, "doc_norm": 1.3862, "encoder_q-embeddings": 5553.5869, "encoder_q-layer.0": 3658.6807, "encoder_q-layer.1": 3893.1011, "encoder_q-layer.10": 7072.3936, "encoder_q-layer.11": 16007.9492, "encoder_q-layer.2": 4295.3638, "encoder_q-layer.3": 4565.7017, "encoder_q-layer.4": 4872.127, "encoder_q-layer.5": 4859.3579, "encoder_q-layer.6": 5482.0732, "encoder_q-layer.7": 6010.8433, "encoder_q-layer.8": 7068.4092, "encoder_q-layer.9": 6423.5454, "epoch": 0.19, "inbatch_neg_score": 0.1581, "inbatch_pos_score": 0.7266, "learning_rate": 4.1315789473684214e-05, "loss": 3.6811, "norm_diff": 0.0944, "norm_loss": 0.0, "num_token_doc": 66.8325, "num_token_overlap": 11.7035, "num_token_query": 31.3836, "num_token_union": 65.1346, "num_word_context": 202.2434, "num_word_doc": 49.8632, "num_word_query": 23.3067, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10115.5807, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1578, "query_norm": 1.2917, "queue_k_norm": 1.3894, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3836, "sent_len_1": 66.8325, "sent_len_max_0": 127.9475, "sent_len_max_1": 209.2075, "stdk": 0.048, "stdq": 0.0427, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43000 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.6922, "doc_norm": 1.3882, "encoder_q-embeddings": 5744.0405, "encoder_q-layer.0": 3786.2786, "encoder_q-layer.1": 3914.3713, "encoder_q-layer.10": 6797.9141, "encoder_q-layer.11": 16130.5166, "encoder_q-layer.2": 4485.1108, "encoder_q-layer.3": 4606.2866, "encoder_q-layer.4": 4960.2251, "encoder_q-layer.5": 5002.0776, "encoder_q-layer.6": 5506.1216, "encoder_q-layer.7": 6236.5063, "encoder_q-layer.8": 7420.8789, "encoder_q-layer.9": 6880.2666, "epoch": 0.19, "inbatch_neg_score": 0.1619, "inbatch_pos_score": 0.7271, "learning_rate": 4.1289473684210526e-05, "loss": 3.6922, "norm_diff": 0.065, "norm_loss": 0.0, "num_token_doc": 66.6537, "num_token_overlap": 11.6631, "num_token_query": 31.3148, "num_token_union": 65.0637, "num_word_context": 202.1867, "num_word_doc": 49.7382, "num_word_query": 23.2478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10335.9217, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1616, "query_norm": 1.3232, "queue_k_norm": 1.3896, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3148, "sent_len_1": 66.6537, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.56, "stdk": 0.0481, "stdq": 0.0432, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43100 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6695, "doc_norm": 1.3916, "encoder_q-embeddings": 5631.1641, "encoder_q-layer.0": 3782.6318, "encoder_q-layer.1": 4139.1055, "encoder_q-layer.10": 7526.4951, "encoder_q-layer.11": 15263.1387, "encoder_q-layer.2": 4585.3008, "encoder_q-layer.3": 4873.0884, "encoder_q-layer.4": 5205.1514, "encoder_q-layer.5": 5465.2158, "encoder_q-layer.6": 5820.5688, "encoder_q-layer.7": 6743.6509, "encoder_q-layer.8": 7536.3188, "encoder_q-layer.9": 6515.3232, "epoch": 0.19, "inbatch_neg_score": 0.1738, "inbatch_pos_score": 0.7549, "learning_rate": 4.1263157894736844e-05, "loss": 3.6695, "norm_diff": 0.0469, "norm_loss": 0.0, "num_token_doc": 66.7567, "num_token_overlap": 11.6947, "num_token_query": 31.4788, "num_token_union": 65.1485, "num_word_context": 202.2922, "num_word_doc": 49.8053, "num_word_query": 23.3755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10333.5701, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.173, "query_norm": 1.3448, "queue_k_norm": 1.3894, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4788, "sent_len_1": 66.7567, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.3288, "stdk": 0.0482, "stdq": 0.0434, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43200 }, { "accuracy": 46.0449, "active_queue_size": 16384.0, "cl_loss": 3.6853, "doc_norm": 1.3895, "encoder_q-embeddings": 12044.499, "encoder_q-layer.0": 7992.7373, "encoder_q-layer.1": 8558.3809, "encoder_q-layer.10": 14402.7422, "encoder_q-layer.11": 30573.9648, "encoder_q-layer.2": 9874.7979, "encoder_q-layer.3": 9947.6406, "encoder_q-layer.4": 10602.1387, "encoder_q-layer.5": 11088.4229, "encoder_q-layer.6": 11496.0264, "encoder_q-layer.7": 12321.2266, "encoder_q-layer.8": 14079.9365, "encoder_q-layer.9": 13006.3809, "epoch": 0.19, "inbatch_neg_score": 0.1779, "inbatch_pos_score": 0.7432, "learning_rate": 4.123684210526316e-05, "loss": 3.6853, "norm_diff": 0.0552, "norm_loss": 0.0, "num_token_doc": 66.8069, "num_token_overlap": 11.7107, "num_token_query": 31.4056, "num_token_union": 65.1103, "num_word_context": 202.3683, "num_word_doc": 49.8476, "num_word_query": 23.3348, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20812.3918, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.178, "query_norm": 1.3343, "queue_k_norm": 1.3915, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4056, "sent_len_1": 66.8069, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.7725, "stdk": 0.0481, "stdq": 0.043, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 43300 }, { "accuracy": 47.2168, "active_queue_size": 16384.0, "cl_loss": 3.6864, "doc_norm": 1.3955, "encoder_q-embeddings": 11237.1758, "encoder_q-layer.0": 7155.8999, "encoder_q-layer.1": 7663.4961, "encoder_q-layer.10": 13005.4053, "encoder_q-layer.11": 31174.3984, "encoder_q-layer.2": 8342.665, "encoder_q-layer.3": 8642.0518, "encoder_q-layer.4": 8955.1797, "encoder_q-layer.5": 9482.2432, "encoder_q-layer.6": 10401.6172, "encoder_q-layer.7": 11533.9004, "encoder_q-layer.8": 13420.334, "encoder_q-layer.9": 12245.9951, "epoch": 0.19, "inbatch_neg_score": 0.1858, "inbatch_pos_score": 0.7588, "learning_rate": 4.1210526315789474e-05, "loss": 3.6864, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.9222, "num_token_overlap": 11.6964, "num_token_query": 31.3886, "num_token_union": 65.1884, "num_word_context": 202.2136, "num_word_doc": 49.9019, "num_word_query": 23.3028, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19665.5575, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1859, "query_norm": 1.325, "queue_k_norm": 1.3914, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3886, "sent_len_1": 66.9222, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.2312, "stdk": 0.0483, "stdq": 0.0426, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43400 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 3.698, "doc_norm": 1.3867, "encoder_q-embeddings": 29585.0996, "encoder_q-layer.0": 21658.7754, "encoder_q-layer.1": 21816.6074, "encoder_q-layer.10": 14459.3311, "encoder_q-layer.11": 31202.1816, "encoder_q-layer.2": 27575.5742, "encoder_q-layer.3": 27804.2812, "encoder_q-layer.4": 27252.0605, "encoder_q-layer.5": 25306.4863, "encoder_q-layer.6": 29213.7188, "encoder_q-layer.7": 25044.6309, "encoder_q-layer.8": 16232.4951, "encoder_q-layer.9": 12819.8877, "epoch": 0.19, "inbatch_neg_score": 0.1932, "inbatch_pos_score": 0.7603, "learning_rate": 4.118421052631579e-05, "loss": 3.698, "norm_diff": 0.0678, "norm_loss": 0.0, "num_token_doc": 66.7256, "num_token_overlap": 11.6579, "num_token_query": 31.3076, "num_token_union": 65.0765, "num_word_context": 202.2287, "num_word_doc": 49.7968, "num_word_query": 23.2441, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37572.1448, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 0.1927, "query_norm": 1.319, "queue_k_norm": 1.3934, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3076, "sent_len_1": 66.7256, "sent_len_max_0": 127.975, "sent_len_max_1": 208.455, "stdk": 0.048, "stdq": 0.0432, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 43500 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.7001, "doc_norm": 1.3969, "encoder_q-embeddings": 11817.9971, "encoder_q-layer.0": 7448.4707, "encoder_q-layer.1": 7819.5088, "encoder_q-layer.10": 13237.6543, "encoder_q-layer.11": 31259.8047, "encoder_q-layer.2": 8665.0156, "encoder_q-layer.3": 9108.1973, "encoder_q-layer.4": 9832.877, "encoder_q-layer.5": 10206.4297, "encoder_q-layer.6": 11475.2324, "encoder_q-layer.7": 12232.3691, "encoder_q-layer.8": 14166.3438, "encoder_q-layer.9": 12877.2695, "epoch": 0.19, "inbatch_neg_score": 0.1876, "inbatch_pos_score": 0.7651, "learning_rate": 4.1157894736842104e-05, "loss": 3.7001, "norm_diff": 0.0902, "norm_loss": 0.0, "num_token_doc": 66.6419, "num_token_overlap": 11.6899, "num_token_query": 31.3719, "num_token_union": 65.0035, "num_word_context": 202.2125, "num_word_doc": 49.7193, "num_word_query": 23.3045, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20762.9534, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.188, "query_norm": 1.3067, "queue_k_norm": 1.3913, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3719, "sent_len_1": 66.6419, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.195, "stdk": 0.0483, "stdq": 0.0432, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 43600 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.686, "doc_norm": 1.3913, "encoder_q-embeddings": 10950.7725, "encoder_q-layer.0": 7160.7559, "encoder_q-layer.1": 7512.7612, "encoder_q-layer.10": 14834.5479, "encoder_q-layer.11": 34295.7695, "encoder_q-layer.2": 8292.0537, "encoder_q-layer.3": 8562.2539, "encoder_q-layer.4": 8914.9902, "encoder_q-layer.5": 9341.7227, "encoder_q-layer.6": 10852.9092, "encoder_q-layer.7": 12637.916, "encoder_q-layer.8": 14996.2598, "encoder_q-layer.9": 13892.6182, "epoch": 0.19, "inbatch_neg_score": 0.1915, "inbatch_pos_score": 0.7495, "learning_rate": 4.113157894736842e-05, "loss": 3.686, "norm_diff": 0.1061, "norm_loss": 0.0, "num_token_doc": 66.8294, "num_token_overlap": 11.6627, "num_token_query": 31.2849, "num_token_union": 65.1285, "num_word_context": 202.4369, "num_word_doc": 49.8587, "num_word_query": 23.2248, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20961.9361, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.1914, "query_norm": 1.2852, "queue_k_norm": 1.3932, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2849, "sent_len_1": 66.8294, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.6425, "stdk": 0.048, "stdq": 0.0426, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43700 }, { "accuracy": 44.5312, "active_queue_size": 16384.0, "cl_loss": 3.7022, "doc_norm": 1.3919, "encoder_q-embeddings": 5961.8218, "encoder_q-layer.0": 3806.7625, "encoder_q-layer.1": 4059.4897, "encoder_q-layer.10": 8022.2812, "encoder_q-layer.11": 17064.2305, "encoder_q-layer.2": 4548.978, "encoder_q-layer.3": 4680.5034, "encoder_q-layer.4": 5035.249, "encoder_q-layer.5": 5218.1348, "encoder_q-layer.6": 5881.4502, "encoder_q-layer.7": 6800.2314, "encoder_q-layer.8": 8029.083, "encoder_q-layer.9": 7721.791, "epoch": 0.19, "inbatch_neg_score": 0.1864, "inbatch_pos_score": 0.7456, "learning_rate": 4.110526315789474e-05, "loss": 3.7022, "norm_diff": 0.1, "norm_loss": 0.0, "num_token_doc": 66.6136, "num_token_overlap": 11.6612, "num_token_query": 31.3757, "num_token_union": 65.0189, "num_word_context": 201.8413, "num_word_doc": 49.6858, "num_word_query": 23.3063, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10939.7188, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1863, "query_norm": 1.2919, "queue_k_norm": 1.3945, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3757, "sent_len_1": 66.6136, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.965, "stdk": 0.048, "stdq": 0.0432, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 43800 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.6909, "doc_norm": 1.3995, "encoder_q-embeddings": 7210.4141, "encoder_q-layer.0": 5056.873, "encoder_q-layer.1": 5668.0986, "encoder_q-layer.10": 7341.5977, "encoder_q-layer.11": 16219.9873, "encoder_q-layer.2": 6849.1475, "encoder_q-layer.3": 6812.1509, "encoder_q-layer.4": 6847.3936, "encoder_q-layer.5": 6450.0693, "encoder_q-layer.6": 6792.6074, "encoder_q-layer.7": 6546.5688, "encoder_q-layer.8": 7659.3745, "encoder_q-layer.9": 6752.6821, "epoch": 0.19, "inbatch_neg_score": 0.1824, "inbatch_pos_score": 0.7422, "learning_rate": 4.107894736842106e-05, "loss": 3.6909, "norm_diff": 0.1366, "norm_loss": 0.0, "num_token_doc": 66.7996, "num_token_overlap": 11.6634, "num_token_query": 31.2948, "num_token_union": 65.1015, "num_word_context": 202.1318, "num_word_doc": 49.864, "num_word_query": 23.2465, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11612.5613, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1829, "query_norm": 1.2629, "queue_k_norm": 1.3932, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2948, "sent_len_1": 66.7996, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.6362, "stdk": 0.0483, "stdq": 0.0422, "stdqueue_k": 0.0481, "stdqueue_q": 0.0, "step": 43900 }, { "accuracy": 46.3867, "active_queue_size": 16384.0, "cl_loss": 3.6884, "doc_norm": 1.3939, "encoder_q-embeddings": 5468.7764, "encoder_q-layer.0": 3555.8416, "encoder_q-layer.1": 3835.033, "encoder_q-layer.10": 8549.4082, "encoder_q-layer.11": 16567.9414, "encoder_q-layer.2": 4371.1851, "encoder_q-layer.3": 4536.8716, "encoder_q-layer.4": 4830.4062, "encoder_q-layer.5": 4856.0786, "encoder_q-layer.6": 5644.7876, "encoder_q-layer.7": 6113.8779, "encoder_q-layer.8": 8278.0449, "encoder_q-layer.9": 7544.1357, "epoch": 0.19, "inbatch_neg_score": 0.1783, "inbatch_pos_score": 0.7471, "learning_rate": 4.105263157894737e-05, "loss": 3.6884, "norm_diff": 0.0992, "norm_loss": 0.0, "num_token_doc": 66.7166, "num_token_overlap": 11.6398, "num_token_query": 31.3397, "num_token_union": 65.1022, "num_word_context": 202.2643, "num_word_doc": 49.7631, "num_word_query": 23.2591, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10518.9035, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1781, "query_norm": 1.2947, "queue_k_norm": 1.3973, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3397, "sent_len_1": 66.7166, "sent_len_max_0": 127.985, "sent_len_max_1": 208.8075, "stdk": 0.0481, "stdq": 0.0436, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44000 }, { "accuracy": 47.7539, "active_queue_size": 16384.0, "cl_loss": 3.691, "doc_norm": 1.3923, "encoder_q-embeddings": 5628.5181, "encoder_q-layer.0": 3797.4136, "encoder_q-layer.1": 3807.0195, "encoder_q-layer.10": 6498.8296, "encoder_q-layer.11": 15463.6025, "encoder_q-layer.2": 4407.8486, "encoder_q-layer.3": 4421.1064, "encoder_q-layer.4": 4645.2012, "encoder_q-layer.5": 4719.5425, "encoder_q-layer.6": 5067.8608, "encoder_q-layer.7": 5613.9482, "encoder_q-layer.8": 6826.4727, "encoder_q-layer.9": 6407.5195, "epoch": 0.19, "inbatch_neg_score": 0.1765, "inbatch_pos_score": 0.748, "learning_rate": 4.102631578947368e-05, "loss": 3.691, "norm_diff": 0.1035, "norm_loss": 0.0, "num_token_doc": 66.7359, "num_token_overlap": 11.7001, "num_token_query": 31.3995, "num_token_union": 65.104, "num_word_context": 202.2139, "num_word_doc": 49.8051, "num_word_query": 23.3375, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9891.653, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1774, "query_norm": 1.2888, "queue_k_norm": 1.3978, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3995, "sent_len_1": 66.7359, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.89, "stdk": 0.048, "stdq": 0.0433, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44100 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.6866, "doc_norm": 1.3999, "encoder_q-embeddings": 5570.6455, "encoder_q-layer.0": 3640.1228, "encoder_q-layer.1": 3903.3174, "encoder_q-layer.10": 6575.7852, "encoder_q-layer.11": 15333.585, "encoder_q-layer.2": 4422.9253, "encoder_q-layer.3": 4568.5918, "encoder_q-layer.4": 4696.8828, "encoder_q-layer.5": 4747.5825, "encoder_q-layer.6": 5195.5088, "encoder_q-layer.7": 6160.8599, "encoder_q-layer.8": 7051.6514, "encoder_q-layer.9": 6351.1987, "epoch": 0.19, "inbatch_neg_score": 0.1739, "inbatch_pos_score": 0.7603, "learning_rate": 4.1e-05, "loss": 3.6866, "norm_diff": 0.1093, "norm_loss": 0.0, "num_token_doc": 66.8442, "num_token_overlap": 11.6599, "num_token_query": 31.3026, "num_token_union": 65.1399, "num_word_context": 202.3824, "num_word_doc": 49.9111, "num_word_query": 23.2344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9927.4847, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.174, "query_norm": 1.2906, "queue_k_norm": 1.3968, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3026, "sent_len_1": 66.8442, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.9175, "stdk": 0.0483, "stdq": 0.0435, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 44200 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.6742, "doc_norm": 1.3997, "encoder_q-embeddings": 5397.5776, "encoder_q-layer.0": 3743.6096, "encoder_q-layer.1": 3877.9485, "encoder_q-layer.10": 6662.126, "encoder_q-layer.11": 15439.7734, "encoder_q-layer.2": 4342.2065, "encoder_q-layer.3": 4478.8213, "encoder_q-layer.4": 4743.6909, "encoder_q-layer.5": 4812.0771, "encoder_q-layer.6": 5251.9575, "encoder_q-layer.7": 6044.584, "encoder_q-layer.8": 7282.8477, "encoder_q-layer.9": 6431.4424, "epoch": 0.19, "inbatch_neg_score": 0.1826, "inbatch_pos_score": 0.75, "learning_rate": 4.097368421052632e-05, "loss": 3.6742, "norm_diff": 0.1311, "norm_loss": 0.0, "num_token_doc": 66.8439, "num_token_overlap": 11.6742, "num_token_query": 31.3333, "num_token_union": 65.1656, "num_word_context": 202.2742, "num_word_doc": 49.8974, "num_word_query": 23.2752, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9888.0253, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1825, "query_norm": 1.2686, "queue_k_norm": 1.3988, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3333, "sent_len_1": 66.8439, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.4575, "stdk": 0.0483, "stdq": 0.0423, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44300 }, { "accuracy": 46.7773, "active_queue_size": 16384.0, "cl_loss": 3.6832, "doc_norm": 1.3998, "encoder_q-embeddings": 5805.7749, "encoder_q-layer.0": 3747.9351, "encoder_q-layer.1": 4023.9761, "encoder_q-layer.10": 6679.6724, "encoder_q-layer.11": 15987.4219, "encoder_q-layer.2": 4567.1401, "encoder_q-layer.3": 4858.814, "encoder_q-layer.4": 5297.0591, "encoder_q-layer.5": 5389.1792, "encoder_q-layer.6": 5962.1426, "encoder_q-layer.7": 6704.2803, "encoder_q-layer.8": 7773.3906, "encoder_q-layer.9": 6835.5874, "epoch": 0.19, "inbatch_neg_score": 0.1778, "inbatch_pos_score": 0.7378, "learning_rate": 4.094736842105264e-05, "loss": 3.6832, "norm_diff": 0.1298, "norm_loss": 0.0, "num_token_doc": 66.8798, "num_token_overlap": 11.6477, "num_token_query": 31.2408, "num_token_union": 65.1221, "num_word_context": 202.4749, "num_word_doc": 49.8945, "num_word_query": 23.2009, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10753.6507, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1783, "query_norm": 1.27, "queue_k_norm": 1.3953, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2408, "sent_len_1": 66.8798, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.2138, "stdk": 0.0483, "stdq": 0.0425, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 44400 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.6868, "doc_norm": 1.3943, "encoder_q-embeddings": 5597.2915, "encoder_q-layer.0": 3720.5503, "encoder_q-layer.1": 4065.7046, "encoder_q-layer.10": 6660.3462, "encoder_q-layer.11": 15005.9209, "encoder_q-layer.2": 4671.5576, "encoder_q-layer.3": 4911.3765, "encoder_q-layer.4": 5454.7588, "encoder_q-layer.5": 5694.2241, "encoder_q-layer.6": 6205.9619, "encoder_q-layer.7": 6377.1963, "encoder_q-layer.8": 7078.646, "encoder_q-layer.9": 6329.6387, "epoch": 0.19, "inbatch_neg_score": 0.1761, "inbatch_pos_score": 0.7578, "learning_rate": 4.092105263157895e-05, "loss": 3.6868, "norm_diff": 0.1205, "norm_loss": 0.0, "num_token_doc": 66.735, "num_token_overlap": 11.6575, "num_token_query": 31.3201, "num_token_union": 65.0627, "num_word_context": 202.1988, "num_word_doc": 49.7667, "num_word_query": 23.2502, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10105.1129, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1763, "query_norm": 1.2738, "queue_k_norm": 1.3986, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3201, "sent_len_1": 66.735, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.965, "stdk": 0.0481, "stdq": 0.0428, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44500 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.679, "doc_norm": 1.3972, "encoder_q-embeddings": 6242.2808, "encoder_q-layer.0": 4236.25, "encoder_q-layer.1": 4696.022, "encoder_q-layer.10": 7584.293, "encoder_q-layer.11": 16009.1084, "encoder_q-layer.2": 5262.2402, "encoder_q-layer.3": 5449.9204, "encoder_q-layer.4": 5701.3696, "encoder_q-layer.5": 5591.9355, "encoder_q-layer.6": 6140.6338, "encoder_q-layer.7": 6767.4453, "encoder_q-layer.8": 8112.9453, "encoder_q-layer.9": 7176.6494, "epoch": 0.19, "inbatch_neg_score": 0.1776, "inbatch_pos_score": 0.7686, "learning_rate": 4.089473684210526e-05, "loss": 3.679, "norm_diff": 0.0858, "norm_loss": 0.0, "num_token_doc": 66.7088, "num_token_overlap": 11.6869, "num_token_query": 31.4234, "num_token_union": 65.0981, "num_word_context": 202.0633, "num_word_doc": 49.7628, "num_word_query": 23.342, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10978.7785, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1783, "query_norm": 1.3113, "queue_k_norm": 1.3984, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4234, "sent_len_1": 66.7088, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.2488, "stdk": 0.0482, "stdq": 0.0442, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44600 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.6736, "doc_norm": 1.3978, "encoder_q-embeddings": 6239.9004, "encoder_q-layer.0": 3972.1792, "encoder_q-layer.1": 4100.5903, "encoder_q-layer.10": 6845.6816, "encoder_q-layer.11": 15320.4492, "encoder_q-layer.2": 4530.8882, "encoder_q-layer.3": 4764.8877, "encoder_q-layer.4": 5018.1484, "encoder_q-layer.5": 4952.9922, "encoder_q-layer.6": 5505.5205, "encoder_q-layer.7": 6156.4238, "encoder_q-layer.8": 7372.1143, "encoder_q-layer.9": 6366.354, "epoch": 0.19, "inbatch_neg_score": 0.1754, "inbatch_pos_score": 0.748, "learning_rate": 4.0868421052631585e-05, "loss": 3.6736, "norm_diff": 0.1335, "norm_loss": 0.0, "num_token_doc": 66.8815, "num_token_overlap": 11.7104, "num_token_query": 31.3906, "num_token_union": 65.1801, "num_word_context": 202.4347, "num_word_doc": 49.9098, "num_word_query": 23.3033, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10356.0976, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1758, "query_norm": 1.2643, "queue_k_norm": 1.3991, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3906, "sent_len_1": 66.8815, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.1625, "stdk": 0.0482, "stdq": 0.0425, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44700 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.6836, "doc_norm": 1.4028, "encoder_q-embeddings": 6730.5605, "encoder_q-layer.0": 4881.6519, "encoder_q-layer.1": 5263.9736, "encoder_q-layer.10": 7093.6929, "encoder_q-layer.11": 16250.625, "encoder_q-layer.2": 5971.4746, "encoder_q-layer.3": 5976.9478, "encoder_q-layer.4": 6334.0693, "encoder_q-layer.5": 5898.123, "encoder_q-layer.6": 6238.5542, "encoder_q-layer.7": 6805.8633, "encoder_q-layer.8": 7794.9468, "encoder_q-layer.9": 6776.4937, "epoch": 0.19, "inbatch_neg_score": 0.1753, "inbatch_pos_score": 0.7471, "learning_rate": 4.08421052631579e-05, "loss": 3.6836, "norm_diff": 0.1394, "norm_loss": 0.0, "num_token_doc": 66.9318, "num_token_overlap": 11.6656, "num_token_query": 31.3351, "num_token_union": 65.2242, "num_word_context": 202.371, "num_word_doc": 49.957, "num_word_query": 23.2703, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11300.1829, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1754, "query_norm": 1.2634, "queue_k_norm": 1.3986, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3351, "sent_len_1": 66.9318, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.0188, "stdk": 0.0484, "stdq": 0.0424, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44800 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.6713, "doc_norm": 1.4001, "encoder_q-embeddings": 6160.2671, "encoder_q-layer.0": 4056.2446, "encoder_q-layer.1": 4371.0146, "encoder_q-layer.10": 7208.6055, "encoder_q-layer.11": 16522.7168, "encoder_q-layer.2": 5031.4258, "encoder_q-layer.3": 5246.6748, "encoder_q-layer.4": 5736.7407, "encoder_q-layer.5": 6063.0645, "encoder_q-layer.6": 6372.1782, "encoder_q-layer.7": 7148.3916, "encoder_q-layer.8": 8344.7881, "encoder_q-layer.9": 7174.2744, "epoch": 0.19, "inbatch_neg_score": 0.1728, "inbatch_pos_score": 0.7441, "learning_rate": 4.0815789473684215e-05, "loss": 3.6713, "norm_diff": 0.1405, "norm_loss": 0.0, "num_token_doc": 66.9204, "num_token_overlap": 11.6928, "num_token_query": 31.4229, "num_token_union": 65.261, "num_word_context": 202.2431, "num_word_doc": 49.9377, "num_word_query": 23.352, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11064.413, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1727, "query_norm": 1.2596, "queue_k_norm": 1.3993, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4229, "sent_len_1": 66.9204, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.6062, "stdk": 0.0483, "stdq": 0.0423, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 44900 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6753, "doc_norm": 1.4007, "encoder_q-embeddings": 5971.2983, "encoder_q-layer.0": 3901.2939, "encoder_q-layer.1": 4069.0486, "encoder_q-layer.10": 6289.1675, "encoder_q-layer.11": 14494.3105, "encoder_q-layer.2": 4475.8677, "encoder_q-layer.3": 4501.6758, "encoder_q-layer.4": 4907.874, "encoder_q-layer.5": 4830.7173, "encoder_q-layer.6": 5193.5493, "encoder_q-layer.7": 5754.0146, "encoder_q-layer.8": 6806.1074, "encoder_q-layer.9": 6236.1299, "epoch": 0.2, "inbatch_neg_score": 0.1734, "inbatch_pos_score": 0.7578, "learning_rate": 4.078947368421053e-05, "loss": 3.6753, "norm_diff": 0.1227, "norm_loss": 0.0, "num_token_doc": 66.7231, "num_token_overlap": 11.6948, "num_token_query": 31.3986, "num_token_union": 65.0854, "num_word_context": 202.3662, "num_word_doc": 49.8034, "num_word_query": 23.3161, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9761.8885, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1733, "query_norm": 1.278, "queue_k_norm": 1.3979, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3986, "sent_len_1": 66.7231, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.1475, "stdk": 0.0483, "stdq": 0.043, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 45000 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.6842, "doc_norm": 1.4005, "encoder_q-embeddings": 26172.1465, "encoder_q-layer.0": 18746.8008, "encoder_q-layer.1": 22196.5801, "encoder_q-layer.10": 6620.1763, "encoder_q-layer.11": 14998.5557, "encoder_q-layer.2": 27239.1738, "encoder_q-layer.3": 31034.4258, "encoder_q-layer.4": 34896.7031, "encoder_q-layer.5": 37111.2461, "encoder_q-layer.6": 34412.9062, "encoder_q-layer.7": 25390.3652, "encoder_q-layer.8": 11180.2256, "encoder_q-layer.9": 7794.5879, "epoch": 0.2, "inbatch_neg_score": 0.174, "inbatch_pos_score": 0.7397, "learning_rate": 4.076315789473684e-05, "loss": 3.6842, "norm_diff": 0.1291, "norm_loss": 0.0, "num_token_doc": 66.6955, "num_token_overlap": 11.6628, "num_token_query": 31.354, "num_token_union": 65.0768, "num_word_context": 202.2696, "num_word_doc": 49.7859, "num_word_query": 23.291, "postclip_grad_norm": 1.0, "preclip_grad_norm": 36359.5917, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 0.1738, "query_norm": 1.2714, "queue_k_norm": 1.4001, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.354, "sent_len_1": 66.6955, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.0037, "stdk": 0.0483, "stdq": 0.0426, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 45100 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.6806, "doc_norm": 1.3962, "encoder_q-embeddings": 5676.022, "encoder_q-layer.0": 3618.2437, "encoder_q-layer.1": 3788.8206, "encoder_q-layer.10": 7022.5781, "encoder_q-layer.11": 15824.0391, "encoder_q-layer.2": 3995.5659, "encoder_q-layer.3": 4199.6367, "encoder_q-layer.4": 4490.3071, "encoder_q-layer.5": 4549.272, "encoder_q-layer.6": 5193.6401, "encoder_q-layer.7": 5974.9155, "encoder_q-layer.8": 7623.9409, "encoder_q-layer.9": 6911.9697, "epoch": 0.2, "inbatch_neg_score": 0.174, "inbatch_pos_score": 0.7588, "learning_rate": 4.0736842105263164e-05, "loss": 3.6806, "norm_diff": 0.104, "norm_loss": 0.0, "num_token_doc": 66.7259, "num_token_overlap": 11.706, "num_token_query": 31.384, "num_token_union": 65.0568, "num_word_context": 202.1613, "num_word_doc": 49.7871, "num_word_query": 23.2915, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10103.8524, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1747, "query_norm": 1.2922, "queue_k_norm": 1.3966, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.384, "sent_len_1": 66.7259, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.8775, "stdk": 0.0481, "stdq": 0.0435, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 45200 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.6684, "doc_norm": 1.3975, "encoder_q-embeddings": 5431.6133, "encoder_q-layer.0": 3612.136, "encoder_q-layer.1": 3838.1443, "encoder_q-layer.10": 6548.3516, "encoder_q-layer.11": 15186.5811, "encoder_q-layer.2": 4372.9019, "encoder_q-layer.3": 4430.5146, "encoder_q-layer.4": 4685.1958, "encoder_q-layer.5": 4586.8394, "encoder_q-layer.6": 5298.0547, "encoder_q-layer.7": 5677.6201, "encoder_q-layer.8": 6711.4541, "encoder_q-layer.9": 6059.3335, "epoch": 0.2, "inbatch_neg_score": 0.1782, "inbatch_pos_score": 0.7578, "learning_rate": 4.0710526315789475e-05, "loss": 3.6684, "norm_diff": 0.1131, "norm_loss": 0.0, "num_token_doc": 66.6331, "num_token_overlap": 11.6495, "num_token_query": 31.3605, "num_token_union": 65.0296, "num_word_context": 202.3223, "num_word_doc": 49.7334, "num_word_query": 23.2893, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9851.9155, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1791, "query_norm": 1.2845, "queue_k_norm": 1.3981, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3605, "sent_len_1": 66.6331, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.8638, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 45300 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.6652, "doc_norm": 1.398, "encoder_q-embeddings": 5215.5146, "encoder_q-layer.0": 3569.8984, "encoder_q-layer.1": 3647.6292, "encoder_q-layer.10": 6430.998, "encoder_q-layer.11": 15290.0244, "encoder_q-layer.2": 4132.8208, "encoder_q-layer.3": 4401.9604, "encoder_q-layer.4": 4558.8145, "encoder_q-layer.5": 4613.3213, "encoder_q-layer.6": 5301.3086, "encoder_q-layer.7": 5852.7251, "encoder_q-layer.8": 6603.2295, "encoder_q-layer.9": 6177.5093, "epoch": 0.2, "inbatch_neg_score": 0.178, "inbatch_pos_score": 0.7456, "learning_rate": 4.0684210526315794e-05, "loss": 3.6652, "norm_diff": 0.1124, "norm_loss": 0.0, "num_token_doc": 66.9355, "num_token_overlap": 11.6837, "num_token_query": 31.35, "num_token_union": 65.155, "num_word_context": 202.4471, "num_word_doc": 49.9378, "num_word_query": 23.2657, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9628.3705, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1785, "query_norm": 1.2856, "queue_k_norm": 1.3983, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.35, "sent_len_1": 66.9355, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.7637, "stdk": 0.0482, "stdq": 0.0426, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 45400 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6631, "doc_norm": 1.401, "encoder_q-embeddings": 5540.3359, "encoder_q-layer.0": 3756.7822, "encoder_q-layer.1": 3960.1753, "encoder_q-layer.10": 6385.1973, "encoder_q-layer.11": 14821.0137, "encoder_q-layer.2": 4338.3179, "encoder_q-layer.3": 4438.8457, "encoder_q-layer.4": 4898.6104, "encoder_q-layer.5": 4962.8379, "encoder_q-layer.6": 5286.0068, "encoder_q-layer.7": 6156.4819, "encoder_q-layer.8": 6722.8516, "encoder_q-layer.9": 6163.7612, "epoch": 0.2, "inbatch_neg_score": 0.1817, "inbatch_pos_score": 0.7617, "learning_rate": 4.0657894736842105e-05, "loss": 3.6631, "norm_diff": 0.0983, "norm_loss": 0.0, "num_token_doc": 66.6193, "num_token_overlap": 11.6976, "num_token_query": 31.3818, "num_token_union": 64.9926, "num_word_context": 201.9813, "num_word_doc": 49.7035, "num_word_query": 23.3132, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9863.4603, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1809, "query_norm": 1.3027, "queue_k_norm": 1.3984, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3818, "sent_len_1": 66.6193, "sent_len_max_0": 127.985, "sent_len_max_1": 208.1775, "stdk": 0.0483, "stdq": 0.043, "stdqueue_k": 0.0482, "stdqueue_q": 0.0, "step": 45500 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6606, "doc_norm": 1.4002, "encoder_q-embeddings": 6389.7305, "encoder_q-layer.0": 4337.5293, "encoder_q-layer.1": 4638.9663, "encoder_q-layer.10": 7016.229, "encoder_q-layer.11": 15804.2842, "encoder_q-layer.2": 5059.3877, "encoder_q-layer.3": 5609.9185, "encoder_q-layer.4": 5797.5923, "encoder_q-layer.5": 5798.8813, "encoder_q-layer.6": 6168.1733, "encoder_q-layer.7": 6436.7593, "encoder_q-layer.8": 7725.5981, "encoder_q-layer.9": 6849.251, "epoch": 0.2, "inbatch_neg_score": 0.1846, "inbatch_pos_score": 0.7695, "learning_rate": 4.0631578947368424e-05, "loss": 3.6606, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.7282, "num_token_overlap": 11.6866, "num_token_query": 31.3098, "num_token_union": 65.0367, "num_word_context": 202.2411, "num_word_doc": 49.8017, "num_word_query": 23.2695, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10675.1765, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1855, "query_norm": 1.3165, "queue_k_norm": 1.4008, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3098, "sent_len_1": 66.7282, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.8113, "stdk": 0.0483, "stdq": 0.043, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 45600 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.6614, "doc_norm": 1.3965, "encoder_q-embeddings": 5591.1885, "encoder_q-layer.0": 3565.1453, "encoder_q-layer.1": 3761.4624, "encoder_q-layer.10": 6800.9248, "encoder_q-layer.11": 15699.2607, "encoder_q-layer.2": 4204.2593, "encoder_q-layer.3": 4411.7627, "encoder_q-layer.4": 4702.4141, "encoder_q-layer.5": 4763.6309, "encoder_q-layer.6": 5133.8496, "encoder_q-layer.7": 6106.4468, "encoder_q-layer.8": 7569.4414, "encoder_q-layer.9": 6602.9746, "epoch": 0.2, "inbatch_neg_score": 0.1887, "inbatch_pos_score": 0.7637, "learning_rate": 4.060526315789474e-05, "loss": 3.6614, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.6276, "num_token_overlap": 11.6853, "num_token_query": 31.4312, "num_token_union": 65.0662, "num_word_context": 202.0354, "num_word_doc": 49.7066, "num_word_query": 23.3542, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10204.2878, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.188, "query_norm": 1.3259, "queue_k_norm": 1.3993, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4312, "sent_len_1": 66.6276, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.785, "stdk": 0.0481, "stdq": 0.0429, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 45700 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.6643, "doc_norm": 1.4017, "encoder_q-embeddings": 17628.9062, "encoder_q-layer.0": 12489.3857, "encoder_q-layer.1": 13410.1299, "encoder_q-layer.10": 13586.5537, "encoder_q-layer.11": 30373.0293, "encoder_q-layer.2": 15307.0625, "encoder_q-layer.3": 15641.4736, "encoder_q-layer.4": 15484.0518, "encoder_q-layer.5": 12440.6299, "encoder_q-layer.6": 12606.1973, "encoder_q-layer.7": 13484.4336, "encoder_q-layer.8": 14549.8037, "encoder_q-layer.9": 13207.9434, "epoch": 0.2, "inbatch_neg_score": 0.1973, "inbatch_pos_score": 0.7886, "learning_rate": 4.0578947368421054e-05, "loss": 3.6643, "norm_diff": 0.0383, "norm_loss": 0.0, "num_token_doc": 66.8067, "num_token_overlap": 11.7109, "num_token_query": 31.3839, "num_token_union": 65.1492, "num_word_context": 202.4195, "num_word_doc": 49.8624, "num_word_query": 23.3335, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24226.5675, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.197, "query_norm": 1.3634, "queue_k_norm": 1.4001, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3839, "sent_len_1": 66.8067, "sent_len_max_0": 127.9525, "sent_len_max_1": 207.0137, "stdk": 0.0483, "stdq": 0.0438, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 45800 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.6451, "doc_norm": 1.4053, "encoder_q-embeddings": 10829.3311, "encoder_q-layer.0": 6940.5605, "encoder_q-layer.1": 7096.5645, "encoder_q-layer.10": 13600.8018, "encoder_q-layer.11": 29286.6172, "encoder_q-layer.2": 7958.8496, "encoder_q-layer.3": 8293.4941, "encoder_q-layer.4": 8737.1621, "encoder_q-layer.5": 8823.3379, "encoder_q-layer.6": 10168.4863, "encoder_q-layer.7": 12241.916, "encoder_q-layer.8": 13714.3662, "encoder_q-layer.9": 12303.1465, "epoch": 0.2, "inbatch_neg_score": 0.2047, "inbatch_pos_score": 0.7808, "learning_rate": 4.055263157894737e-05, "loss": 3.6451, "norm_diff": 0.0605, "norm_loss": 0.0, "num_token_doc": 66.8888, "num_token_overlap": 11.7107, "num_token_query": 31.4166, "num_token_union": 65.1871, "num_word_context": 202.2565, "num_word_doc": 49.9071, "num_word_query": 23.3412, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19175.5466, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.2052, "query_norm": 1.3448, "queue_k_norm": 1.4032, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4166, "sent_len_1": 66.8888, "sent_len_max_0": 127.975, "sent_len_max_1": 209.4588, "stdk": 0.0484, "stdq": 0.0428, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 45900 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.6581, "doc_norm": 1.4075, "encoder_q-embeddings": 5596.1089, "encoder_q-layer.0": 3735.9651, "encoder_q-layer.1": 3902.2761, "encoder_q-layer.10": 6313.6416, "encoder_q-layer.11": 15051.4277, "encoder_q-layer.2": 4188.6353, "encoder_q-layer.3": 4305.6709, "encoder_q-layer.4": 4417.9463, "encoder_q-layer.5": 4699.5962, "encoder_q-layer.6": 5102.9126, "encoder_q-layer.7": 5690.4556, "encoder_q-layer.8": 6734.4214, "encoder_q-layer.9": 6286.7754, "epoch": 0.2, "inbatch_neg_score": 0.2126, "inbatch_pos_score": 0.7954, "learning_rate": 4.0526315789473684e-05, "loss": 3.6581, "norm_diff": 0.0646, "norm_loss": 0.0, "num_token_doc": 66.789, "num_token_overlap": 11.6947, "num_token_query": 31.4727, "num_token_union": 65.1736, "num_word_context": 202.4084, "num_word_doc": 49.8702, "num_word_query": 23.3883, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9796.2026, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2129, "query_norm": 1.3429, "queue_k_norm": 1.4032, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4727, "sent_len_1": 66.789, "sent_len_max_0": 127.9988, "sent_len_max_1": 206.8562, "stdk": 0.0484, "stdq": 0.0433, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 46000 }, { "accuracy": 46.9238, "active_queue_size": 16384.0, "cl_loss": 3.671, "doc_norm": 1.4063, "encoder_q-embeddings": 6159.146, "encoder_q-layer.0": 4216.0264, "encoder_q-layer.1": 4386.4043, "encoder_q-layer.10": 7132.6362, "encoder_q-layer.11": 16224.5049, "encoder_q-layer.2": 4837.8179, "encoder_q-layer.3": 4834.7139, "encoder_q-layer.4": 5191.02, "encoder_q-layer.5": 5513.1221, "encoder_q-layer.6": 5849.5957, "encoder_q-layer.7": 6689.6709, "encoder_q-layer.8": 8024.3467, "encoder_q-layer.9": 6860.2046, "epoch": 0.2, "inbatch_neg_score": 0.2159, "inbatch_pos_score": 0.7866, "learning_rate": 4.05e-05, "loss": 3.671, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.6216, "num_token_overlap": 11.6719, "num_token_query": 31.3885, "num_token_union": 65.0658, "num_word_context": 202.3638, "num_word_doc": 49.7212, "num_word_query": 23.3082, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10805.0188, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2148, "query_norm": 1.3364, "queue_k_norm": 1.4058, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3885, "sent_len_1": 66.6216, "sent_len_max_0": 127.985, "sent_len_max_1": 207.4512, "stdk": 0.0483, "stdq": 0.0435, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46100 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.6774, "doc_norm": 1.4063, "encoder_q-embeddings": 5212.7656, "encoder_q-layer.0": 3512.7861, "encoder_q-layer.1": 3552.582, "encoder_q-layer.10": 6958.1299, "encoder_q-layer.11": 15426.2783, "encoder_q-layer.2": 3948.3037, "encoder_q-layer.3": 3967.8726, "encoder_q-layer.4": 4206.0425, "encoder_q-layer.5": 4352.5356, "encoder_q-layer.6": 4894.7861, "encoder_q-layer.7": 5905.8105, "encoder_q-layer.8": 6903.6934, "encoder_q-layer.9": 6284.3135, "epoch": 0.2, "inbatch_neg_score": 0.2156, "inbatch_pos_score": 0.7949, "learning_rate": 4.047368421052632e-05, "loss": 3.6774, "norm_diff": 0.0978, "norm_loss": 0.0, "num_token_doc": 66.6528, "num_token_overlap": 11.6291, "num_token_query": 31.2609, "num_token_union": 65.0282, "num_word_context": 202.26, "num_word_doc": 49.7531, "num_word_query": 23.2253, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9657.7081, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2152, "query_norm": 1.3085, "queue_k_norm": 1.405, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2609, "sent_len_1": 66.6528, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.8162, "stdk": 0.0483, "stdq": 0.043, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 46200 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.6697, "doc_norm": 1.4042, "encoder_q-embeddings": 5425.4985, "encoder_q-layer.0": 3596.3027, "encoder_q-layer.1": 3704.323, "encoder_q-layer.10": 6432.2358, "encoder_q-layer.11": 15003.0967, "encoder_q-layer.2": 4187.4185, "encoder_q-layer.3": 4113.4849, "encoder_q-layer.4": 4275.6924, "encoder_q-layer.5": 4443.1455, "encoder_q-layer.6": 5009.7524, "encoder_q-layer.7": 5499.79, "encoder_q-layer.8": 6988.0952, "encoder_q-layer.9": 6308.5371, "epoch": 0.2, "inbatch_neg_score": 0.2132, "inbatch_pos_score": 0.79, "learning_rate": 4.044736842105263e-05, "loss": 3.6697, "norm_diff": 0.1075, "norm_loss": 0.0, "num_token_doc": 66.8408, "num_token_overlap": 11.7015, "num_token_query": 31.4262, "num_token_union": 65.1819, "num_word_context": 202.3909, "num_word_doc": 49.9032, "num_word_query": 23.3442, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9717.1722, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2128, "query_norm": 1.2967, "queue_k_norm": 1.4068, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4262, "sent_len_1": 66.8408, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.1875, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46300 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.673, "doc_norm": 1.4034, "encoder_q-embeddings": 5660.2041, "encoder_q-layer.0": 3606.2463, "encoder_q-layer.1": 3835.79, "encoder_q-layer.10": 7358.7217, "encoder_q-layer.11": 15868.4326, "encoder_q-layer.2": 4302.3389, "encoder_q-layer.3": 4464.2793, "encoder_q-layer.4": 4772.0269, "encoder_q-layer.5": 4899.5605, "encoder_q-layer.6": 5438.6543, "encoder_q-layer.7": 6220.5425, "encoder_q-layer.8": 7846.0166, "encoder_q-layer.9": 6975.897, "epoch": 0.2, "inbatch_neg_score": 0.2107, "inbatch_pos_score": 0.7803, "learning_rate": 4.042105263157895e-05, "loss": 3.673, "norm_diff": 0.0999, "norm_loss": 0.0, "num_token_doc": 66.633, "num_token_overlap": 11.6935, "num_token_query": 31.4131, "num_token_union": 65.0405, "num_word_context": 202.0089, "num_word_doc": 49.7128, "num_word_query": 23.3336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10355.4081, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2098, "query_norm": 1.3035, "queue_k_norm": 1.4075, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4131, "sent_len_1": 66.633, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.0525, "stdk": 0.0481, "stdq": 0.0433, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46400 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.6748, "doc_norm": 1.4098, "encoder_q-embeddings": 5934.6953, "encoder_q-layer.0": 3984.6362, "encoder_q-layer.1": 4187.0981, "encoder_q-layer.10": 6511.8862, "encoder_q-layer.11": 14983.458, "encoder_q-layer.2": 4850.3594, "encoder_q-layer.3": 5322.7666, "encoder_q-layer.4": 6104.2393, "encoder_q-layer.5": 6412.9268, "encoder_q-layer.6": 7192.854, "encoder_q-layer.7": 7404.4639, "encoder_q-layer.8": 8418.4551, "encoder_q-layer.9": 6365.3198, "epoch": 0.2, "inbatch_neg_score": 0.2098, "inbatch_pos_score": 0.7764, "learning_rate": 4.039473684210526e-05, "loss": 3.6748, "norm_diff": 0.1236, "norm_loss": 0.0, "num_token_doc": 66.8208, "num_token_overlap": 11.6363, "num_token_query": 31.3179, "num_token_union": 65.1736, "num_word_context": 202.4275, "num_word_doc": 49.8465, "num_word_query": 23.2371, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10759.2146, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2101, "query_norm": 1.2862, "queue_k_norm": 1.4078, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3179, "sent_len_1": 66.8208, "sent_len_max_0": 127.9575, "sent_len_max_1": 207.7625, "stdk": 0.0484, "stdq": 0.0427, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 46500 }, { "accuracy": 47.3633, "active_queue_size": 16384.0, "cl_loss": 3.662, "doc_norm": 1.4088, "encoder_q-embeddings": 5336.6763, "encoder_q-layer.0": 3632.7065, "encoder_q-layer.1": 3805.531, "encoder_q-layer.10": 6623.3896, "encoder_q-layer.11": 15124.6338, "encoder_q-layer.2": 4241.0039, "encoder_q-layer.3": 4559.1572, "encoder_q-layer.4": 4667.7998, "encoder_q-layer.5": 4750.7925, "encoder_q-layer.6": 5332.7295, "encoder_q-layer.7": 6498.6709, "encoder_q-layer.8": 7480.5801, "encoder_q-layer.9": 6359.3755, "epoch": 0.2, "inbatch_neg_score": 0.2068, "inbatch_pos_score": 0.7856, "learning_rate": 4.036842105263158e-05, "loss": 3.662, "norm_diff": 0.1174, "norm_loss": 0.0, "num_token_doc": 66.734, "num_token_overlap": 11.677, "num_token_query": 31.3727, "num_token_union": 65.1017, "num_word_context": 202.338, "num_word_doc": 49.7794, "num_word_query": 23.2996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10021.2926, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2068, "query_norm": 1.2914, "queue_k_norm": 1.4085, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3727, "sent_len_1": 66.734, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.21, "stdk": 0.0483, "stdq": 0.043, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46600 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.6524, "doc_norm": 1.4075, "encoder_q-embeddings": 8026.7588, "encoder_q-layer.0": 5673.5815, "encoder_q-layer.1": 5942.0947, "encoder_q-layer.10": 6762.999, "encoder_q-layer.11": 15567.0693, "encoder_q-layer.2": 6695.0649, "encoder_q-layer.3": 6349.2329, "encoder_q-layer.4": 6130.7568, "encoder_q-layer.5": 5533.2837, "encoder_q-layer.6": 5865.2773, "encoder_q-layer.7": 6255.0732, "encoder_q-layer.8": 7100.5391, "encoder_q-layer.9": 6395.3677, "epoch": 0.2, "inbatch_neg_score": 0.2022, "inbatch_pos_score": 0.7749, "learning_rate": 4.03421052631579e-05, "loss": 3.6524, "norm_diff": 0.1362, "norm_loss": 0.0, "num_token_doc": 66.7627, "num_token_overlap": 11.6862, "num_token_query": 31.366, "num_token_union": 65.1252, "num_word_context": 202.2153, "num_word_doc": 49.8138, "num_word_query": 23.2983, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11206.5339, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2026, "query_norm": 1.2713, "queue_k_norm": 1.4084, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.366, "sent_len_1": 66.7627, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.845, "stdk": 0.0483, "stdq": 0.0423, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46700 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.6776, "doc_norm": 1.4083, "encoder_q-embeddings": 5560.9937, "encoder_q-layer.0": 3669.8362, "encoder_q-layer.1": 3876.3123, "encoder_q-layer.10": 6550.7534, "encoder_q-layer.11": 16343.6445, "encoder_q-layer.2": 4306.1289, "encoder_q-layer.3": 4386.8345, "encoder_q-layer.4": 4610.0103, "encoder_q-layer.5": 4607.7607, "encoder_q-layer.6": 5167.7471, "encoder_q-layer.7": 5649.0762, "encoder_q-layer.8": 6929.084, "encoder_q-layer.9": 6273.2969, "epoch": 0.2, "inbatch_neg_score": 0.1972, "inbatch_pos_score": 0.7627, "learning_rate": 4.031578947368421e-05, "loss": 3.6776, "norm_diff": 0.1412, "norm_loss": 0.0, "num_token_doc": 66.8108, "num_token_overlap": 11.6552, "num_token_query": 31.3654, "num_token_union": 65.1301, "num_word_context": 202.2516, "num_word_doc": 49.8646, "num_word_query": 23.3013, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10003.4977, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1973, "query_norm": 1.2671, "queue_k_norm": 1.4085, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3654, "sent_len_1": 66.8108, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.84, "stdk": 0.0483, "stdq": 0.0423, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 46800 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.6631, "doc_norm": 1.4088, "encoder_q-embeddings": 5657.1055, "encoder_q-layer.0": 3655.2346, "encoder_q-layer.1": 3850.1348, "encoder_q-layer.10": 6576.2568, "encoder_q-layer.11": 15258.6396, "encoder_q-layer.2": 4305.9409, "encoder_q-layer.3": 4393.3237, "encoder_q-layer.4": 4871.8496, "encoder_q-layer.5": 4914.2358, "encoder_q-layer.6": 5457.5342, "encoder_q-layer.7": 5928.6611, "encoder_q-layer.8": 7352.8164, "encoder_q-layer.9": 6470.1797, "epoch": 0.2, "inbatch_neg_score": 0.1962, "inbatch_pos_score": 0.7769, "learning_rate": 4.028947368421053e-05, "loss": 3.6631, "norm_diff": 0.1092, "norm_loss": 0.0, "num_token_doc": 66.7206, "num_token_overlap": 11.6631, "num_token_query": 31.3335, "num_token_union": 65.0832, "num_word_context": 202.2449, "num_word_doc": 49.806, "num_word_query": 23.2756, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10007.0506, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1962, "query_norm": 1.2996, "queue_k_norm": 1.4107, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3335, "sent_len_1": 66.7206, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.6037, "stdk": 0.0483, "stdq": 0.0435, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 46900 }, { "accuracy": 47.5586, "active_queue_size": 16384.0, "cl_loss": 3.6497, "doc_norm": 1.4092, "encoder_q-embeddings": 5587.7319, "encoder_q-layer.0": 3881.717, "encoder_q-layer.1": 4141.7075, "encoder_q-layer.10": 7365.7866, "encoder_q-layer.11": 15675.4297, "encoder_q-layer.2": 4562.3906, "encoder_q-layer.3": 4807.4741, "encoder_q-layer.4": 5140.7007, "encoder_q-layer.5": 5467.9951, "encoder_q-layer.6": 6149.2744, "encoder_q-layer.7": 6807.9263, "encoder_q-layer.8": 7433.3472, "encoder_q-layer.9": 7001.688, "epoch": 0.2, "inbatch_neg_score": 0.1975, "inbatch_pos_score": 0.7861, "learning_rate": 4.026315789473684e-05, "loss": 3.6497, "norm_diff": 0.1044, "norm_loss": 0.0, "num_token_doc": 66.8067, "num_token_overlap": 11.6585, "num_token_query": 31.3115, "num_token_union": 65.0999, "num_word_context": 202.4658, "num_word_doc": 49.8353, "num_word_query": 23.2336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10433.8167, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1971, "query_norm": 1.3048, "queue_k_norm": 1.411, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3115, "sent_len_1": 66.8067, "sent_len_max_0": 127.99, "sent_len_max_1": 208.17, "stdk": 0.0483, "stdq": 0.0436, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47000 }, { "accuracy": 45.3613, "active_queue_size": 16384.0, "cl_loss": 3.6395, "doc_norm": 1.4065, "encoder_q-embeddings": 6022.2393, "encoder_q-layer.0": 3926.3433, "encoder_q-layer.1": 4205.1729, "encoder_q-layer.10": 7425.7744, "encoder_q-layer.11": 16830.2324, "encoder_q-layer.2": 4633.2144, "encoder_q-layer.3": 4947.9634, "encoder_q-layer.4": 5421.6973, "encoder_q-layer.5": 5601.9399, "encoder_q-layer.6": 5862.3818, "encoder_q-layer.7": 6764.9751, "encoder_q-layer.8": 7791.8525, "encoder_q-layer.9": 7156.4136, "epoch": 0.2, "inbatch_neg_score": 0.1987, "inbatch_pos_score": 0.7637, "learning_rate": 4.023684210526316e-05, "loss": 3.6395, "norm_diff": 0.1184, "norm_loss": 0.0, "num_token_doc": 66.9226, "num_token_overlap": 11.704, "num_token_query": 31.4213, "num_token_union": 65.2095, "num_word_context": 202.3261, "num_word_doc": 49.9164, "num_word_query": 23.3467, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10946.9001, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1987, "query_norm": 1.2881, "queue_k_norm": 1.4123, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4213, "sent_len_1": 66.9226, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.8, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 47100 }, { "accuracy": 46.7773, "active_queue_size": 16384.0, "cl_loss": 3.6669, "doc_norm": 1.4113, "encoder_q-embeddings": 7315.8652, "encoder_q-layer.0": 4936.5679, "encoder_q-layer.1": 5779.8472, "encoder_q-layer.10": 7094.6055, "encoder_q-layer.11": 16259.0947, "encoder_q-layer.2": 6842.8247, "encoder_q-layer.3": 7355.5869, "encoder_q-layer.4": 7893.7295, "encoder_q-layer.5": 7890.9673, "encoder_q-layer.6": 8411.5068, "encoder_q-layer.7": 8758.1182, "encoder_q-layer.8": 8535.4609, "encoder_q-layer.9": 7040.0342, "epoch": 0.2, "inbatch_neg_score": 0.1946, "inbatch_pos_score": 0.7705, "learning_rate": 4.021052631578948e-05, "loss": 3.6669, "norm_diff": 0.1198, "norm_loss": 0.0, "num_token_doc": 66.6948, "num_token_overlap": 11.6575, "num_token_query": 31.3954, "num_token_union": 65.0837, "num_word_context": 202.2947, "num_word_doc": 49.7881, "num_word_query": 23.3254, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12588.407, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1942, "query_norm": 1.2915, "queue_k_norm": 1.41, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3954, "sent_len_1": 66.6948, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.6438, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47200 }, { "accuracy": 48.1445, "active_queue_size": 16384.0, "cl_loss": 3.6594, "doc_norm": 1.4058, "encoder_q-embeddings": 5287.5952, "encoder_q-layer.0": 3617.9028, "encoder_q-layer.1": 3783.9094, "encoder_q-layer.10": 6291.3213, "encoder_q-layer.11": 15381.3291, "encoder_q-layer.2": 4357.5781, "encoder_q-layer.3": 4456.1953, "encoder_q-layer.4": 4675.8301, "encoder_q-layer.5": 4739.0435, "encoder_q-layer.6": 5261.5723, "encoder_q-layer.7": 5718.0264, "encoder_q-layer.8": 6659.8271, "encoder_q-layer.9": 6221.8506, "epoch": 0.21, "inbatch_neg_score": 0.1968, "inbatch_pos_score": 0.7773, "learning_rate": 4.018421052631579e-05, "loss": 3.6594, "norm_diff": 0.1192, "norm_loss": 0.0, "num_token_doc": 66.7147, "num_token_overlap": 11.6533, "num_token_query": 31.3133, "num_token_union": 65.0736, "num_word_context": 202.3194, "num_word_doc": 49.7944, "num_word_query": 23.2462, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9988.0621, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1971, "query_norm": 1.2866, "queue_k_norm": 1.4095, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3133, "sent_len_1": 66.7147, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.3288, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47300 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.6518, "doc_norm": 1.41, "encoder_q-embeddings": 5828.8442, "encoder_q-layer.0": 3773.5583, "encoder_q-layer.1": 3885.4326, "encoder_q-layer.10": 6872.3594, "encoder_q-layer.11": 16161.1514, "encoder_q-layer.2": 4291.5093, "encoder_q-layer.3": 4450.1528, "encoder_q-layer.4": 4746.3911, "encoder_q-layer.5": 4796.7271, "encoder_q-layer.6": 5262.9302, "encoder_q-layer.7": 5792.4141, "encoder_q-layer.8": 7344.1572, "encoder_q-layer.9": 6492.4546, "epoch": 0.21, "inbatch_neg_score": 0.1918, "inbatch_pos_score": 0.7686, "learning_rate": 4.015789473684211e-05, "loss": 3.6518, "norm_diff": 0.135, "norm_loss": 0.0, "num_token_doc": 66.8835, "num_token_overlap": 11.6778, "num_token_query": 31.335, "num_token_union": 65.1465, "num_word_context": 202.2445, "num_word_doc": 49.8807, "num_word_query": 23.2582, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10302.6532, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1918, "query_norm": 1.275, "queue_k_norm": 1.4094, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.335, "sent_len_1": 66.8835, "sent_len_max_0": 127.99, "sent_len_max_1": 210.0375, "stdk": 0.0483, "stdq": 0.0426, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47400 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.6553, "doc_norm": 1.4072, "encoder_q-embeddings": 5104.3042, "encoder_q-layer.0": 3432.6243, "encoder_q-layer.1": 3711.0444, "encoder_q-layer.10": 6929.0479, "encoder_q-layer.11": 15901.8223, "encoder_q-layer.2": 4006.8486, "encoder_q-layer.3": 4197.5098, "encoder_q-layer.4": 4440.5225, "encoder_q-layer.5": 4574.1597, "encoder_q-layer.6": 5258.1655, "encoder_q-layer.7": 5748.1934, "encoder_q-layer.8": 7091.5288, "encoder_q-layer.9": 6275.8701, "epoch": 0.21, "inbatch_neg_score": 0.1969, "inbatch_pos_score": 0.7744, "learning_rate": 4.0131578947368425e-05, "loss": 3.6553, "norm_diff": 0.1229, "norm_loss": 0.0, "num_token_doc": 66.6342, "num_token_overlap": 11.6554, "num_token_query": 31.3407, "num_token_union": 64.9998, "num_word_context": 202.0366, "num_word_doc": 49.7121, "num_word_query": 23.2874, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9911.3574, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1962, "query_norm": 1.2844, "queue_k_norm": 1.4095, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3407, "sent_len_1": 66.6342, "sent_len_max_0": 127.98, "sent_len_max_1": 209.3762, "stdk": 0.0482, "stdq": 0.0428, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47500 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.6724, "doc_norm": 1.4101, "encoder_q-embeddings": 5750.7808, "encoder_q-layer.0": 3847.0176, "encoder_q-layer.1": 3934.1646, "encoder_q-layer.10": 6584.6318, "encoder_q-layer.11": 14741.2627, "encoder_q-layer.2": 4353.5107, "encoder_q-layer.3": 4811.2251, "encoder_q-layer.4": 4990.3896, "encoder_q-layer.5": 5099.9023, "encoder_q-layer.6": 5802.9927, "encoder_q-layer.7": 6316.9429, "encoder_q-layer.8": 7147.6753, "encoder_q-layer.9": 6458.6172, "epoch": 0.21, "inbatch_neg_score": 0.1899, "inbatch_pos_score": 0.7715, "learning_rate": 4.010526315789474e-05, "loss": 3.6724, "norm_diff": 0.1209, "norm_loss": 0.0, "num_token_doc": 66.5824, "num_token_overlap": 11.6307, "num_token_query": 31.2705, "num_token_union": 64.9865, "num_word_context": 202.0305, "num_word_doc": 49.683, "num_word_query": 23.2168, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10050.7872, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.1896, "query_norm": 1.2892, "queue_k_norm": 1.4093, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2705, "sent_len_1": 66.5824, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.3475, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47600 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.6401, "doc_norm": 1.4127, "encoder_q-embeddings": 2920.9136, "encoder_q-layer.0": 1947.7849, "encoder_q-layer.1": 2034.4401, "encoder_q-layer.10": 3408.1851, "encoder_q-layer.11": 7725.9331, "encoder_q-layer.2": 2387.8816, "encoder_q-layer.3": 2427.6858, "encoder_q-layer.4": 2640.5881, "encoder_q-layer.5": 2699.8162, "encoder_q-layer.6": 2803.3123, "encoder_q-layer.7": 2988.7566, "encoder_q-layer.8": 3506.167, "encoder_q-layer.9": 3130.2617, "epoch": 0.21, "inbatch_neg_score": 0.1925, "inbatch_pos_score": 0.769, "learning_rate": 4.0078947368421055e-05, "loss": 3.6401, "norm_diff": 0.1282, "norm_loss": 0.0, "num_token_doc": 66.73, "num_token_overlap": 11.6696, "num_token_query": 31.3703, "num_token_union": 65.0961, "num_word_context": 202.4126, "num_word_doc": 49.8337, "num_word_query": 23.3025, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5083.3957, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1925, "query_norm": 1.2845, "queue_k_norm": 1.4093, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3703, "sent_len_1": 66.73, "sent_len_max_0": 127.985, "sent_len_max_1": 208.6725, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47700 }, { "accuracy": 45.8496, "active_queue_size": 16384.0, "cl_loss": 3.6549, "doc_norm": 1.4063, "encoder_q-embeddings": 2748.3057, "encoder_q-layer.0": 1786.7235, "encoder_q-layer.1": 1889.1217, "encoder_q-layer.10": 3298.8362, "encoder_q-layer.11": 7895.6909, "encoder_q-layer.2": 2173.3269, "encoder_q-layer.3": 2248.6765, "encoder_q-layer.4": 2386.741, "encoder_q-layer.5": 2374.4321, "encoder_q-layer.6": 2628.5369, "encoder_q-layer.7": 2989.3459, "encoder_q-layer.8": 3521.2332, "encoder_q-layer.9": 3154.3931, "epoch": 0.21, "inbatch_neg_score": 0.1969, "inbatch_pos_score": 0.7598, "learning_rate": 4.0052631578947367e-05, "loss": 3.6549, "norm_diff": 0.1127, "norm_loss": 0.0, "num_token_doc": 66.8136, "num_token_overlap": 11.675, "num_token_query": 31.3768, "num_token_union": 65.1297, "num_word_context": 202.3527, "num_word_doc": 49.871, "num_word_query": 23.2935, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5052.3111, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1954, "query_norm": 1.2937, "queue_k_norm": 1.4102, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3768, "sent_len_1": 66.8136, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.6213, "stdk": 0.0482, "stdq": 0.0429, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 47800 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.6503, "doc_norm": 1.4103, "encoder_q-embeddings": 2937.1709, "encoder_q-layer.0": 1888.5948, "encoder_q-layer.1": 2060.6125, "encoder_q-layer.10": 3371.8274, "encoder_q-layer.11": 7626.0542, "encoder_q-layer.2": 2262.95, "encoder_q-layer.3": 2342.054, "encoder_q-layer.4": 2526.5132, "encoder_q-layer.5": 2493.054, "encoder_q-layer.6": 2842.9453, "encoder_q-layer.7": 3107.0911, "encoder_q-layer.8": 3590.8994, "encoder_q-layer.9": 3271.5996, "epoch": 0.21, "inbatch_neg_score": 0.1982, "inbatch_pos_score": 0.7817, "learning_rate": 4.0026315789473685e-05, "loss": 3.6503, "norm_diff": 0.115, "norm_loss": 0.0, "num_token_doc": 66.8166, "num_token_overlap": 11.7065, "num_token_query": 31.3061, "num_token_union": 65.0559, "num_word_context": 202.0829, "num_word_doc": 49.8523, "num_word_query": 23.2404, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5091.5882, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.1978, "query_norm": 1.2953, "queue_k_norm": 1.4113, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3061, "sent_len_1": 66.8166, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.1987, "stdk": 0.0484, "stdq": 0.0428, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 47900 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.6572, "doc_norm": 1.4209, "encoder_q-embeddings": 2869.7722, "encoder_q-layer.0": 1881.1644, "encoder_q-layer.1": 1993.0613, "encoder_q-layer.10": 3274.9333, "encoder_q-layer.11": 7656.5439, "encoder_q-layer.2": 2238.6055, "encoder_q-layer.3": 2359.9338, "encoder_q-layer.4": 2391.855, "encoder_q-layer.5": 2484.6445, "encoder_q-layer.6": 2688.4421, "encoder_q-layer.7": 2964.4299, "encoder_q-layer.8": 3427.7104, "encoder_q-layer.9": 3152.4814, "epoch": 0.21, "inbatch_neg_score": 0.2025, "inbatch_pos_score": 0.792, "learning_rate": 4e-05, "loss": 3.6572, "norm_diff": 0.1073, "norm_loss": 0.0, "num_token_doc": 66.6961, "num_token_overlap": 11.6625, "num_token_query": 31.3676, "num_token_union": 65.0421, "num_word_context": 202.0622, "num_word_doc": 49.7499, "num_word_query": 23.2797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5058.45, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2028, "query_norm": 1.3135, "queue_k_norm": 1.4097, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3676, "sent_len_1": 66.6961, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.0087, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 48000 }, { "accuracy": 46.2891, "active_queue_size": 16384.0, "cl_loss": 3.6498, "doc_norm": 1.4167, "encoder_q-embeddings": 3221.3845, "encoder_q-layer.0": 2263.9189, "encoder_q-layer.1": 2574.8535, "encoder_q-layer.10": 3761.5706, "encoder_q-layer.11": 8046.4761, "encoder_q-layer.2": 2965.2507, "encoder_q-layer.3": 3033.031, "encoder_q-layer.4": 3199.9514, "encoder_q-layer.5": 3252.3799, "encoder_q-layer.6": 3407.4507, "encoder_q-layer.7": 3665.1379, "encoder_q-layer.8": 4127.3965, "encoder_q-layer.9": 3464.9509, "epoch": 0.21, "inbatch_neg_score": 0.2062, "inbatch_pos_score": 0.7925, "learning_rate": 3.997368421052632e-05, "loss": 3.6498, "norm_diff": 0.072, "norm_loss": 0.0, "num_token_doc": 66.7257, "num_token_overlap": 11.6818, "num_token_query": 31.4023, "num_token_union": 65.0902, "num_word_context": 202.381, "num_word_doc": 49.7937, "num_word_query": 23.3371, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5623.8955, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2058, "query_norm": 1.3447, "queue_k_norm": 1.4099, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4023, "sent_len_1": 66.7257, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.23, "stdk": 0.0486, "stdq": 0.0438, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 48100 }, { "accuracy": 47.5586, "active_queue_size": 16384.0, "cl_loss": 3.651, "doc_norm": 1.409, "encoder_q-embeddings": 1366.1505, "encoder_q-layer.0": 902.3486, "encoder_q-layer.1": 948.6983, "encoder_q-layer.10": 1711.2354, "encoder_q-layer.11": 3924.8413, "encoder_q-layer.2": 1044.6512, "encoder_q-layer.3": 1123.4127, "encoder_q-layer.4": 1137.3049, "encoder_q-layer.5": 1148.8163, "encoder_q-layer.6": 1366.2463, "encoder_q-layer.7": 1443.7529, "encoder_q-layer.8": 1790.8171, "encoder_q-layer.9": 1661.2534, "epoch": 0.21, "inbatch_neg_score": 0.2135, "inbatch_pos_score": 0.79, "learning_rate": 3.994736842105263e-05, "loss": 3.651, "norm_diff": 0.057, "norm_loss": 0.0, "num_token_doc": 66.8739, "num_token_overlap": 11.7012, "num_token_query": 31.3928, "num_token_union": 65.1685, "num_word_context": 202.3284, "num_word_doc": 49.9127, "num_word_query": 23.3155, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2530.5307, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2126, "query_norm": 1.352, "queue_k_norm": 1.412, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3928, "sent_len_1": 66.8739, "sent_len_max_0": 127.995, "sent_len_max_1": 207.975, "stdk": 0.0483, "stdq": 0.0436, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48200 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.6661, "doc_norm": 1.405, "encoder_q-embeddings": 2498.7896, "encoder_q-layer.0": 1906.9036, "encoder_q-layer.1": 1981.6519, "encoder_q-layer.10": 1661.4558, "encoder_q-layer.11": 3819.0269, "encoder_q-layer.2": 1978.1196, "encoder_q-layer.3": 1812.0645, "encoder_q-layer.4": 1928.4481, "encoder_q-layer.5": 1709.4181, "encoder_q-layer.6": 1866.007, "encoder_q-layer.7": 1835.3407, "encoder_q-layer.8": 1848.75, "encoder_q-layer.9": 1629.5621, "epoch": 0.21, "inbatch_neg_score": 0.2219, "inbatch_pos_score": 0.7935, "learning_rate": 3.9921052631578945e-05, "loss": 3.6661, "norm_diff": 0.0589, "norm_loss": 0.0, "num_token_doc": 66.6822, "num_token_overlap": 11.6505, "num_token_query": 31.3275, "num_token_union": 65.0642, "num_word_context": 201.7789, "num_word_doc": 49.7207, "num_word_query": 23.2627, "postclip_grad_norm": 1.0, "preclip_grad_norm": 3171.3235, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2213, "query_norm": 1.3461, "queue_k_norm": 1.412, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3275, "sent_len_1": 66.6822, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.7775, "stdk": 0.0481, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48300 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.6552, "doc_norm": 1.4174, "encoder_q-embeddings": 1381.3837, "encoder_q-layer.0": 890.6868, "encoder_q-layer.1": 923.5435, "encoder_q-layer.10": 1743.5955, "encoder_q-layer.11": 3904.4858, "encoder_q-layer.2": 1037.2321, "encoder_q-layer.3": 1077.5321, "encoder_q-layer.4": 1125.3113, "encoder_q-layer.5": 1176.3308, "encoder_q-layer.6": 1307.2982, "encoder_q-layer.7": 1512.0435, "encoder_q-layer.8": 1818.6555, "encoder_q-layer.9": 1665.6292, "epoch": 0.21, "inbatch_neg_score": 0.2313, "inbatch_pos_score": 0.8145, "learning_rate": 3.989473684210526e-05, "loss": 3.6552, "norm_diff": 0.0549, "norm_loss": 0.0, "num_token_doc": 66.8627, "num_token_overlap": 11.6442, "num_token_query": 31.2711, "num_token_union": 65.1378, "num_word_context": 202.4005, "num_word_doc": 49.8598, "num_word_query": 23.2117, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2536.786, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2314, "query_norm": 1.3625, "queue_k_norm": 1.413, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2711, "sent_len_1": 66.8627, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.2463, "stdk": 0.0486, "stdq": 0.0435, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48400 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.6521, "doc_norm": 1.4208, "encoder_q-embeddings": 1483.2859, "encoder_q-layer.0": 1013.3452, "encoder_q-layer.1": 1069.3018, "encoder_q-layer.10": 1649.5854, "encoder_q-layer.11": 3810.1567, "encoder_q-layer.2": 1183.6545, "encoder_q-layer.3": 1217.0488, "encoder_q-layer.4": 1305.6461, "encoder_q-layer.5": 1263.2477, "encoder_q-layer.6": 1324.9479, "encoder_q-layer.7": 1467.11, "encoder_q-layer.8": 1765.053, "encoder_q-layer.9": 1586.2974, "epoch": 0.21, "inbatch_neg_score": 0.2434, "inbatch_pos_score": 0.8232, "learning_rate": 3.986842105263158e-05, "loss": 3.6521, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.9254, "num_token_overlap": 11.6846, "num_token_query": 31.3508, "num_token_union": 65.1751, "num_word_context": 202.412, "num_word_doc": 49.937, "num_word_query": 23.295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2545.3979, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2432, "query_norm": 1.3378, "queue_k_norm": 1.415, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3508, "sent_len_1": 66.9254, "sent_len_max_0": 127.995, "sent_len_max_1": 208.4938, "stdk": 0.0486, "stdq": 0.0426, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48500 }, { "accuracy": 47.3145, "active_queue_size": 16384.0, "cl_loss": 3.6443, "doc_norm": 1.4149, "encoder_q-embeddings": 1374.5292, "encoder_q-layer.0": 915.4401, "encoder_q-layer.1": 952.2777, "encoder_q-layer.10": 1629.142, "encoder_q-layer.11": 3771.4517, "encoder_q-layer.2": 1091.2676, "encoder_q-layer.3": 1126.791, "encoder_q-layer.4": 1182.917, "encoder_q-layer.5": 1197.9725, "encoder_q-layer.6": 1385.9889, "encoder_q-layer.7": 1495.1777, "encoder_q-layer.8": 1740.5576, "encoder_q-layer.9": 1581.0171, "epoch": 0.21, "inbatch_neg_score": 0.2475, "inbatch_pos_score": 0.8223, "learning_rate": 3.98421052631579e-05, "loss": 3.6443, "norm_diff": 0.0748, "norm_loss": 0.0, "num_token_doc": 66.7468, "num_token_overlap": 11.6712, "num_token_query": 31.3854, "num_token_union": 65.1491, "num_word_context": 202.1233, "num_word_doc": 49.838, "num_word_query": 23.3119, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2465.9724, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2476, "query_norm": 1.34, "queue_k_norm": 1.4174, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3854, "sent_len_1": 66.7468, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.9625, "stdk": 0.0484, "stdq": 0.0432, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48600 }, { "accuracy": 46.5332, "active_queue_size": 16384.0, "cl_loss": 3.6478, "doc_norm": 1.4137, "encoder_q-embeddings": 1359.4177, "encoder_q-layer.0": 883.9133, "encoder_q-layer.1": 902.1669, "encoder_q-layer.10": 1600.9005, "encoder_q-layer.11": 3859.1479, "encoder_q-layer.2": 1008.1733, "encoder_q-layer.3": 1041.2291, "encoder_q-layer.4": 1147.5947, "encoder_q-layer.5": 1164.0388, "encoder_q-layer.6": 1240.1313, "encoder_q-layer.7": 1361.3459, "encoder_q-layer.8": 1708.1293, "encoder_q-layer.9": 1562.4546, "epoch": 0.21, "inbatch_neg_score": 0.2437, "inbatch_pos_score": 0.8062, "learning_rate": 3.981578947368421e-05, "loss": 3.6478, "norm_diff": 0.1024, "norm_loss": 0.0, "num_token_doc": 66.7077, "num_token_overlap": 11.6583, "num_token_query": 31.3204, "num_token_union": 65.0895, "num_word_context": 202.2191, "num_word_doc": 49.7691, "num_word_query": 23.2567, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2478.0636, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2445, "query_norm": 1.3113, "queue_k_norm": 1.418, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3204, "sent_len_1": 66.7077, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.0662, "stdk": 0.0483, "stdq": 0.0426, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48700 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.6469, "doc_norm": 1.42, "encoder_q-embeddings": 1417.9918, "encoder_q-layer.0": 940.9628, "encoder_q-layer.1": 1018.2664, "encoder_q-layer.10": 1897.0673, "encoder_q-layer.11": 3951.0625, "encoder_q-layer.2": 1140.3394, "encoder_q-layer.3": 1190.3044, "encoder_q-layer.4": 1269.5461, "encoder_q-layer.5": 1298.8029, "encoder_q-layer.6": 1466.4513, "encoder_q-layer.7": 1646.5515, "encoder_q-layer.8": 1906.0396, "encoder_q-layer.9": 1800.3395, "epoch": 0.21, "inbatch_neg_score": 0.246, "inbatch_pos_score": 0.8457, "learning_rate": 3.978947368421053e-05, "loss": 3.6469, "norm_diff": 0.1039, "norm_loss": 0.0, "num_token_doc": 66.847, "num_token_overlap": 11.6758, "num_token_query": 31.3481, "num_token_union": 65.1544, "num_word_context": 202.5538, "num_word_doc": 49.8666, "num_word_query": 23.2818, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2602.5569, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2468, "query_norm": 1.3161, "queue_k_norm": 1.419, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3481, "sent_len_1": 66.847, "sent_len_max_0": 127.98, "sent_len_max_1": 206.3537, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48800 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.6385, "doc_norm": 1.4194, "encoder_q-embeddings": 1387.1252, "encoder_q-layer.0": 908.9297, "encoder_q-layer.1": 975.0134, "encoder_q-layer.10": 1626.8071, "encoder_q-layer.11": 4044.783, "encoder_q-layer.2": 1129.7026, "encoder_q-layer.3": 1190.6707, "encoder_q-layer.4": 1311.837, "encoder_q-layer.5": 1329.6196, "encoder_q-layer.6": 1404.1788, "encoder_q-layer.7": 1535.4071, "encoder_q-layer.8": 1935.573, "encoder_q-layer.9": 1662.1609, "epoch": 0.21, "inbatch_neg_score": 0.2468, "inbatch_pos_score": 0.8188, "learning_rate": 3.976315789473685e-05, "loss": 3.6385, "norm_diff": 0.1188, "norm_loss": 0.0, "num_token_doc": 66.8403, "num_token_overlap": 11.679, "num_token_query": 31.4082, "num_token_union": 65.2087, "num_word_context": 202.5032, "num_word_doc": 49.8777, "num_word_query": 23.3134, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2578.6398, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2458, "query_norm": 1.3006, "queue_k_norm": 1.4201, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4082, "sent_len_1": 66.8403, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.365, "stdk": 0.0484, "stdq": 0.0426, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 48900 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.6491, "doc_norm": 1.4183, "encoder_q-embeddings": 1307.6459, "encoder_q-layer.0": 870.2969, "encoder_q-layer.1": 891.972, "encoder_q-layer.10": 1623.0955, "encoder_q-layer.11": 3929.1323, "encoder_q-layer.2": 984.2192, "encoder_q-layer.3": 1021.9614, "encoder_q-layer.4": 1080.9231, "encoder_q-layer.5": 1040.3973, "encoder_q-layer.6": 1199.8087, "encoder_q-layer.7": 1363.2034, "encoder_q-layer.8": 1732.7957, "encoder_q-layer.9": 1605.399, "epoch": 0.21, "inbatch_neg_score": 0.2386, "inbatch_pos_score": 0.8101, "learning_rate": 3.973684210526316e-05, "loss": 3.6491, "norm_diff": 0.1244, "norm_loss": 0.0, "num_token_doc": 66.8416, "num_token_overlap": 11.666, "num_token_query": 31.3336, "num_token_union": 65.1272, "num_word_context": 202.3196, "num_word_doc": 49.8634, "num_word_query": 23.277, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2417.1074, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2395, "query_norm": 1.294, "queue_k_norm": 1.417, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3336, "sent_len_1": 66.8416, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.8587, "stdk": 0.0484, "stdq": 0.0426, "stdqueue_k": 0.0483, "stdqueue_q": 0.0, "step": 49000 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.6445, "doc_norm": 1.4216, "encoder_q-embeddings": 1303.1193, "encoder_q-layer.0": 870.0363, "encoder_q-layer.1": 920.4363, "encoder_q-layer.10": 1509.8782, "encoder_q-layer.11": 3726.6794, "encoder_q-layer.2": 1004.2389, "encoder_q-layer.3": 1030.7252, "encoder_q-layer.4": 1111.7883, "encoder_q-layer.5": 1114.2871, "encoder_q-layer.6": 1220.887, "encoder_q-layer.7": 1395.7405, "encoder_q-layer.8": 1619.2977, "encoder_q-layer.9": 1475.5554, "epoch": 0.21, "inbatch_neg_score": 0.2353, "inbatch_pos_score": 0.8262, "learning_rate": 3.971052631578948e-05, "loss": 3.6445, "norm_diff": 0.1143, "norm_loss": 0.0, "num_token_doc": 66.7944, "num_token_overlap": 11.7013, "num_token_query": 31.4141, "num_token_union": 65.1446, "num_word_context": 202.3859, "num_word_doc": 49.8146, "num_word_query": 23.3248, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2343.729, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2352, "query_norm": 1.3073, "queue_k_norm": 1.4201, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4141, "sent_len_1": 66.7944, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.2887, "stdk": 0.0484, "stdq": 0.0432, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 49100 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.6558, "doc_norm": 1.423, "encoder_q-embeddings": 1413.9845, "encoder_q-layer.0": 936.4066, "encoder_q-layer.1": 985.1771, "encoder_q-layer.10": 1764.8676, "encoder_q-layer.11": 3929.9199, "encoder_q-layer.2": 1081.0239, "encoder_q-layer.3": 1123.4047, "encoder_q-layer.4": 1206.0267, "encoder_q-layer.5": 1262.0386, "encoder_q-layer.6": 1398.6541, "encoder_q-layer.7": 1616.8546, "encoder_q-layer.8": 1965.106, "encoder_q-layer.9": 1696.7723, "epoch": 0.21, "inbatch_neg_score": 0.2319, "inbatch_pos_score": 0.811, "learning_rate": 3.968421052631579e-05, "loss": 3.6558, "norm_diff": 0.1071, "norm_loss": 0.0, "num_token_doc": 66.8241, "num_token_overlap": 11.6461, "num_token_query": 31.2997, "num_token_union": 65.1453, "num_word_context": 202.4271, "num_word_doc": 49.8491, "num_word_query": 23.2469, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2588.6944, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2322, "query_norm": 1.3159, "queue_k_norm": 1.4223, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2997, "sent_len_1": 66.8241, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.5588, "stdk": 0.0485, "stdq": 0.0435, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49200 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.6444, "doc_norm": 1.4242, "encoder_q-embeddings": 1386.7883, "encoder_q-layer.0": 916.6213, "encoder_q-layer.1": 936.803, "encoder_q-layer.10": 1637.1167, "encoder_q-layer.11": 3875.4805, "encoder_q-layer.2": 1033.7756, "encoder_q-layer.3": 1090.7139, "encoder_q-layer.4": 1153.9091, "encoder_q-layer.5": 1177.3857, "encoder_q-layer.6": 1355.3453, "encoder_q-layer.7": 1454.0471, "encoder_q-layer.8": 1749.5441, "encoder_q-layer.9": 1571.6383, "epoch": 0.21, "inbatch_neg_score": 0.2282, "inbatch_pos_score": 0.8018, "learning_rate": 3.965789473684211e-05, "loss": 3.6444, "norm_diff": 0.1293, "norm_loss": 0.0, "num_token_doc": 66.8472, "num_token_overlap": 11.6934, "num_token_query": 31.3385, "num_token_union": 65.1644, "num_word_context": 202.1871, "num_word_doc": 49.9017, "num_word_query": 23.2884, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2522.6261, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2294, "query_norm": 1.2949, "queue_k_norm": 1.4227, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3385, "sent_len_1": 66.8472, "sent_len_max_0": 127.995, "sent_len_max_1": 209.9688, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49300 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.653, "doc_norm": 1.4231, "encoder_q-embeddings": 1299.1637, "encoder_q-layer.0": 882.1058, "encoder_q-layer.1": 924.3115, "encoder_q-layer.10": 1819.5081, "encoder_q-layer.11": 3963.2219, "encoder_q-layer.2": 1037.0822, "encoder_q-layer.3": 1064.1511, "encoder_q-layer.4": 1122.1307, "encoder_q-layer.5": 1123.2711, "encoder_q-layer.6": 1306.2147, "encoder_q-layer.7": 1460.0194, "encoder_q-layer.8": 1866.9791, "encoder_q-layer.9": 1575.2816, "epoch": 0.21, "inbatch_neg_score": 0.2201, "inbatch_pos_score": 0.8042, "learning_rate": 3.9631578947368426e-05, "loss": 3.653, "norm_diff": 0.1257, "norm_loss": 0.0, "num_token_doc": 66.887, "num_token_overlap": 11.6531, "num_token_query": 31.355, "num_token_union": 65.1968, "num_word_context": 202.5867, "num_word_doc": 49.8897, "num_word_query": 23.2866, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2489.6559, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2218, "query_norm": 1.2974, "queue_k_norm": 1.423, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.355, "sent_len_1": 66.887, "sent_len_max_0": 127.99, "sent_len_max_1": 208.3875, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49400 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.6391, "doc_norm": 1.4233, "encoder_q-embeddings": 1751.3075, "encoder_q-layer.0": 1236.4487, "encoder_q-layer.1": 1277.9885, "encoder_q-layer.10": 1583.413, "encoder_q-layer.11": 3613.0952, "encoder_q-layer.2": 1473.7404, "encoder_q-layer.3": 1606.1279, "encoder_q-layer.4": 1924.8977, "encoder_q-layer.5": 1975.9445, "encoder_q-layer.6": 1878.6984, "encoder_q-layer.7": 1833.6707, "encoder_q-layer.8": 1912.0192, "encoder_q-layer.9": 1532.8394, "epoch": 0.21, "inbatch_neg_score": 0.2287, "inbatch_pos_score": 0.7959, "learning_rate": 3.960526315789474e-05, "loss": 3.6391, "norm_diff": 0.1202, "norm_loss": 0.0, "num_token_doc": 66.829, "num_token_overlap": 11.6753, "num_token_query": 31.4408, "num_token_union": 65.2139, "num_word_context": 202.4378, "num_word_doc": 49.8468, "num_word_query": 23.3541, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2844.7421, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2285, "query_norm": 1.3031, "queue_k_norm": 1.4219, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4408, "sent_len_1": 66.829, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.5175, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49500 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6376, "doc_norm": 1.4227, "encoder_q-embeddings": 1326.0149, "encoder_q-layer.0": 865.1708, "encoder_q-layer.1": 913.7433, "encoder_q-layer.10": 1620.2454, "encoder_q-layer.11": 3800.6199, "encoder_q-layer.2": 1001.938, "encoder_q-layer.3": 1031.2041, "encoder_q-layer.4": 1110.3549, "encoder_q-layer.5": 1096.9348, "encoder_q-layer.6": 1236.7711, "encoder_q-layer.7": 1398.3385, "encoder_q-layer.8": 1782.6863, "encoder_q-layer.9": 1572.2792, "epoch": 0.22, "inbatch_neg_score": 0.2299, "inbatch_pos_score": 0.8184, "learning_rate": 3.9578947368421056e-05, "loss": 3.6376, "norm_diff": 0.1189, "norm_loss": 0.0, "num_token_doc": 66.9796, "num_token_overlap": 11.6754, "num_token_query": 31.3841, "num_token_union": 65.2432, "num_word_context": 202.5099, "num_word_doc": 49.96, "num_word_query": 23.3271, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2431.6807, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2302, "query_norm": 1.3038, "queue_k_norm": 1.4222, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3841, "sent_len_1": 66.9796, "sent_len_max_0": 127.9488, "sent_len_max_1": 211.0213, "stdk": 0.0484, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49600 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.6455, "doc_norm": 1.4148, "encoder_q-embeddings": 1634.8494, "encoder_q-layer.0": 1085.3842, "encoder_q-layer.1": 1109.1228, "encoder_q-layer.10": 1644.4855, "encoder_q-layer.11": 3818.9475, "encoder_q-layer.2": 1231.8666, "encoder_q-layer.3": 1247.8108, "encoder_q-layer.4": 1348.481, "encoder_q-layer.5": 1348.5687, "encoder_q-layer.6": 1437.2518, "encoder_q-layer.7": 1611.3235, "encoder_q-layer.8": 1699.0656, "encoder_q-layer.9": 1597.3572, "epoch": 0.22, "inbatch_neg_score": 0.2313, "inbatch_pos_score": 0.8042, "learning_rate": 3.955263157894737e-05, "loss": 3.6455, "norm_diff": 0.1118, "norm_loss": 0.0, "num_token_doc": 66.7321, "num_token_overlap": 11.6527, "num_token_query": 31.2906, "num_token_union": 65.0787, "num_word_context": 202.1859, "num_word_doc": 49.815, "num_word_query": 23.2449, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2617.5351, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2312, "query_norm": 1.3031, "queue_k_norm": 1.4229, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2906, "sent_len_1": 66.7321, "sent_len_max_0": 127.995, "sent_len_max_1": 208.2463, "stdk": 0.0481, "stdq": 0.0429, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49700 }, { "accuracy": 45.8984, "active_queue_size": 16384.0, "cl_loss": 3.6406, "doc_norm": 1.4167, "encoder_q-embeddings": 1324.6956, "encoder_q-layer.0": 866.0333, "encoder_q-layer.1": 903.3735, "encoder_q-layer.10": 1623.4878, "encoder_q-layer.11": 3726.7542, "encoder_q-layer.2": 998.9889, "encoder_q-layer.3": 1039.9807, "encoder_q-layer.4": 1108.6078, "encoder_q-layer.5": 1121.5652, "encoder_q-layer.6": 1214.2246, "encoder_q-layer.7": 1389.2939, "encoder_q-layer.8": 1690.6475, "encoder_q-layer.9": 1560.0676, "epoch": 0.22, "inbatch_neg_score": 0.2316, "inbatch_pos_score": 0.7881, "learning_rate": 3.9526315789473686e-05, "loss": 3.6406, "norm_diff": 0.136, "norm_loss": 0.0, "num_token_doc": 66.7026, "num_token_overlap": 11.6567, "num_token_query": 31.3547, "num_token_union": 65.0922, "num_word_context": 202.2697, "num_word_doc": 49.7775, "num_word_query": 23.2905, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2408.702, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2316, "query_norm": 1.2807, "queue_k_norm": 1.4226, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3547, "sent_len_1": 66.7026, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.2175, "stdk": 0.0482, "stdq": 0.0421, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49800 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.6496, "doc_norm": 1.4248, "encoder_q-embeddings": 1339.9943, "encoder_q-layer.0": 861.3505, "encoder_q-layer.1": 902.0754, "encoder_q-layer.10": 1915.3566, "encoder_q-layer.11": 4025.3938, "encoder_q-layer.2": 1045.1786, "encoder_q-layer.3": 1103.8761, "encoder_q-layer.4": 1158.5928, "encoder_q-layer.5": 1183.2743, "encoder_q-layer.6": 1340.8634, "encoder_q-layer.7": 1490.3032, "encoder_q-layer.8": 1815.6439, "encoder_q-layer.9": 1723.9702, "epoch": 0.22, "inbatch_neg_score": 0.2305, "inbatch_pos_score": 0.8027, "learning_rate": 3.9500000000000005e-05, "loss": 3.6496, "norm_diff": 0.1219, "norm_loss": 0.0, "num_token_doc": 66.8949, "num_token_overlap": 11.6503, "num_token_query": 31.34, "num_token_union": 65.1752, "num_word_context": 202.3482, "num_word_doc": 49.8745, "num_word_query": 23.2786, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2555.7912, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2317, "query_norm": 1.3029, "queue_k_norm": 1.4227, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.34, "sent_len_1": 66.8949, "sent_len_max_0": 127.965, "sent_len_max_1": 211.5775, "stdk": 0.0485, "stdq": 0.0428, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 49900 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.6276, "doc_norm": 1.4262, "encoder_q-embeddings": 1619.2167, "encoder_q-layer.0": 1083.8812, "encoder_q-layer.1": 1180.235, "encoder_q-layer.10": 1842.0288, "encoder_q-layer.11": 3992.4053, "encoder_q-layer.2": 1386.5199, "encoder_q-layer.3": 1418.6287, "encoder_q-layer.4": 1534.9543, "encoder_q-layer.5": 1558.345, "encoder_q-layer.6": 1610.976, "encoder_q-layer.7": 1769.0481, "encoder_q-layer.8": 1830.1207, "encoder_q-layer.9": 1699.7039, "epoch": 0.22, "inbatch_neg_score": 0.2334, "inbatch_pos_score": 0.8262, "learning_rate": 3.9473684210526316e-05, "loss": 3.6276, "norm_diff": 0.1178, "norm_loss": 0.0, "num_token_doc": 66.6939, "num_token_overlap": 11.7351, "num_token_query": 31.4913, "num_token_union": 65.0989, "num_word_context": 202.121, "num_word_doc": 49.7518, "num_word_query": 23.3979, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2763.3154, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2339, "query_norm": 1.3083, "queue_k_norm": 1.4236, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4913, "sent_len_1": 66.6939, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.3938, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50000 }, { "dev_runtime": 31.3234, "dev_samples_per_second": 1.022, "dev_steps_per_second": 0.032, "epoch": 0.22, "step": 50000, "test_accuracy": 93.51806640625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.37315481901168823, "test_doc_norm": 1.392085313796997, "test_inbatch_neg_score": 0.504136860370636, "test_inbatch_pos_score": 1.5030062198638916, "test_loss": 0.37315481901168823, "test_loss_align": 1.0806453227996826, "test_loss_unif": 3.8785688877105713, "test_loss_unif_q@queue": 3.8785688877105713, "test_norm_diff": 0.06153655797243118, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.2290598750114441, "test_query_norm": 1.4536218643188477, "test_queue_k_norm": 1.423865556716919, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.0432722270488739, "test_stdq": 0.04415716230869293, "test_stdqueue_k": 0.04850912094116211, "test_stdqueue_q": 0.0 }, { "dev_runtime": 31.3234, "dev_samples_per_second": 1.022, "dev_steps_per_second": 0.032, "epoch": 0.22, "eval_beir-arguana_ndcg@10": 0.36186, "eval_beir-arguana_recall@10": 0.61522, "eval_beir-arguana_recall@100": 0.91892, "eval_beir-arguana_recall@20": 0.75249, "eval_beir-avg_ndcg@10": 0.38594908333333333, "eval_beir-avg_recall@10": 0.45513900000000007, "eval_beir-avg_recall@100": 0.6385996666666667, "eval_beir-avg_recall@20": 0.5167944999999999, "eval_beir-cqadupstack_ndcg@10": 0.2672208333333333, "eval_beir-cqadupstack_recall@10": 0.36172000000000004, "eval_beir-cqadupstack_recall@100": 0.6009066666666667, "eval_beir-cqadupstack_recall@20": 0.429595, "eval_beir-fiqa_ndcg@10": 0.2513, "eval_beir-fiqa_recall@10": 0.3117, "eval_beir-fiqa_recall@100": 0.58456, "eval_beir-fiqa_recall@20": 0.38921, "eval_beir-nfcorpus_ndcg@10": 0.30546, "eval_beir-nfcorpus_recall@10": 0.1503, "eval_beir-nfcorpus_recall@100": 0.28889, "eval_beir-nfcorpus_recall@20": 0.18193, "eval_beir-nq_ndcg@10": 0.30007, "eval_beir-nq_recall@10": 0.48233, "eval_beir-nq_recall@100": 0.81619, "eval_beir-nq_recall@20": 0.60366, "eval_beir-quora_ndcg@10": 0.76809, "eval_beir-quora_recall@10": 0.8787, "eval_beir-quora_recall@100": 0.97659, "eval_beir-quora_recall@20": 0.92405, "eval_beir-scidocs_ndcg@10": 0.15289, "eval_beir-scidocs_recall@10": 0.16112, "eval_beir-scidocs_recall@100": 0.37468, "eval_beir-scidocs_recall@20": 0.22072, "eval_beir-scifact_ndcg@10": 0.63957, "eval_beir-scifact_recall@10": 0.79333, "eval_beir-scifact_recall@100": 0.92156, "eval_beir-scifact_recall@20": 0.82911, "eval_beir-trec-covid_ndcg@10": 0.61789, "eval_beir-trec-covid_recall@10": 0.66, "eval_beir-trec-covid_recall@100": 0.4676, "eval_beir-trec-covid_recall@20": 0.622, "eval_beir-webis-touche2020_ndcg@10": 0.19514, "eval_beir-webis-touche2020_recall@10": 0.13697, "eval_beir-webis-touche2020_recall@100": 0.4361, "eval_beir-webis-touche2020_recall@20": 0.21518, "eval_senteval-avg_sts": 0.7623288647772744, "eval_senteval-sickr_spearman": 0.7287704746060207, "eval_senteval-stsb_spearman": 0.7958872549485279, "step": 50000, "test_accuracy": 93.51806640625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.37315481901168823, "test_doc_norm": 1.392085313796997, "test_inbatch_neg_score": 0.504136860370636, "test_inbatch_pos_score": 1.5030062198638916, "test_loss": 0.37315481901168823, "test_loss_align": 1.0806453227996826, "test_loss_unif": 3.8785688877105713, "test_loss_unif_q@queue": 3.8785688877105713, "test_norm_diff": 0.06153655797243118, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.2290598750114441, "test_query_norm": 1.4536218643188477, "test_queue_k_norm": 1.423865556716919, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.0432722270488739, "test_stdq": 0.04415716230869293, "test_stdqueue_k": 0.04850912094116211, "test_stdqueue_q": 0.0 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.6543, "doc_norm": 1.4278, "encoder_q-embeddings": 1663.2047, "encoder_q-layer.0": 1156.4902, "encoder_q-layer.1": 1156.0325, "encoder_q-layer.10": 1634.6058, "encoder_q-layer.11": 3709.1223, "encoder_q-layer.2": 1237.8145, "encoder_q-layer.3": 1232.9314, "encoder_q-layer.4": 1233.1567, "encoder_q-layer.5": 1197.094, "encoder_q-layer.6": 1329.236, "encoder_q-layer.7": 1460.0613, "encoder_q-layer.8": 1721.8126, "encoder_q-layer.9": 1551.5623, "epoch": 0.22, "inbatch_neg_score": 0.2332, "inbatch_pos_score": 0.8154, "learning_rate": 3.9447368421052635e-05, "loss": 3.6543, "norm_diff": 0.1188, "norm_loss": 0.0, "num_token_doc": 66.7227, "num_token_overlap": 11.6582, "num_token_query": 31.3738, "num_token_union": 65.1129, "num_word_context": 202.4924, "num_word_doc": 49.7908, "num_word_query": 23.3145, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2564.9258, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2327, "query_norm": 1.309, "queue_k_norm": 1.4227, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3738, "sent_len_1": 66.7227, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.31, "stdk": 0.0486, "stdq": 0.0428, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50100 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.6361, "doc_norm": 1.4201, "encoder_q-embeddings": 5364.0366, "encoder_q-layer.0": 3771.5408, "encoder_q-layer.1": 4526.7866, "encoder_q-layer.10": 3640.0581, "encoder_q-layer.11": 7676.8569, "encoder_q-layer.2": 5871.9663, "encoder_q-layer.3": 6357.8784, "encoder_q-layer.4": 7386.3428, "encoder_q-layer.5": 7949.9297, "encoder_q-layer.6": 7879.2979, "encoder_q-layer.7": 7072.874, "encoder_q-layer.8": 5558.7715, "encoder_q-layer.9": 3800.7039, "epoch": 0.22, "inbatch_neg_score": 0.2328, "inbatch_pos_score": 0.8159, "learning_rate": 3.9421052631578946e-05, "loss": 3.6361, "norm_diff": 0.0874, "norm_loss": 0.0, "num_token_doc": 66.6634, "num_token_overlap": 11.6726, "num_token_query": 31.4429, "num_token_union": 65.0921, "num_word_context": 202.5374, "num_word_doc": 49.7394, "num_word_query": 23.3468, "postclip_grad_norm": 1.0, "preclip_grad_norm": 8922.8455, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2321, "query_norm": 1.3327, "queue_k_norm": 1.4214, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4429, "sent_len_1": 66.6634, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.9613, "stdk": 0.0483, "stdq": 0.0435, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 50200 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.6349, "doc_norm": 1.4253, "encoder_q-embeddings": 2907.3831, "encoder_q-layer.0": 1959.359, "encoder_q-layer.1": 2128.8284, "encoder_q-layer.10": 3219.6831, "encoder_q-layer.11": 7416.8486, "encoder_q-layer.2": 2351.168, "encoder_q-layer.3": 2424.26, "encoder_q-layer.4": 2539.2646, "encoder_q-layer.5": 2525.188, "encoder_q-layer.6": 2847.3887, "encoder_q-layer.7": 3054.9751, "encoder_q-layer.8": 3568.3164, "encoder_q-layer.9": 3279.9163, "epoch": 0.22, "inbatch_neg_score": 0.2343, "inbatch_pos_score": 0.8267, "learning_rate": 3.9394736842105265e-05, "loss": 3.6349, "norm_diff": 0.0881, "norm_loss": 0.0, "num_token_doc": 66.7811, "num_token_overlap": 11.6781, "num_token_query": 31.3932, "num_token_union": 65.1448, "num_word_context": 202.2331, "num_word_doc": 49.8305, "num_word_query": 23.3421, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5017.6867, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2341, "query_norm": 1.3373, "queue_k_norm": 1.4237, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3932, "sent_len_1": 66.7811, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.9175, "stdk": 0.0485, "stdq": 0.0435, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50300 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.6389, "doc_norm": 1.4242, "encoder_q-embeddings": 2615.9856, "encoder_q-layer.0": 1709.4092, "encoder_q-layer.1": 1789.6691, "encoder_q-layer.10": 3347.8909, "encoder_q-layer.11": 7447.9307, "encoder_q-layer.2": 2019.9657, "encoder_q-layer.3": 2042.0098, "encoder_q-layer.4": 2172.761, "encoder_q-layer.5": 2189.0117, "encoder_q-layer.6": 2524.1472, "encoder_q-layer.7": 2746.4719, "encoder_q-layer.8": 3334.032, "encoder_q-layer.9": 3060.5164, "epoch": 0.22, "inbatch_neg_score": 0.2366, "inbatch_pos_score": 0.814, "learning_rate": 3.936842105263158e-05, "loss": 3.6389, "norm_diff": 0.0939, "norm_loss": 0.0, "num_token_doc": 66.7135, "num_token_overlap": 11.6561, "num_token_query": 31.3835, "num_token_union": 65.1217, "num_word_context": 202.287, "num_word_doc": 49.8075, "num_word_query": 23.2963, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4750.4114, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2369, "query_norm": 1.3303, "queue_k_norm": 1.4227, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3835, "sent_len_1": 66.7135, "sent_len_max_0": 128.0, "sent_len_max_1": 208.0888, "stdk": 0.0485, "stdq": 0.0426, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 50400 }, { "accuracy": 46.2891, "active_queue_size": 16384.0, "cl_loss": 3.629, "doc_norm": 1.4163, "encoder_q-embeddings": 2654.9114, "encoder_q-layer.0": 1771.0425, "encoder_q-layer.1": 1818.8792, "encoder_q-layer.10": 3339.0337, "encoder_q-layer.11": 7615.2588, "encoder_q-layer.2": 2082.2161, "encoder_q-layer.3": 2119.8057, "encoder_q-layer.4": 2231.6865, "encoder_q-layer.5": 2268.5227, "encoder_q-layer.6": 2585.2434, "encoder_q-layer.7": 2942.7952, "encoder_q-layer.8": 3504.561, "encoder_q-layer.9": 3380.2266, "epoch": 0.22, "inbatch_neg_score": 0.2432, "inbatch_pos_score": 0.8096, "learning_rate": 3.9342105263157895e-05, "loss": 3.629, "norm_diff": 0.0661, "norm_loss": 0.0, "num_token_doc": 66.8699, "num_token_overlap": 11.6746, "num_token_query": 31.3775, "num_token_union": 65.2039, "num_word_context": 202.6274, "num_word_doc": 49.9132, "num_word_query": 23.3176, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4842.5469, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2422, "query_norm": 1.3502, "queue_k_norm": 1.4247, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3775, "sent_len_1": 66.8699, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.6687, "stdk": 0.0481, "stdq": 0.0431, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50500 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.6241, "doc_norm": 1.4259, "encoder_q-embeddings": 2679.011, "encoder_q-layer.0": 1751.1436, "encoder_q-layer.1": 1892.3911, "encoder_q-layer.10": 3185.116, "encoder_q-layer.11": 7429.2158, "encoder_q-layer.2": 2101.3818, "encoder_q-layer.3": 2109.4167, "encoder_q-layer.4": 2285.707, "encoder_q-layer.5": 2340.0017, "encoder_q-layer.6": 2517.3401, "encoder_q-layer.7": 2869.6658, "encoder_q-layer.8": 3666.9736, "encoder_q-layer.9": 3115.0791, "epoch": 0.22, "inbatch_neg_score": 0.2528, "inbatch_pos_score": 0.833, "learning_rate": 3.931578947368421e-05, "loss": 3.6241, "norm_diff": 0.0678, "norm_loss": 0.0, "num_token_doc": 66.7712, "num_token_overlap": 11.6822, "num_token_query": 31.318, "num_token_union": 65.0828, "num_word_context": 202.0362, "num_word_doc": 49.7955, "num_word_query": 23.2369, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4878.2987, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2534, "query_norm": 1.3581, "queue_k_norm": 1.425, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.318, "sent_len_1": 66.7712, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.7212, "stdk": 0.0485, "stdq": 0.0429, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50600 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.6193, "doc_norm": 1.4232, "encoder_q-embeddings": 3521.4736, "encoder_q-layer.0": 2509.4219, "encoder_q-layer.1": 2704.8896, "encoder_q-layer.10": 3656.9346, "encoder_q-layer.11": 7990.6865, "encoder_q-layer.2": 3113.3577, "encoder_q-layer.3": 3187.0552, "encoder_q-layer.4": 3236.1836, "encoder_q-layer.5": 3167.3921, "encoder_q-layer.6": 3206.9285, "encoder_q-layer.7": 3259.3486, "encoder_q-layer.8": 3850.1125, "encoder_q-layer.9": 3501.7654, "epoch": 0.22, "inbatch_neg_score": 0.262, "inbatch_pos_score": 0.8179, "learning_rate": 3.9289473684210524e-05, "loss": 3.6193, "norm_diff": 0.0668, "norm_loss": 0.0, "num_token_doc": 66.7038, "num_token_overlap": 11.687, "num_token_query": 31.3303, "num_token_union": 65.0626, "num_word_context": 202.1316, "num_word_doc": 49.7811, "num_word_query": 23.2797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5637.8305, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2622, "query_norm": 1.3564, "queue_k_norm": 1.4242, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3303, "sent_len_1": 66.7038, "sent_len_max_0": 127.975, "sent_len_max_1": 209.0087, "stdk": 0.0484, "stdq": 0.0422, "stdqueue_k": 0.0484, "stdqueue_q": 0.0, "step": 50700 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.6307, "doc_norm": 1.4233, "encoder_q-embeddings": 2691.0095, "encoder_q-layer.0": 1773.7103, "encoder_q-layer.1": 1859.5768, "encoder_q-layer.10": 3223.8503, "encoder_q-layer.11": 7471.4302, "encoder_q-layer.2": 2052.4722, "encoder_q-layer.3": 2115.8689, "encoder_q-layer.4": 2209.8411, "encoder_q-layer.5": 2176.4541, "encoder_q-layer.6": 2571.1089, "encoder_q-layer.7": 2760.1174, "encoder_q-layer.8": 3310.6423, "encoder_q-layer.9": 3099.9983, "epoch": 0.22, "inbatch_neg_score": 0.2761, "inbatch_pos_score": 0.855, "learning_rate": 3.926315789473684e-05, "loss": 3.6307, "norm_diff": 0.0486, "norm_loss": 0.0, "num_token_doc": 66.8886, "num_token_overlap": 11.6517, "num_token_query": 31.2661, "num_token_union": 65.1168, "num_word_context": 202.3536, "num_word_doc": 49.9078, "num_word_query": 23.2113, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4799.0359, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2754, "query_norm": 1.3747, "queue_k_norm": 1.4273, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2661, "sent_len_1": 66.8886, "sent_len_max_0": 127.9887, "sent_len_max_1": 211.5238, "stdk": 0.0483, "stdq": 0.0429, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50800 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.6246, "doc_norm": 1.4298, "encoder_q-embeddings": 8672.6875, "encoder_q-layer.0": 5926.9194, "encoder_q-layer.1": 6063.4492, "encoder_q-layer.10": 3318.3777, "encoder_q-layer.11": 7768.5073, "encoder_q-layer.2": 7710.7769, "encoder_q-layer.3": 7570.3086, "encoder_q-layer.4": 8187.6362, "encoder_q-layer.5": 8252.9121, "encoder_q-layer.6": 8362.9121, "encoder_q-layer.7": 7618.5688, "encoder_q-layer.8": 5912.1792, "encoder_q-layer.9": 3857.5359, "epoch": 0.22, "inbatch_neg_score": 0.2841, "inbatch_pos_score": 0.8545, "learning_rate": 3.923684210526316e-05, "loss": 3.6246, "norm_diff": 0.0697, "norm_loss": 0.0, "num_token_doc": 66.9094, "num_token_overlap": 11.7406, "num_token_query": 31.5694, "num_token_union": 65.2531, "num_word_context": 202.546, "num_word_doc": 49.9501, "num_word_query": 23.4757, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10607.5517, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2834, "query_norm": 1.3601, "queue_k_norm": 1.4288, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5694, "sent_len_1": 66.9094, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.3562, "stdk": 0.0485, "stdq": 0.0429, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 50900 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.6404, "doc_norm": 1.4306, "encoder_q-embeddings": 2630.8843, "encoder_q-layer.0": 1808.9183, "encoder_q-layer.1": 1912.5117, "encoder_q-layer.10": 3522.5745, "encoder_q-layer.11": 7703.772, "encoder_q-layer.2": 2056.6501, "encoder_q-layer.3": 2087.6118, "encoder_q-layer.4": 2224.2292, "encoder_q-layer.5": 2256.0239, "encoder_q-layer.6": 2672.4167, "encoder_q-layer.7": 3020.9812, "encoder_q-layer.8": 3669.6602, "encoder_q-layer.9": 3354.0735, "epoch": 0.22, "inbatch_neg_score": 0.2896, "inbatch_pos_score": 0.877, "learning_rate": 3.921052631578947e-05, "loss": 3.6404, "norm_diff": 0.0694, "norm_loss": 0.0, "num_token_doc": 67.1028, "num_token_overlap": 11.7018, "num_token_query": 31.4322, "num_token_union": 65.3772, "num_word_context": 202.7879, "num_word_doc": 50.0868, "num_word_query": 23.3611, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4991.2604, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2905, "query_norm": 1.3612, "queue_k_norm": 1.4298, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4322, "sent_len_1": 67.1028, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.0938, "stdk": 0.0485, "stdq": 0.0433, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 51000 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.6318, "doc_norm": 1.4314, "encoder_q-embeddings": 2876.6526, "encoder_q-layer.0": 1933.0571, "encoder_q-layer.1": 2062.1675, "encoder_q-layer.10": 3807.6267, "encoder_q-layer.11": 7659.4189, "encoder_q-layer.2": 2283.0156, "encoder_q-layer.3": 2353.9385, "encoder_q-layer.4": 2544.9404, "encoder_q-layer.5": 2537.739, "encoder_q-layer.6": 2864.8132, "encoder_q-layer.7": 3407.4487, "encoder_q-layer.8": 3913.3193, "encoder_q-layer.9": 3593.1497, "epoch": 0.22, "inbatch_neg_score": 0.2892, "inbatch_pos_score": 0.8872, "learning_rate": 3.918421052631579e-05, "loss": 3.6318, "norm_diff": 0.0593, "norm_loss": 0.0, "num_token_doc": 66.9322, "num_token_overlap": 11.6769, "num_token_query": 31.2916, "num_token_union": 65.1435, "num_word_context": 202.2577, "num_word_doc": 49.9007, "num_word_query": 23.2345, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5130.0664, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2886, "query_norm": 1.3722, "queue_k_norm": 1.4321, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2916, "sent_len_1": 66.9322, "sent_len_max_0": 127.99, "sent_len_max_1": 209.1887, "stdk": 0.0485, "stdq": 0.0444, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 51100 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.6225, "doc_norm": 1.438, "encoder_q-embeddings": 2604.3032, "encoder_q-layer.0": 1759.6146, "encoder_q-layer.1": 1886.9248, "encoder_q-layer.10": 3367.2917, "encoder_q-layer.11": 7748.6255, "encoder_q-layer.2": 2182.884, "encoder_q-layer.3": 2135.4192, "encoder_q-layer.4": 2261.321, "encoder_q-layer.5": 2239.8945, "encoder_q-layer.6": 2510.2161, "encoder_q-layer.7": 2755.5559, "encoder_q-layer.8": 3496.8538, "encoder_q-layer.9": 3184.0996, "epoch": 0.22, "inbatch_neg_score": 0.2863, "inbatch_pos_score": 0.876, "learning_rate": 3.91578947368421e-05, "loss": 3.6225, "norm_diff": 0.0985, "norm_loss": 0.0, "num_token_doc": 66.9033, "num_token_overlap": 11.7388, "num_token_query": 31.5404, "num_token_union": 65.2615, "num_word_context": 202.5976, "num_word_doc": 49.9438, "num_word_query": 23.4271, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4807.7416, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2864, "query_norm": 1.3395, "queue_k_norm": 1.4341, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5404, "sent_len_1": 66.9033, "sent_len_max_0": 127.995, "sent_len_max_1": 208.0362, "stdk": 0.0487, "stdq": 0.0434, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 51200 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.6334, "doc_norm": 1.4357, "encoder_q-embeddings": 3002.7764, "encoder_q-layer.0": 1983.0372, "encoder_q-layer.1": 2116.2537, "encoder_q-layer.10": 3468.6846, "encoder_q-layer.11": 7661.7256, "encoder_q-layer.2": 2379.0063, "encoder_q-layer.3": 2472.1501, "encoder_q-layer.4": 2649.2129, "encoder_q-layer.5": 2758.7651, "encoder_q-layer.6": 3077.1306, "encoder_q-layer.7": 3809.5505, "encoder_q-layer.8": 4292.373, "encoder_q-layer.9": 3548.8779, "epoch": 0.22, "inbatch_neg_score": 0.2847, "inbatch_pos_score": 0.855, "learning_rate": 3.913157894736843e-05, "loss": 3.6334, "norm_diff": 0.1144, "norm_loss": 0.0, "num_token_doc": 66.8246, "num_token_overlap": 11.6917, "num_token_query": 31.3529, "num_token_union": 65.1045, "num_word_context": 202.1574, "num_word_doc": 49.8647, "num_word_query": 23.2859, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5296.0962, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2839, "query_norm": 1.3213, "queue_k_norm": 1.4357, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3529, "sent_len_1": 66.8246, "sent_len_max_0": 127.9688, "sent_len_max_1": 208.94, "stdk": 0.0485, "stdq": 0.0429, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 51300 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.627, "doc_norm": 1.4365, "encoder_q-embeddings": 2852.6614, "encoder_q-layer.0": 1866.7351, "encoder_q-layer.1": 1978.1868, "encoder_q-layer.10": 3348.9219, "encoder_q-layer.11": 7777.8613, "encoder_q-layer.2": 2230.2422, "encoder_q-layer.3": 2182.5479, "encoder_q-layer.4": 2228.6074, "encoder_q-layer.5": 2264.51, "encoder_q-layer.6": 2580.3296, "encoder_q-layer.7": 3004.748, "encoder_q-layer.8": 3591.2495, "encoder_q-layer.9": 3457.7139, "epoch": 0.22, "inbatch_neg_score": 0.2875, "inbatch_pos_score": 0.8555, "learning_rate": 3.910526315789474e-05, "loss": 3.627, "norm_diff": 0.12, "norm_loss": 0.0, "num_token_doc": 66.822, "num_token_overlap": 11.6814, "num_token_query": 31.3316, "num_token_union": 65.0943, "num_word_context": 202.4865, "num_word_doc": 49.8125, "num_word_query": 23.2634, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5037.7255, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2869, "query_norm": 1.3166, "queue_k_norm": 1.4352, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3316, "sent_len_1": 66.822, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.2413, "stdk": 0.0485, "stdq": 0.0426, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 51400 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.6372, "doc_norm": 1.436, "encoder_q-embeddings": 2562.3459, "encoder_q-layer.0": 1739.6104, "encoder_q-layer.1": 1761.9529, "encoder_q-layer.10": 3325.9524, "encoder_q-layer.11": 7840.0049, "encoder_q-layer.2": 2015.2244, "encoder_q-layer.3": 2071.8691, "encoder_q-layer.4": 2176.5989, "encoder_q-layer.5": 2248.7451, "encoder_q-layer.6": 2610.6743, "encoder_q-layer.7": 2749.7778, "encoder_q-layer.8": 3387.6655, "encoder_q-layer.9": 3156.2766, "epoch": 0.22, "inbatch_neg_score": 0.2838, "inbatch_pos_score": 0.8774, "learning_rate": 3.907894736842105e-05, "loss": 3.6372, "norm_diff": 0.1058, "norm_loss": 0.0, "num_token_doc": 66.7351, "num_token_overlap": 11.666, "num_token_query": 31.3048, "num_token_union": 65.0314, "num_word_context": 202.4065, "num_word_doc": 49.7828, "num_word_query": 23.2482, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4848.4635, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2849, "query_norm": 1.3303, "queue_k_norm": 1.4367, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3048, "sent_len_1": 66.7351, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.4325, "stdk": 0.0485, "stdq": 0.0433, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 51500 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.6263, "doc_norm": 1.4339, "encoder_q-embeddings": 3999.9243, "encoder_q-layer.0": 2910.9502, "encoder_q-layer.1": 3312.5671, "encoder_q-layer.10": 3339.261, "encoder_q-layer.11": 7434.4824, "encoder_q-layer.2": 3030.3291, "encoder_q-layer.3": 2915.8823, "encoder_q-layer.4": 2878.3232, "encoder_q-layer.5": 2678.4314, "encoder_q-layer.6": 2756.9597, "encoder_q-layer.7": 2930.157, "encoder_q-layer.8": 3368.4307, "encoder_q-layer.9": 3127.5747, "epoch": 0.22, "inbatch_neg_score": 0.2882, "inbatch_pos_score": 0.8696, "learning_rate": 3.905263157894737e-05, "loss": 3.6263, "norm_diff": 0.103, "norm_loss": 0.0, "num_token_doc": 66.7423, "num_token_overlap": 11.6755, "num_token_query": 31.3943, "num_token_union": 65.1108, "num_word_context": 202.2919, "num_word_doc": 49.8141, "num_word_query": 23.3301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5476.896, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2883, "query_norm": 1.3309, "queue_k_norm": 1.4384, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3943, "sent_len_1": 66.7423, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.5137, "stdk": 0.0483, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 51600 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.6254, "doc_norm": 1.44, "encoder_q-embeddings": 3104.5071, "encoder_q-layer.0": 2181.2524, "encoder_q-layer.1": 2292.8823, "encoder_q-layer.10": 3394.0493, "encoder_q-layer.11": 7766.481, "encoder_q-layer.2": 2686.1707, "encoder_q-layer.3": 2753.5654, "encoder_q-layer.4": 2870.4458, "encoder_q-layer.5": 2844.803, "encoder_q-layer.6": 3156.7725, "encoder_q-layer.7": 3457.416, "encoder_q-layer.8": 3746.905, "encoder_q-layer.9": 3250.4509, "epoch": 0.22, "inbatch_neg_score": 0.2893, "inbatch_pos_score": 0.8779, "learning_rate": 3.902631578947369e-05, "loss": 3.6254, "norm_diff": 0.0968, "norm_loss": 0.0, "num_token_doc": 66.5714, "num_token_overlap": 11.6299, "num_token_query": 31.2125, "num_token_union": 64.9385, "num_word_context": 202.058, "num_word_doc": 49.6955, "num_word_query": 23.165, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5379.5522, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.29, "query_norm": 1.3432, "queue_k_norm": 1.4375, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2125, "sent_len_1": 66.5714, "sent_len_max_0": 127.985, "sent_len_max_1": 206.6513, "stdk": 0.0485, "stdq": 0.0436, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 51700 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.62, "doc_norm": 1.4398, "encoder_q-embeddings": 3039.7253, "encoder_q-layer.0": 2145.8943, "encoder_q-layer.1": 2234.9312, "encoder_q-layer.10": 3145.0007, "encoder_q-layer.11": 7385.8389, "encoder_q-layer.2": 2489.3394, "encoder_q-layer.3": 2588.7527, "encoder_q-layer.4": 2751.3901, "encoder_q-layer.5": 2724.6545, "encoder_q-layer.6": 2913.2166, "encoder_q-layer.7": 3107.5139, "encoder_q-layer.8": 3512.3501, "encoder_q-layer.9": 3077.3682, "epoch": 0.22, "inbatch_neg_score": 0.2839, "inbatch_pos_score": 0.8657, "learning_rate": 3.9000000000000006e-05, "loss": 3.62, "norm_diff": 0.1219, "norm_loss": 0.0, "num_token_doc": 66.7585, "num_token_overlap": 11.6697, "num_token_query": 31.3802, "num_token_union": 65.0897, "num_word_context": 202.182, "num_word_doc": 49.8029, "num_word_query": 23.3084, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5102.6492, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2834, "query_norm": 1.3179, "queue_k_norm": 1.4396, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3802, "sent_len_1": 66.7585, "sent_len_max_0": 127.945, "sent_len_max_1": 208.42, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 51800 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.6365, "doc_norm": 1.4429, "encoder_q-embeddings": 2599.3521, "encoder_q-layer.0": 1760.7561, "encoder_q-layer.1": 1878.2678, "encoder_q-layer.10": 3266.054, "encoder_q-layer.11": 7526.5566, "encoder_q-layer.2": 2053.5903, "encoder_q-layer.3": 2083.4858, "encoder_q-layer.4": 2237.6394, "encoder_q-layer.5": 2242.0757, "encoder_q-layer.6": 2498.8254, "encoder_q-layer.7": 2801.9553, "encoder_q-layer.8": 3427.7168, "encoder_q-layer.9": 3138.8967, "epoch": 0.23, "inbatch_neg_score": 0.2836, "inbatch_pos_score": 0.856, "learning_rate": 3.897368421052632e-05, "loss": 3.6365, "norm_diff": 0.1288, "norm_loss": 0.0, "num_token_doc": 66.7, "num_token_overlap": 11.6706, "num_token_query": 31.3157, "num_token_union": 65.0607, "num_word_context": 202.2427, "num_word_doc": 49.8004, "num_word_query": 23.2646, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4866.8415, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.283, "query_norm": 1.3142, "queue_k_norm": 1.441, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3157, "sent_len_1": 66.7, "sent_len_max_0": 127.965, "sent_len_max_1": 206.7438, "stdk": 0.0486, "stdq": 0.0426, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 51900 }, { "accuracy": 46.6309, "active_queue_size": 16384.0, "cl_loss": 3.6254, "doc_norm": 1.4413, "encoder_q-embeddings": 2819.8394, "encoder_q-layer.0": 1920.9208, "encoder_q-layer.1": 1946.3239, "encoder_q-layer.10": 3470.9956, "encoder_q-layer.11": 7900.0513, "encoder_q-layer.2": 2152.0623, "encoder_q-layer.3": 2148.1873, "encoder_q-layer.4": 2254.2844, "encoder_q-layer.5": 2295.9526, "encoder_q-layer.6": 2662.2795, "encoder_q-layer.7": 2842.8264, "encoder_q-layer.8": 3425.0725, "encoder_q-layer.9": 3193.0356, "epoch": 0.23, "inbatch_neg_score": 0.2844, "inbatch_pos_score": 0.8447, "learning_rate": 3.894736842105263e-05, "loss": 3.6254, "norm_diff": 0.1264, "norm_loss": 0.0, "num_token_doc": 66.8854, "num_token_overlap": 11.6675, "num_token_query": 31.3627, "num_token_union": 65.1589, "num_word_context": 202.4089, "num_word_doc": 49.901, "num_word_query": 23.2874, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5055.6856, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2842, "query_norm": 1.315, "queue_k_norm": 1.4425, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3627, "sent_len_1": 66.8854, "sent_len_max_0": 127.99, "sent_len_max_1": 208.2075, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52000 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.6322, "doc_norm": 1.4384, "encoder_q-embeddings": 2709.1409, "encoder_q-layer.0": 1783.5483, "encoder_q-layer.1": 1850.181, "encoder_q-layer.10": 3181.7168, "encoder_q-layer.11": 7701.5967, "encoder_q-layer.2": 2049.7935, "encoder_q-layer.3": 2133.1418, "encoder_q-layer.4": 2247.8591, "encoder_q-layer.5": 2325.478, "encoder_q-layer.6": 2646.895, "encoder_q-layer.7": 3110.252, "encoder_q-layer.8": 3669.4617, "encoder_q-layer.9": 3134.5806, "epoch": 0.23, "inbatch_neg_score": 0.2842, "inbatch_pos_score": 0.8613, "learning_rate": 3.892105263157895e-05, "loss": 3.6322, "norm_diff": 0.1204, "norm_loss": 0.0, "num_token_doc": 66.7645, "num_token_overlap": 11.6889, "num_token_query": 31.4137, "num_token_union": 65.1135, "num_word_context": 202.2459, "num_word_doc": 49.8269, "num_word_query": 23.3249, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4992.3571, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.2822, "query_norm": 1.3181, "queue_k_norm": 1.4409, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4137, "sent_len_1": 66.7645, "sent_len_max_0": 127.99, "sent_len_max_1": 208.8875, "stdk": 0.0484, "stdq": 0.0428, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52100 }, { "accuracy": 47.0703, "active_queue_size": 16384.0, "cl_loss": 3.6209, "doc_norm": 1.4386, "encoder_q-embeddings": 5862.7744, "encoder_q-layer.0": 3877.9895, "encoder_q-layer.1": 4268.7197, "encoder_q-layer.10": 7202.2715, "encoder_q-layer.11": 16917.3262, "encoder_q-layer.2": 4644.0586, "encoder_q-layer.3": 4828.562, "encoder_q-layer.4": 5191.7095, "encoder_q-layer.5": 5289.0288, "encoder_q-layer.6": 5827.0024, "encoder_q-layer.7": 6200.2173, "encoder_q-layer.8": 7395.561, "encoder_q-layer.9": 6646.8457, "epoch": 0.23, "inbatch_neg_score": 0.2874, "inbatch_pos_score": 0.8721, "learning_rate": 3.8894736842105266e-05, "loss": 3.6209, "norm_diff": 0.108, "norm_loss": 0.0, "num_token_doc": 66.7204, "num_token_overlap": 11.6668, "num_token_query": 31.3315, "num_token_union": 65.0505, "num_word_context": 202.3808, "num_word_doc": 49.7885, "num_word_query": 23.2659, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10780.0568, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2878, "query_norm": 1.3307, "queue_k_norm": 1.4404, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3315, "sent_len_1": 66.7204, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.6575, "stdk": 0.0484, "stdq": 0.0432, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 52200 }, { "accuracy": 46.9727, "active_queue_size": 16384.0, "cl_loss": 3.6118, "doc_norm": 1.4403, "encoder_q-embeddings": 9218.4639, "encoder_q-layer.0": 6350.627, "encoder_q-layer.1": 6683.5776, "encoder_q-layer.10": 6545.3442, "encoder_q-layer.11": 14967.9043, "encoder_q-layer.2": 8687.8965, "encoder_q-layer.3": 8556.7344, "encoder_q-layer.4": 7161.8096, "encoder_q-layer.5": 7252.7412, "encoder_q-layer.6": 7966.8926, "encoder_q-layer.7": 7352.2363, "encoder_q-layer.8": 7640.8638, "encoder_q-layer.9": 6667.1538, "epoch": 0.23, "inbatch_neg_score": 0.2885, "inbatch_pos_score": 0.8569, "learning_rate": 3.8868421052631584e-05, "loss": 3.6118, "norm_diff": 0.1112, "norm_loss": 0.0, "num_token_doc": 66.9399, "num_token_overlap": 11.7222, "num_token_query": 31.4743, "num_token_union": 65.2435, "num_word_context": 202.2463, "num_word_doc": 49.9471, "num_word_query": 23.3832, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12582.1152, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2878, "query_norm": 1.3291, "queue_k_norm": 1.4434, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4743, "sent_len_1": 66.9399, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.1887, "stdk": 0.0484, "stdq": 0.0432, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52300 }, { "accuracy": 46.7285, "active_queue_size": 16384.0, "cl_loss": 3.6321, "doc_norm": 1.4448, "encoder_q-embeddings": 5871.1318, "encoder_q-layer.0": 3771.7744, "encoder_q-layer.1": 3912.0913, "encoder_q-layer.10": 6945.2358, "encoder_q-layer.11": 16295.4678, "encoder_q-layer.2": 4404.2749, "encoder_q-layer.3": 4527.8179, "encoder_q-layer.4": 4847.2427, "encoder_q-layer.5": 5019.6475, "encoder_q-layer.6": 5549.667, "encoder_q-layer.7": 6258.4067, "encoder_q-layer.8": 6926.5234, "encoder_q-layer.9": 6174.7554, "epoch": 0.23, "inbatch_neg_score": 0.2888, "inbatch_pos_score": 0.8599, "learning_rate": 3.8842105263157896e-05, "loss": 3.6321, "norm_diff": 0.1263, "norm_loss": 0.0, "num_token_doc": 66.882, "num_token_overlap": 11.6902, "num_token_query": 31.3262, "num_token_union": 65.1328, "num_word_context": 202.4056, "num_word_doc": 49.9033, "num_word_query": 23.2593, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10420.4587, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2891, "query_norm": 1.3184, "queue_k_norm": 1.4431, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3262, "sent_len_1": 66.882, "sent_len_max_0": 127.985, "sent_len_max_1": 209.8288, "stdk": 0.0486, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52400 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.6241, "doc_norm": 1.446, "encoder_q-embeddings": 5196.5645, "encoder_q-layer.0": 3451.4109, "encoder_q-layer.1": 3735.0271, "encoder_q-layer.10": 6180.6963, "encoder_q-layer.11": 14975.5596, "encoder_q-layer.2": 4119.3428, "encoder_q-layer.3": 4174.5742, "encoder_q-layer.4": 4322.4829, "encoder_q-layer.5": 4391.4375, "encoder_q-layer.6": 4905.4419, "encoder_q-layer.7": 5584.8281, "encoder_q-layer.8": 7087.0674, "encoder_q-layer.9": 6096.1978, "epoch": 0.23, "inbatch_neg_score": 0.2873, "inbatch_pos_score": 0.873, "learning_rate": 3.8815789473684214e-05, "loss": 3.6241, "norm_diff": 0.1237, "norm_loss": 0.0, "num_token_doc": 66.6319, "num_token_overlap": 11.7, "num_token_query": 31.4689, "num_token_union": 65.0783, "num_word_context": 202.2157, "num_word_doc": 49.7514, "num_word_query": 23.3842, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9647.1887, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2869, "query_norm": 1.3223, "queue_k_norm": 1.4427, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4689, "sent_len_1": 66.6319, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.6825, "stdk": 0.0486, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52500 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.6262, "doc_norm": 1.4443, "encoder_q-embeddings": 5442.1157, "encoder_q-layer.0": 3683.624, "encoder_q-layer.1": 3953.8875, "encoder_q-layer.10": 6915.1445, "encoder_q-layer.11": 15468.1279, "encoder_q-layer.2": 4412.3237, "encoder_q-layer.3": 4618.8955, "encoder_q-layer.4": 5011.0566, "encoder_q-layer.5": 5005.2627, "encoder_q-layer.6": 5492.624, "encoder_q-layer.7": 6032.4404, "encoder_q-layer.8": 7180.9893, "encoder_q-layer.9": 6589.5327, "epoch": 0.23, "inbatch_neg_score": 0.2855, "inbatch_pos_score": 0.855, "learning_rate": 3.8789473684210526e-05, "loss": 3.6262, "norm_diff": 0.1272, "norm_loss": 0.0, "num_token_doc": 66.6852, "num_token_overlap": 11.6844, "num_token_query": 31.3526, "num_token_union": 65.0383, "num_word_context": 202.0307, "num_word_doc": 49.7716, "num_word_query": 23.2742, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10113.1234, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2852, "query_norm": 1.3171, "queue_k_norm": 1.4438, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3526, "sent_len_1": 66.6852, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.635, "stdk": 0.0486, "stdq": 0.0428, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52600 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.6288, "doc_norm": 1.4448, "encoder_q-embeddings": 7377.0981, "encoder_q-layer.0": 5302.9585, "encoder_q-layer.1": 5550.6494, "encoder_q-layer.10": 6593.0449, "encoder_q-layer.11": 14877.0508, "encoder_q-layer.2": 6353.0132, "encoder_q-layer.3": 6919.0991, "encoder_q-layer.4": 7315.5112, "encoder_q-layer.5": 6403.8359, "encoder_q-layer.6": 6746.1562, "encoder_q-layer.7": 6451.165, "encoder_q-layer.8": 7137.7515, "encoder_q-layer.9": 6300.6436, "epoch": 0.23, "inbatch_neg_score": 0.2862, "inbatch_pos_score": 0.8667, "learning_rate": 3.8763157894736844e-05, "loss": 3.6288, "norm_diff": 0.1223, "norm_loss": 0.0, "num_token_doc": 67.023, "num_token_overlap": 11.6771, "num_token_query": 31.3787, "num_token_union": 65.2496, "num_word_context": 202.8646, "num_word_doc": 49.9844, "num_word_query": 23.3257, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11171.4655, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2859, "query_norm": 1.3226, "queue_k_norm": 1.4439, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3787, "sent_len_1": 67.023, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.6525, "stdk": 0.0486, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52700 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.6094, "doc_norm": 1.4424, "encoder_q-embeddings": 5461.5542, "encoder_q-layer.0": 3637.8218, "encoder_q-layer.1": 3818.0105, "encoder_q-layer.10": 6375.2104, "encoder_q-layer.11": 14329.208, "encoder_q-layer.2": 4285.5142, "encoder_q-layer.3": 4311.1274, "encoder_q-layer.4": 4578.1035, "encoder_q-layer.5": 4509.7197, "encoder_q-layer.6": 5198.5605, "encoder_q-layer.7": 5890.6807, "encoder_q-layer.8": 6757.2188, "encoder_q-layer.9": 6078.6279, "epoch": 0.23, "inbatch_neg_score": 0.29, "inbatch_pos_score": 0.8647, "learning_rate": 3.873684210526316e-05, "loss": 3.6094, "norm_diff": 0.1283, "norm_loss": 0.0, "num_token_doc": 66.7259, "num_token_overlap": 11.6869, "num_token_query": 31.3575, "num_token_union": 65.0487, "num_word_context": 202.2253, "num_word_doc": 49.7934, "num_word_query": 23.2962, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9566.8847, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2893, "query_norm": 1.3142, "queue_k_norm": 1.4437, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3575, "sent_len_1": 66.7259, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.6062, "stdk": 0.0485, "stdq": 0.0426, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52800 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.6205, "doc_norm": 1.4445, "encoder_q-embeddings": 5556.5962, "encoder_q-layer.0": 3742.385, "encoder_q-layer.1": 3824.5532, "encoder_q-layer.10": 6908.1592, "encoder_q-layer.11": 15794.2646, "encoder_q-layer.2": 4322.8682, "encoder_q-layer.3": 4461.5146, "encoder_q-layer.4": 4735.0083, "encoder_q-layer.5": 4780.9624, "encoder_q-layer.6": 5395.2202, "encoder_q-layer.7": 6040.0713, "encoder_q-layer.8": 7300.6802, "encoder_q-layer.9": 6532.7939, "epoch": 0.23, "inbatch_neg_score": 0.2867, "inbatch_pos_score": 0.8521, "learning_rate": 3.8710526315789474e-05, "loss": 3.6205, "norm_diff": 0.144, "norm_loss": 0.0, "num_token_doc": 66.749, "num_token_overlap": 11.7019, "num_token_query": 31.3888, "num_token_union": 65.0762, "num_word_context": 202.0255, "num_word_doc": 49.7927, "num_word_query": 23.3101, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10185.4543, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2869, "query_norm": 1.3005, "queue_k_norm": 1.4441, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3888, "sent_len_1": 66.749, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.1962, "stdk": 0.0485, "stdq": 0.0421, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 52900 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.6182, "doc_norm": 1.4469, "encoder_q-embeddings": 5415.6846, "encoder_q-layer.0": 3486.8345, "encoder_q-layer.1": 3758.0242, "encoder_q-layer.10": 6727.3647, "encoder_q-layer.11": 15156.9199, "encoder_q-layer.2": 4117.7305, "encoder_q-layer.3": 4073.9666, "encoder_q-layer.4": 4383.9585, "encoder_q-layer.5": 4488.9385, "encoder_q-layer.6": 5158.2954, "encoder_q-layer.7": 5898.8589, "encoder_q-layer.8": 7346.8691, "encoder_q-layer.9": 6297.4868, "epoch": 0.23, "inbatch_neg_score": 0.2878, "inbatch_pos_score": 0.8687, "learning_rate": 3.868421052631579e-05, "loss": 3.6182, "norm_diff": 0.1276, "norm_loss": 0.0, "num_token_doc": 66.7593, "num_token_overlap": 11.704, "num_token_query": 31.4765, "num_token_union": 65.1735, "num_word_context": 202.3764, "num_word_doc": 49.8224, "num_word_query": 23.3837, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9847.4356, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2881, "query_norm": 1.3194, "queue_k_norm": 1.4436, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4765, "sent_len_1": 66.7593, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.9462, "stdk": 0.0486, "stdq": 0.0428, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 53000 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.619, "doc_norm": 1.4403, "encoder_q-embeddings": 5059.9468, "encoder_q-layer.0": 3518.6494, "encoder_q-layer.1": 3566.0442, "encoder_q-layer.10": 7233.4326, "encoder_q-layer.11": 15570.1602, "encoder_q-layer.2": 3976.2229, "encoder_q-layer.3": 4004.3394, "encoder_q-layer.4": 4341.5537, "encoder_q-layer.5": 4530.9595, "encoder_q-layer.6": 5017.4219, "encoder_q-layer.7": 5442.2305, "encoder_q-layer.8": 7117.9302, "encoder_q-layer.9": 6665.0684, "epoch": 0.23, "inbatch_neg_score": 0.2872, "inbatch_pos_score": 0.8711, "learning_rate": 3.865789473684211e-05, "loss": 3.619, "norm_diff": 0.1058, "norm_loss": 0.0, "num_token_doc": 66.8191, "num_token_overlap": 11.6803, "num_token_query": 31.2685, "num_token_union": 65.0807, "num_word_context": 202.0281, "num_word_doc": 49.8375, "num_word_query": 23.2196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9712.1016, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2878, "query_norm": 1.3345, "queue_k_norm": 1.4422, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2685, "sent_len_1": 66.8191, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.6175, "stdk": 0.0484, "stdq": 0.0433, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 53100 }, { "accuracy": 47.5586, "active_queue_size": 16384.0, "cl_loss": 3.6213, "doc_norm": 1.4463, "encoder_q-embeddings": 6508.2524, "encoder_q-layer.0": 4398.0698, "encoder_q-layer.1": 4908.6616, "encoder_q-layer.10": 6584.7568, "encoder_q-layer.11": 14918.4785, "encoder_q-layer.2": 5309.3306, "encoder_q-layer.3": 5717.6992, "encoder_q-layer.4": 6344.0742, "encoder_q-layer.5": 6427.3164, "encoder_q-layer.6": 6858.9492, "encoder_q-layer.7": 6950.9233, "encoder_q-layer.8": 7597.8179, "encoder_q-layer.9": 6466.897, "epoch": 0.23, "inbatch_neg_score": 0.2903, "inbatch_pos_score": 0.875, "learning_rate": 3.863157894736842e-05, "loss": 3.6213, "norm_diff": 0.1021, "norm_loss": 0.0, "num_token_doc": 66.7278, "num_token_overlap": 11.6289, "num_token_query": 31.3759, "num_token_union": 65.1525, "num_word_context": 202.6511, "num_word_doc": 49.7874, "num_word_query": 23.3032, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10882.1005, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2898, "query_norm": 1.3442, "queue_k_norm": 1.4456, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3759, "sent_len_1": 66.7278, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.1287, "stdk": 0.0486, "stdq": 0.0436, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53200 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.6132, "doc_norm": 1.4418, "encoder_q-embeddings": 5874.998, "encoder_q-layer.0": 3939.5659, "encoder_q-layer.1": 4330.7476, "encoder_q-layer.10": 6739.0791, "encoder_q-layer.11": 15378.8887, "encoder_q-layer.2": 4702.7554, "encoder_q-layer.3": 4755.5669, "encoder_q-layer.4": 5053.0781, "encoder_q-layer.5": 5166.7607, "encoder_q-layer.6": 5646.8262, "encoder_q-layer.7": 6214.4863, "encoder_q-layer.8": 7172.7148, "encoder_q-layer.9": 6492.0444, "epoch": 0.23, "inbatch_neg_score": 0.2885, "inbatch_pos_score": 0.8613, "learning_rate": 3.860526315789474e-05, "loss": 3.6132, "norm_diff": 0.1132, "norm_loss": 0.0, "num_token_doc": 66.7365, "num_token_overlap": 11.6846, "num_token_query": 31.4591, "num_token_union": 65.1654, "num_word_context": 202.1566, "num_word_doc": 49.784, "num_word_query": 23.3632, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10353.3909, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2888, "query_norm": 1.3286, "queue_k_norm": 1.4453, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4591, "sent_len_1": 66.7365, "sent_len_max_0": 127.995, "sent_len_max_1": 209.1838, "stdk": 0.0484, "stdq": 0.0428, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53300 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.6004, "doc_norm": 1.4412, "encoder_q-embeddings": 5882.8394, "encoder_q-layer.0": 3651.1274, "encoder_q-layer.1": 3990.9592, "encoder_q-layer.10": 6673.4634, "encoder_q-layer.11": 14883.0068, "encoder_q-layer.2": 4384.8833, "encoder_q-layer.3": 4576.8081, "encoder_q-layer.4": 4849.728, "encoder_q-layer.5": 5071.6826, "encoder_q-layer.6": 5668.2158, "encoder_q-layer.7": 6772.3418, "encoder_q-layer.8": 7742.4668, "encoder_q-layer.9": 6586.1729, "epoch": 0.23, "inbatch_neg_score": 0.296, "inbatch_pos_score": 0.876, "learning_rate": 3.857894736842105e-05, "loss": 3.6004, "norm_diff": 0.1005, "norm_loss": 0.0, "num_token_doc": 66.716, "num_token_overlap": 11.6825, "num_token_query": 31.3674, "num_token_union": 65.0913, "num_word_context": 202.3506, "num_word_doc": 49.8003, "num_word_query": 23.3051, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10201.1626, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2959, "query_norm": 1.3407, "queue_k_norm": 1.447, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3674, "sent_len_1": 66.716, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.6525, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53400 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.589, "doc_norm": 1.4435, "encoder_q-embeddings": 5374.6631, "encoder_q-layer.0": 3461.2356, "encoder_q-layer.1": 3608.5728, "encoder_q-layer.10": 6609.9087, "encoder_q-layer.11": 14836.7852, "encoder_q-layer.2": 4035.1372, "encoder_q-layer.3": 4133.5151, "encoder_q-layer.4": 4329.9727, "encoder_q-layer.5": 4450.0259, "encoder_q-layer.6": 5051.998, "encoder_q-layer.7": 5605.3784, "encoder_q-layer.8": 6696.9502, "encoder_q-layer.9": 6399.1636, "epoch": 0.23, "inbatch_neg_score": 0.2979, "inbatch_pos_score": 0.8804, "learning_rate": 3.855263157894737e-05, "loss": 3.589, "norm_diff": 0.0935, "norm_loss": 0.0, "num_token_doc": 66.8197, "num_token_overlap": 11.731, "num_token_query": 31.5191, "num_token_union": 65.1646, "num_word_context": 202.232, "num_word_doc": 49.8608, "num_word_query": 23.4192, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9572.4271, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.2971, "query_norm": 1.35, "queue_k_norm": 1.4434, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5191, "sent_len_1": 66.8197, "sent_len_max_0": 127.995, "sent_len_max_1": 209.505, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0485, "stdqueue_q": 0.0, "step": 53500 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.6135, "doc_norm": 1.4493, "encoder_q-embeddings": 5303.7124, "encoder_q-layer.0": 3570.2351, "encoder_q-layer.1": 3813.5916, "encoder_q-layer.10": 6766.4438, "encoder_q-layer.11": 15037.9971, "encoder_q-layer.2": 4198.9497, "encoder_q-layer.3": 4205.7563, "encoder_q-layer.4": 4635.4458, "encoder_q-layer.5": 4565.2803, "encoder_q-layer.6": 5285.895, "encoder_q-layer.7": 5935.1191, "encoder_q-layer.8": 7008.7871, "encoder_q-layer.9": 6342.5576, "epoch": 0.23, "inbatch_neg_score": 0.3029, "inbatch_pos_score": 0.8945, "learning_rate": 3.852631578947369e-05, "loss": 3.6135, "norm_diff": 0.0853, "norm_loss": 0.0, "num_token_doc": 66.8704, "num_token_overlap": 11.6808, "num_token_query": 31.3995, "num_token_union": 65.1874, "num_word_context": 202.4978, "num_word_doc": 49.9172, "num_word_query": 23.3149, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9730.164, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3027, "query_norm": 1.364, "queue_k_norm": 1.4472, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3995, "sent_len_1": 66.8704, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.2463, "stdk": 0.0486, "stdq": 0.0434, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53600 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.607, "doc_norm": 1.4439, "encoder_q-embeddings": 5052.3125, "encoder_q-layer.0": 3481.9785, "encoder_q-layer.1": 3663.7234, "encoder_q-layer.10": 6406.6333, "encoder_q-layer.11": 15208.2695, "encoder_q-layer.2": 4010.9985, "encoder_q-layer.3": 4081.4258, "encoder_q-layer.4": 4359.6538, "encoder_q-layer.5": 4540.4526, "encoder_q-layer.6": 5105.2354, "encoder_q-layer.7": 5732.9673, "encoder_q-layer.8": 7253.2261, "encoder_q-layer.9": 6610.2607, "epoch": 0.23, "inbatch_neg_score": 0.3095, "inbatch_pos_score": 0.8916, "learning_rate": 3.85e-05, "loss": 3.607, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.9402, "num_token_overlap": 11.6929, "num_token_query": 31.4078, "num_token_union": 65.2111, "num_word_context": 202.3118, "num_word_doc": 49.9344, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9730.0387, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3101, "query_norm": 1.3756, "queue_k_norm": 1.4473, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4078, "sent_len_1": 66.9402, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.6788, "stdk": 0.0484, "stdq": 0.0433, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53700 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.6104, "doc_norm": 1.4474, "encoder_q-embeddings": 5368.8472, "encoder_q-layer.0": 3549.9172, "encoder_q-layer.1": 3640.3674, "encoder_q-layer.10": 6512.7974, "encoder_q-layer.11": 14391.9062, "encoder_q-layer.2": 4097.4399, "encoder_q-layer.3": 4149.9092, "encoder_q-layer.4": 4319.0493, "encoder_q-layer.5": 4262.9297, "encoder_q-layer.6": 4774.7847, "encoder_q-layer.7": 5614.4292, "encoder_q-layer.8": 6679.2026, "encoder_q-layer.9": 6031.7554, "epoch": 0.23, "inbatch_neg_score": 0.3213, "inbatch_pos_score": 0.9009, "learning_rate": 3.847368421052632e-05, "loss": 3.6104, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.8634, "num_token_overlap": 11.6836, "num_token_query": 31.3909, "num_token_union": 65.1563, "num_word_context": 202.1802, "num_word_doc": 49.89, "num_word_query": 23.3087, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9411.5047, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3213, "query_norm": 1.3678, "queue_k_norm": 1.449, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3909, "sent_len_1": 66.8634, "sent_len_max_0": 127.98, "sent_len_max_1": 210.2875, "stdk": 0.0485, "stdq": 0.0428, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53800 }, { "accuracy": 46.4355, "active_queue_size": 16384.0, "cl_loss": 3.6148, "doc_norm": 1.4452, "encoder_q-embeddings": 6493.9048, "encoder_q-layer.0": 4428.5156, "encoder_q-layer.1": 4646.3179, "encoder_q-layer.10": 6842.8926, "encoder_q-layer.11": 15266.8555, "encoder_q-layer.2": 5349.3765, "encoder_q-layer.3": 5489.0972, "encoder_q-layer.4": 5950.813, "encoder_q-layer.5": 6197.9224, "encoder_q-layer.6": 6607.2534, "encoder_q-layer.7": 7131.6997, "encoder_q-layer.8": 7743.0312, "encoder_q-layer.9": 6670.8481, "epoch": 0.23, "inbatch_neg_score": 0.33, "inbatch_pos_score": 0.8955, "learning_rate": 3.844736842105263e-05, "loss": 3.6148, "norm_diff": 0.0646, "norm_loss": 0.0, "num_token_doc": 66.8362, "num_token_overlap": 11.6435, "num_token_query": 31.2811, "num_token_union": 65.129, "num_word_context": 202.4154, "num_word_doc": 49.8893, "num_word_query": 23.223, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10874.6491, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3296, "query_norm": 1.3805, "queue_k_norm": 1.4498, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2811, "sent_len_1": 66.8362, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.535, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 53900 }, { "accuracy": 47.6074, "active_queue_size": 16384.0, "cl_loss": 3.6312, "doc_norm": 1.4483, "encoder_q-embeddings": 5467.7773, "encoder_q-layer.0": 3581.936, "encoder_q-layer.1": 3848.728, "encoder_q-layer.10": 7372.0859, "encoder_q-layer.11": 14610.4932, "encoder_q-layer.2": 4100.8862, "encoder_q-layer.3": 4274.9268, "encoder_q-layer.4": 4585.9458, "encoder_q-layer.5": 4616.79, "encoder_q-layer.6": 5040.6704, "encoder_q-layer.7": 5828.2188, "encoder_q-layer.8": 6760.6802, "encoder_q-layer.9": 6464.7246, "epoch": 0.23, "inbatch_neg_score": 0.3361, "inbatch_pos_score": 0.9072, "learning_rate": 3.842105263157895e-05, "loss": 3.6312, "norm_diff": 0.0792, "norm_loss": 0.0, "num_token_doc": 66.7602, "num_token_overlap": 11.638, "num_token_query": 31.2595, "num_token_union": 65.0705, "num_word_context": 202.3686, "num_word_doc": 49.857, "num_word_query": 23.2305, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9666.5479, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3352, "query_norm": 1.3691, "queue_k_norm": 1.4518, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2595, "sent_len_1": 66.7602, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.9375, "stdk": 0.0484, "stdq": 0.0429, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54000 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.6074, "doc_norm": 1.4564, "encoder_q-embeddings": 5132.8037, "encoder_q-layer.0": 3398.0315, "encoder_q-layer.1": 3646.5354, "encoder_q-layer.10": 7305.6797, "encoder_q-layer.11": 14973.7324, "encoder_q-layer.2": 3983.4055, "encoder_q-layer.3": 4140.7593, "encoder_q-layer.4": 4430.7959, "encoder_q-layer.5": 4327.6514, "encoder_q-layer.6": 4885.0244, "encoder_q-layer.7": 5529.0127, "encoder_q-layer.8": 6964.6763, "encoder_q-layer.9": 6286.0762, "epoch": 0.23, "inbatch_neg_score": 0.3418, "inbatch_pos_score": 0.9287, "learning_rate": 3.839473684210527e-05, "loss": 3.6074, "norm_diff": 0.0884, "norm_loss": 0.0, "num_token_doc": 66.9706, "num_token_overlap": 11.7136, "num_token_query": 31.4721, "num_token_union": 65.2807, "num_word_context": 202.5248, "num_word_doc": 49.9494, "num_word_query": 23.3651, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9555.5434, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3406, "query_norm": 1.368, "queue_k_norm": 1.4529, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4721, "sent_len_1": 66.9706, "sent_len_max_0": 128.0, "sent_len_max_1": 208.3625, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54100 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.6156, "doc_norm": 1.4548, "encoder_q-embeddings": 10445.832, "encoder_q-layer.0": 6993.7363, "encoder_q-layer.1": 7384.2256, "encoder_q-layer.10": 14825.4258, "encoder_q-layer.11": 31442.1875, "encoder_q-layer.2": 8238.3926, "encoder_q-layer.3": 8458.4746, "encoder_q-layer.4": 8779.4336, "encoder_q-layer.5": 9191.7891, "encoder_q-layer.6": 10015.8096, "encoder_q-layer.7": 11050.1582, "encoder_q-layer.8": 13935.3721, "encoder_q-layer.9": 13275.792, "epoch": 0.24, "inbatch_neg_score": 0.3373, "inbatch_pos_score": 0.9204, "learning_rate": 3.836842105263158e-05, "loss": 3.6156, "norm_diff": 0.1038, "norm_loss": 0.0, "num_token_doc": 66.7159, "num_token_overlap": 11.682, "num_token_query": 31.3527, "num_token_union": 65.0706, "num_word_context": 202.4352, "num_word_doc": 49.7949, "num_word_query": 23.2767, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19453.489, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3376, "query_norm": 1.351, "queue_k_norm": 1.4542, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3527, "sent_len_1": 66.7159, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.3013, "stdk": 0.0486, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54200 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.6056, "doc_norm": 1.4501, "encoder_q-embeddings": 10139.5576, "encoder_q-layer.0": 6973.7329, "encoder_q-layer.1": 7399.9951, "encoder_q-layer.10": 12573.6035, "encoder_q-layer.11": 31389.9824, "encoder_q-layer.2": 8274.1826, "encoder_q-layer.3": 8236.3047, "encoder_q-layer.4": 8951.7373, "encoder_q-layer.5": 8818.374, "encoder_q-layer.6": 10231.832, "encoder_q-layer.7": 11239.6406, "encoder_q-layer.8": 13127.1826, "encoder_q-layer.9": 11987.9199, "epoch": 0.24, "inbatch_neg_score": 0.3338, "inbatch_pos_score": 0.9111, "learning_rate": 3.83421052631579e-05, "loss": 3.6056, "norm_diff": 0.1141, "norm_loss": 0.0, "num_token_doc": 66.8507, "num_token_overlap": 11.7222, "num_token_query": 31.4788, "num_token_union": 65.1759, "num_word_context": 202.1637, "num_word_doc": 49.8861, "num_word_query": 23.3864, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19137.3652, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3342, "query_norm": 1.336, "queue_k_norm": 1.455, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4788, "sent_len_1": 66.8507, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.48, "stdk": 0.0484, "stdq": 0.0426, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54300 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.6076, "doc_norm": 1.451, "encoder_q-embeddings": 13797.5869, "encoder_q-layer.0": 9855.1895, "encoder_q-layer.1": 11256.8535, "encoder_q-layer.10": 14052.041, "encoder_q-layer.11": 31524.2344, "encoder_q-layer.2": 12782.2695, "encoder_q-layer.3": 12810.4033, "encoder_q-layer.4": 11789.5625, "encoder_q-layer.5": 12118.373, "encoder_q-layer.6": 13297.252, "encoder_q-layer.7": 13570.3857, "encoder_q-layer.8": 16089.6816, "encoder_q-layer.9": 13941.8457, "epoch": 0.24, "inbatch_neg_score": 0.3314, "inbatch_pos_score": 0.9028, "learning_rate": 3.831578947368421e-05, "loss": 3.6076, "norm_diff": 0.1065, "norm_loss": 0.0, "num_token_doc": 66.6682, "num_token_overlap": 11.676, "num_token_query": 31.3896, "num_token_union": 65.0836, "num_word_context": 202.6123, "num_word_doc": 49.7909, "num_word_query": 23.3109, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22523.417, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3311, "query_norm": 1.3444, "queue_k_norm": 1.4557, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3896, "sent_len_1": 66.6682, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.5037, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54400 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.6105, "doc_norm": 1.4585, "encoder_q-embeddings": 10636.1729, "encoder_q-layer.0": 7318.6943, "encoder_q-layer.1": 7633.8052, "encoder_q-layer.10": 12983.498, "encoder_q-layer.11": 29810.3652, "encoder_q-layer.2": 8508.6357, "encoder_q-layer.3": 8881.1211, "encoder_q-layer.4": 9403.6895, "encoder_q-layer.5": 9401.333, "encoder_q-layer.6": 10933.7139, "encoder_q-layer.7": 11680.6465, "encoder_q-layer.8": 13606.2383, "encoder_q-layer.9": 12735.2734, "epoch": 0.24, "inbatch_neg_score": 0.3314, "inbatch_pos_score": 0.9355, "learning_rate": 3.828947368421053e-05, "loss": 3.6105, "norm_diff": 0.1052, "norm_loss": 0.0, "num_token_doc": 66.9167, "num_token_overlap": 11.6709, "num_token_query": 31.331, "num_token_union": 65.1875, "num_word_context": 202.127, "num_word_doc": 49.9263, "num_word_query": 23.2688, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19341.0221, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3315, "query_norm": 1.3533, "queue_k_norm": 1.4554, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.331, "sent_len_1": 66.9167, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.16, "stdk": 0.0486, "stdq": 0.0436, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54500 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.6011, "doc_norm": 1.4632, "encoder_q-embeddings": 10216.4814, "encoder_q-layer.0": 6671.6943, "encoder_q-layer.1": 6928.1108, "encoder_q-layer.10": 13354.835, "encoder_q-layer.11": 30268.0352, "encoder_q-layer.2": 7579.9175, "encoder_q-layer.3": 7815.7319, "encoder_q-layer.4": 8328.3213, "encoder_q-layer.5": 8340.2725, "encoder_q-layer.6": 9918.4707, "encoder_q-layer.7": 10842.9834, "encoder_q-layer.8": 13400.1924, "encoder_q-layer.9": 12094.6777, "epoch": 0.24, "inbatch_neg_score": 0.3338, "inbatch_pos_score": 0.9194, "learning_rate": 3.8263157894736846e-05, "loss": 3.6011, "norm_diff": 0.1303, "norm_loss": 0.0, "num_token_doc": 66.8106, "num_token_overlap": 11.6963, "num_token_query": 31.3855, "num_token_union": 65.1542, "num_word_context": 202.2709, "num_word_doc": 49.8724, "num_word_query": 23.3024, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19250.599, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3337, "query_norm": 1.3329, "queue_k_norm": 1.4571, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3855, "sent_len_1": 66.8106, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.7038, "stdk": 0.0488, "stdq": 0.0426, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54600 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.6055, "doc_norm": 1.4603, "encoder_q-embeddings": 10888.8428, "encoder_q-layer.0": 7182.2583, "encoder_q-layer.1": 7583.6758, "encoder_q-layer.10": 13897.3984, "encoder_q-layer.11": 29790.9043, "encoder_q-layer.2": 8375.083, "encoder_q-layer.3": 8673.916, "encoder_q-layer.4": 9060.4766, "encoder_q-layer.5": 9680.7764, "encoder_q-layer.6": 10648.4639, "encoder_q-layer.7": 11941.9854, "encoder_q-layer.8": 14112.7666, "encoder_q-layer.9": 12736.9561, "epoch": 0.24, "inbatch_neg_score": 0.3285, "inbatch_pos_score": 0.9048, "learning_rate": 3.823684210526316e-05, "loss": 3.6055, "norm_diff": 0.1325, "norm_loss": 0.0, "num_token_doc": 66.6778, "num_token_overlap": 11.669, "num_token_query": 31.3411, "num_token_union": 65.0437, "num_word_context": 202.2308, "num_word_doc": 49.764, "num_word_query": 23.2823, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19498.8477, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3291, "query_norm": 1.3278, "queue_k_norm": 1.4582, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3411, "sent_len_1": 66.6778, "sent_len_max_0": 127.99, "sent_len_max_1": 207.055, "stdk": 0.0487, "stdq": 0.0425, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 54700 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.6137, "doc_norm": 1.4577, "encoder_q-embeddings": 16592.1758, "encoder_q-layer.0": 11703.9756, "encoder_q-layer.1": 11107.6094, "encoder_q-layer.10": 16304.1318, "encoder_q-layer.11": 33522.0312, "encoder_q-layer.2": 11437.7012, "encoder_q-layer.3": 10498.3818, "encoder_q-layer.4": 11165.46, "encoder_q-layer.5": 10862.7158, "encoder_q-layer.6": 12285.9404, "encoder_q-layer.7": 13255.7012, "encoder_q-layer.8": 16703.5762, "encoder_q-layer.9": 14850.0918, "epoch": 0.24, "inbatch_neg_score": 0.3313, "inbatch_pos_score": 0.9189, "learning_rate": 3.8210526315789476e-05, "loss": 3.6137, "norm_diff": 0.0965, "norm_loss": 0.0, "num_token_doc": 66.6988, "num_token_overlap": 11.6252, "num_token_query": 31.2785, "num_token_union": 65.0524, "num_word_context": 202.5099, "num_word_doc": 49.7645, "num_word_query": 23.2243, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23515.6008, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3308, "query_norm": 1.3612, "queue_k_norm": 1.4563, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2785, "sent_len_1": 66.6988, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.0788, "stdk": 0.0486, "stdq": 0.0438, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54800 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.6048, "doc_norm": 1.4593, "encoder_q-embeddings": 10485.6572, "encoder_q-layer.0": 6961.0967, "encoder_q-layer.1": 7488.9434, "encoder_q-layer.10": 12905.6123, "encoder_q-layer.11": 29483.6445, "encoder_q-layer.2": 8425.2227, "encoder_q-layer.3": 8825.1191, "encoder_q-layer.4": 9138.1289, "encoder_q-layer.5": 9540.8975, "encoder_q-layer.6": 10328.873, "encoder_q-layer.7": 11549.6279, "encoder_q-layer.8": 13209.998, "encoder_q-layer.9": 12501.8447, "epoch": 0.24, "inbatch_neg_score": 0.33, "inbatch_pos_score": 0.9224, "learning_rate": 3.818421052631579e-05, "loss": 3.6048, "norm_diff": 0.111, "norm_loss": 0.0, "num_token_doc": 66.68, "num_token_overlap": 11.6882, "num_token_query": 31.4244, "num_token_union": 65.082, "num_word_context": 202.2413, "num_word_doc": 49.7479, "num_word_query": 23.3458, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19037.3619, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3303, "query_norm": 1.3483, "queue_k_norm": 1.4585, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4244, "sent_len_1": 66.68, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.245, "stdk": 0.0486, "stdq": 0.0433, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 54900 }, { "accuracy": 46.9727, "active_queue_size": 16384.0, "cl_loss": 3.6129, "doc_norm": 1.4554, "encoder_q-embeddings": 10310.3594, "encoder_q-layer.0": 6888.8354, "encoder_q-layer.1": 7145.8696, "encoder_q-layer.10": 12744.125, "encoder_q-layer.11": 29144.8047, "encoder_q-layer.2": 7698.687, "encoder_q-layer.3": 7855.3086, "encoder_q-layer.4": 8269.3223, "encoder_q-layer.5": 8848.666, "encoder_q-layer.6": 9997.8789, "encoder_q-layer.7": 11302.6494, "encoder_q-layer.8": 13485.7275, "encoder_q-layer.9": 12519.7637, "epoch": 0.24, "inbatch_neg_score": 0.329, "inbatch_pos_score": 0.9038, "learning_rate": 3.815789473684211e-05, "loss": 3.6129, "norm_diff": 0.1165, "norm_loss": 0.0, "num_token_doc": 66.6282, "num_token_overlap": 11.664, "num_token_query": 31.4533, "num_token_union": 65.1049, "num_word_context": 202.3849, "num_word_doc": 49.7216, "num_word_query": 23.3526, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18829.8843, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3289, "query_norm": 1.3389, "queue_k_norm": 1.4574, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4533, "sent_len_1": 66.6282, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.1538, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 55000 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.6107, "doc_norm": 1.4583, "encoder_q-embeddings": 10105.2686, "encoder_q-layer.0": 6914.2837, "encoder_q-layer.1": 7226.8081, "encoder_q-layer.10": 14018.8652, "encoder_q-layer.11": 31571.0684, "encoder_q-layer.2": 7691.1211, "encoder_q-layer.3": 7688.4565, "encoder_q-layer.4": 8207.0557, "encoder_q-layer.5": 8436.043, "encoder_q-layer.6": 9466.9502, "encoder_q-layer.7": 11766.7568, "encoder_q-layer.8": 13891.6289, "encoder_q-layer.9": 13146.4785, "epoch": 0.24, "inbatch_neg_score": 0.329, "inbatch_pos_score": 0.9209, "learning_rate": 3.8131578947368424e-05, "loss": 3.6107, "norm_diff": 0.1172, "norm_loss": 0.0, "num_token_doc": 66.9138, "num_token_overlap": 11.6946, "num_token_query": 31.434, "num_token_union": 65.2249, "num_word_context": 202.4108, "num_word_doc": 49.8924, "num_word_query": 23.3502, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19664.4491, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3293, "query_norm": 1.3411, "queue_k_norm": 1.4618, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.434, "sent_len_1": 66.9138, "sent_len_max_0": 127.9925, "sent_len_max_1": 211.215, "stdk": 0.0486, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 55100 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.5979, "doc_norm": 1.4597, "encoder_q-embeddings": 10413.8096, "encoder_q-layer.0": 7034.8955, "encoder_q-layer.1": 7097.7251, "encoder_q-layer.10": 13692.7119, "encoder_q-layer.11": 30809.4883, "encoder_q-layer.2": 7922.2529, "encoder_q-layer.3": 8067.9775, "encoder_q-layer.4": 8598.002, "encoder_q-layer.5": 9029.8506, "encoder_q-layer.6": 9941.0791, "encoder_q-layer.7": 11389.0801, "encoder_q-layer.8": 13844.5264, "encoder_q-layer.9": 12783.9229, "epoch": 0.24, "inbatch_neg_score": 0.328, "inbatch_pos_score": 0.9058, "learning_rate": 3.8105263157894735e-05, "loss": 3.5979, "norm_diff": 0.125, "norm_loss": 0.0, "num_token_doc": 66.802, "num_token_overlap": 11.7179, "num_token_query": 31.3983, "num_token_union": 65.1235, "num_word_context": 202.2802, "num_word_doc": 49.8523, "num_word_query": 23.3293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19460.9549, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3276, "query_norm": 1.3347, "queue_k_norm": 1.4591, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3983, "sent_len_1": 66.802, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.425, "stdk": 0.0486, "stdq": 0.0429, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 55200 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.6028, "doc_norm": 1.4574, "encoder_q-embeddings": 10566.2598, "encoder_q-layer.0": 7183.6924, "encoder_q-layer.1": 7396.0044, "encoder_q-layer.10": 13167.8926, "encoder_q-layer.11": 29155.0547, "encoder_q-layer.2": 8210.0732, "encoder_q-layer.3": 8508.0811, "encoder_q-layer.4": 8949.4238, "encoder_q-layer.5": 8953.1592, "encoder_q-layer.6": 10131.1943, "encoder_q-layer.7": 11514.8896, "encoder_q-layer.8": 13448.4668, "encoder_q-layer.9": 12312.1416, "epoch": 0.24, "inbatch_neg_score": 0.3307, "inbatch_pos_score": 0.9111, "learning_rate": 3.8078947368421054e-05, "loss": 3.6028, "norm_diff": 0.1219, "norm_loss": 0.0, "num_token_doc": 66.7812, "num_token_overlap": 11.6588, "num_token_query": 31.3559, "num_token_union": 65.1331, "num_word_context": 202.3254, "num_word_doc": 49.8624, "num_word_query": 23.2875, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19210.7103, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3306, "query_norm": 1.3355, "queue_k_norm": 1.4593, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3559, "sent_len_1": 66.7812, "sent_len_max_0": 127.99, "sent_len_max_1": 207.1575, "stdk": 0.0485, "stdq": 0.0429, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 55300 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.6248, "doc_norm": 1.4621, "encoder_q-embeddings": 10739.7373, "encoder_q-layer.0": 6903.8438, "encoder_q-layer.1": 7095.3477, "encoder_q-layer.10": 13507.5723, "encoder_q-layer.11": 30265.418, "encoder_q-layer.2": 7885.5488, "encoder_q-layer.3": 8015.8926, "encoder_q-layer.4": 8501.2363, "encoder_q-layer.5": 8806.793, "encoder_q-layer.6": 9828.9111, "encoder_q-layer.7": 11638.7773, "encoder_q-layer.8": 13579.4746, "encoder_q-layer.9": 13055.9619, "epoch": 0.24, "inbatch_neg_score": 0.3291, "inbatch_pos_score": 0.9233, "learning_rate": 3.8052631578947365e-05, "loss": 3.6248, "norm_diff": 0.1174, "norm_loss": 0.0, "num_token_doc": 66.6141, "num_token_overlap": 11.624, "num_token_query": 31.2447, "num_token_union": 64.9957, "num_word_context": 202.1814, "num_word_doc": 49.7269, "num_word_query": 23.1929, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19586.2485, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3298, "query_norm": 1.3447, "queue_k_norm": 1.4596, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2447, "sent_len_1": 66.6141, "sent_len_max_0": 127.985, "sent_len_max_1": 207.675, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 55400 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.5888, "doc_norm": 1.4631, "encoder_q-embeddings": 10891.5645, "encoder_q-layer.0": 7195.9976, "encoder_q-layer.1": 7584.7969, "encoder_q-layer.10": 12727.9023, "encoder_q-layer.11": 29827.2754, "encoder_q-layer.2": 8613.2646, "encoder_q-layer.3": 8619.583, "encoder_q-layer.4": 8844.1172, "encoder_q-layer.5": 9470.418, "encoder_q-layer.6": 10406.4688, "encoder_q-layer.7": 11786.0107, "encoder_q-layer.8": 14178.4541, "encoder_q-layer.9": 12541.2197, "epoch": 0.24, "inbatch_neg_score": 0.3295, "inbatch_pos_score": 0.9165, "learning_rate": 3.802631578947369e-05, "loss": 3.5888, "norm_diff": 0.126, "norm_loss": 0.0, "num_token_doc": 66.8089, "num_token_overlap": 11.7077, "num_token_query": 31.4456, "num_token_union": 65.1926, "num_word_context": 202.3291, "num_word_doc": 49.8332, "num_word_query": 23.3531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19756.6211, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3296, "query_norm": 1.337, "queue_k_norm": 1.4615, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4456, "sent_len_1": 66.8089, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7688, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 55500 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.6053, "doc_norm": 1.4629, "encoder_q-embeddings": 10848.9697, "encoder_q-layer.0": 6925.6997, "encoder_q-layer.1": 7213.3926, "encoder_q-layer.10": 13534.1289, "encoder_q-layer.11": 31460.457, "encoder_q-layer.2": 7803.8359, "encoder_q-layer.3": 8260.1572, "encoder_q-layer.4": 8650.9619, "encoder_q-layer.5": 8887.0986, "encoder_q-layer.6": 9823.6641, "encoder_q-layer.7": 11289.0303, "encoder_q-layer.8": 13863.8018, "encoder_q-layer.9": 12846.292, "epoch": 0.24, "inbatch_neg_score": 0.3274, "inbatch_pos_score": 0.9146, "learning_rate": 3.8e-05, "loss": 3.6053, "norm_diff": 0.1246, "norm_loss": 0.0, "num_token_doc": 66.7361, "num_token_overlap": 11.6763, "num_token_query": 31.4377, "num_token_union": 65.125, "num_word_context": 202.3229, "num_word_doc": 49.7997, "num_word_query": 23.3521, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19277.1596, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3269, "query_norm": 1.3383, "queue_k_norm": 1.4634, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4377, "sent_len_1": 66.7361, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.8137, "stdk": 0.0487, "stdq": 0.043, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 55600 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.6107, "doc_norm": 1.4629, "encoder_q-embeddings": 10963.0605, "encoder_q-layer.0": 7375.0298, "encoder_q-layer.1": 7732.0981, "encoder_q-layer.10": 13257.7637, "encoder_q-layer.11": 29713.6738, "encoder_q-layer.2": 8561.5488, "encoder_q-layer.3": 8330.7695, "encoder_q-layer.4": 8903.0127, "encoder_q-layer.5": 9149.4893, "encoder_q-layer.6": 10182.7305, "encoder_q-layer.7": 11325.25, "encoder_q-layer.8": 13924.6592, "encoder_q-layer.9": 12754.8809, "epoch": 0.24, "inbatch_neg_score": 0.3344, "inbatch_pos_score": 0.9316, "learning_rate": 3.7973684210526314e-05, "loss": 3.6107, "norm_diff": 0.1092, "norm_loss": 0.0, "num_token_doc": 66.8097, "num_token_overlap": 11.7321, "num_token_query": 31.5439, "num_token_union": 65.1863, "num_word_context": 202.371, "num_word_doc": 49.8254, "num_word_query": 23.4429, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19292.4103, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3342, "query_norm": 1.3536, "queue_k_norm": 1.4618, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5439, "sent_len_1": 66.8097, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.1863, "stdk": 0.0486, "stdq": 0.0434, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 55700 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.6019, "doc_norm": 1.4638, "encoder_q-embeddings": 11362.7998, "encoder_q-layer.0": 7269.8628, "encoder_q-layer.1": 7553.9092, "encoder_q-layer.10": 13795.0166, "encoder_q-layer.11": 30759.2695, "encoder_q-layer.2": 8333.7656, "encoder_q-layer.3": 8809.46, "encoder_q-layer.4": 9360.4922, "encoder_q-layer.5": 9200.9014, "encoder_q-layer.6": 10581.5381, "encoder_q-layer.7": 12094.4424, "encoder_q-layer.8": 14677.9258, "encoder_q-layer.9": 12833.1777, "epoch": 0.24, "inbatch_neg_score": 0.3327, "inbatch_pos_score": 0.9175, "learning_rate": 3.794736842105263e-05, "loss": 3.6019, "norm_diff": 0.1066, "norm_loss": 0.0, "num_token_doc": 66.7968, "num_token_overlap": 11.6796, "num_token_query": 31.438, "num_token_union": 65.1785, "num_word_context": 202.3418, "num_word_doc": 49.8597, "num_word_query": 23.3535, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19936.1621, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3335, "query_norm": 1.3572, "queue_k_norm": 1.4611, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.438, "sent_len_1": 66.7968, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.4588, "stdk": 0.0487, "stdq": 0.0436, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 55800 }, { "accuracy": 46.875, "active_queue_size": 16384.0, "cl_loss": 3.6088, "doc_norm": 1.4614, "encoder_q-embeddings": 10816.0293, "encoder_q-layer.0": 7429.9932, "encoder_q-layer.1": 7630.0171, "encoder_q-layer.10": 12907.5059, "encoder_q-layer.11": 29547.8203, "encoder_q-layer.2": 8693.8135, "encoder_q-layer.3": 8758.7578, "encoder_q-layer.4": 9171.3896, "encoder_q-layer.5": 9084.958, "encoder_q-layer.6": 10711.5283, "encoder_q-layer.7": 11887.9004, "encoder_q-layer.8": 14301.1133, "encoder_q-layer.9": 12495.0029, "epoch": 0.24, "inbatch_neg_score": 0.3344, "inbatch_pos_score": 0.9062, "learning_rate": 3.792105263157895e-05, "loss": 3.6088, "norm_diff": 0.1202, "norm_loss": 0.0, "num_token_doc": 66.7113, "num_token_overlap": 11.6526, "num_token_query": 31.3637, "num_token_union": 65.0882, "num_word_context": 202.3142, "num_word_doc": 49.7621, "num_word_query": 23.2862, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19535.6405, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3347, "query_norm": 1.3412, "queue_k_norm": 1.4621, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3637, "sent_len_1": 66.7113, "sent_len_max_0": 127.985, "sent_len_max_1": 209.89, "stdk": 0.0486, "stdq": 0.0428, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 55900 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.6013, "doc_norm": 1.4575, "encoder_q-embeddings": 10782.5107, "encoder_q-layer.0": 7059.5425, "encoder_q-layer.1": 7332.375, "encoder_q-layer.10": 14198.0107, "encoder_q-layer.11": 31479.5371, "encoder_q-layer.2": 8170.0029, "encoder_q-layer.3": 8325.9922, "encoder_q-layer.4": 9300.4482, "encoder_q-layer.5": 9210.3467, "encoder_q-layer.6": 10990.5693, "encoder_q-layer.7": 12065.2656, "encoder_q-layer.8": 14610.2754, "encoder_q-layer.9": 13238.0928, "epoch": 0.24, "inbatch_neg_score": 0.3324, "inbatch_pos_score": 0.9238, "learning_rate": 3.789473684210527e-05, "loss": 3.6013, "norm_diff": 0.1079, "norm_loss": 0.0, "num_token_doc": 66.9543, "num_token_overlap": 11.6683, "num_token_query": 31.3089, "num_token_union": 65.1964, "num_word_context": 202.4871, "num_word_doc": 49.9657, "num_word_query": 23.2613, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19789.7176, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3315, "query_norm": 1.3496, "queue_k_norm": 1.4622, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3089, "sent_len_1": 66.9543, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.515, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56000 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.6031, "doc_norm": 1.4607, "encoder_q-embeddings": 10693.0303, "encoder_q-layer.0": 7059.5884, "encoder_q-layer.1": 7356.2563, "encoder_q-layer.10": 13255.8838, "encoder_q-layer.11": 30171.7969, "encoder_q-layer.2": 8225.4609, "encoder_q-layer.3": 8358.9385, "encoder_q-layer.4": 8577.4648, "encoder_q-layer.5": 8674.8301, "encoder_q-layer.6": 9886.2812, "encoder_q-layer.7": 11099.8438, "encoder_q-layer.8": 13656.7275, "encoder_q-layer.9": 12383.6309, "epoch": 0.24, "inbatch_neg_score": 0.3367, "inbatch_pos_score": 0.9214, "learning_rate": 3.786842105263158e-05, "loss": 3.6031, "norm_diff": 0.1033, "norm_loss": 0.0, "num_token_doc": 66.7963, "num_token_overlap": 11.664, "num_token_query": 31.3266, "num_token_union": 65.1211, "num_word_context": 202.3395, "num_word_doc": 49.8492, "num_word_query": 23.2796, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19322.5204, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3364, "query_norm": 1.3574, "queue_k_norm": 1.4615, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3266, "sent_len_1": 66.7963, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.8775, "stdk": 0.0485, "stdq": 0.0431, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56100 }, { "accuracy": 47.5586, "active_queue_size": 16384.0, "cl_loss": 3.595, "doc_norm": 1.4597, "encoder_q-embeddings": 21665.2422, "encoder_q-layer.0": 14261.7578, "encoder_q-layer.1": 14770.5752, "encoder_q-layer.10": 29949.5215, "encoder_q-layer.11": 66015.8594, "encoder_q-layer.2": 16511.3281, "encoder_q-layer.3": 17252.6348, "encoder_q-layer.4": 18357.9648, "encoder_q-layer.5": 18579.4766, "encoder_q-layer.6": 19952.9023, "encoder_q-layer.7": 21533.293, "encoder_q-layer.8": 26254.8887, "encoder_q-layer.9": 26105.1094, "epoch": 0.24, "inbatch_neg_score": 0.3395, "inbatch_pos_score": 0.916, "learning_rate": 3.78421052631579e-05, "loss": 3.595, "norm_diff": 0.0908, "norm_loss": 0.0, "num_token_doc": 66.8136, "num_token_overlap": 11.691, "num_token_query": 31.4675, "num_token_union": 65.2035, "num_word_context": 202.312, "num_word_doc": 49.8587, "num_word_query": 23.3751, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39768.9491, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 0.3401, "query_norm": 1.3689, "queue_k_norm": 1.4618, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4675, "sent_len_1": 66.8136, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.0913, "stdk": 0.0485, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56200 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.5842, "doc_norm": 1.4587, "encoder_q-embeddings": 21404.8047, "encoder_q-layer.0": 14409.7344, "encoder_q-layer.1": 15461.6162, "encoder_q-layer.10": 26207.0098, "encoder_q-layer.11": 59666.8398, "encoder_q-layer.2": 17616.7051, "encoder_q-layer.3": 17325.2148, "encoder_q-layer.4": 18845.8828, "encoder_q-layer.5": 18762.6387, "encoder_q-layer.6": 21417.2754, "encoder_q-layer.7": 23804.7031, "encoder_q-layer.8": 29795.2012, "encoder_q-layer.9": 25856.1152, "epoch": 0.24, "inbatch_neg_score": 0.3475, "inbatch_pos_score": 0.9404, "learning_rate": 3.781578947368421e-05, "loss": 3.5842, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.8171, "num_token_overlap": 11.7105, "num_token_query": 31.3591, "num_token_union": 65.1198, "num_word_context": 202.0466, "num_word_doc": 49.8702, "num_word_query": 23.2953, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38946.3549, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 0.3479, "query_norm": 1.3825, "queue_k_norm": 1.4633, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3591, "sent_len_1": 66.8171, "sent_len_max_0": 128.0, "sent_len_max_1": 207.9625, "stdk": 0.0484, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 56300 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.59, "doc_norm": 1.4625, "encoder_q-embeddings": 20776.1191, "encoder_q-layer.0": 13704.0635, "encoder_q-layer.1": 14263.5537, "encoder_q-layer.10": 24513.4844, "encoder_q-layer.11": 57883.5, "encoder_q-layer.2": 15706.3564, "encoder_q-layer.3": 16786.2734, "encoder_q-layer.4": 17858.2266, "encoder_q-layer.5": 17691.2695, "encoder_q-layer.6": 19820.377, "encoder_q-layer.7": 22501.9102, "encoder_q-layer.8": 26828.1387, "encoder_q-layer.9": 24292.3691, "epoch": 0.24, "inbatch_neg_score": 0.3567, "inbatch_pos_score": 0.938, "learning_rate": 3.778947368421053e-05, "loss": 3.59, "norm_diff": 0.073, "norm_loss": 0.0, "num_token_doc": 66.8919, "num_token_overlap": 11.69, "num_token_query": 31.4128, "num_token_union": 65.1702, "num_word_context": 202.4953, "num_word_doc": 49.8682, "num_word_query": 23.3302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37394.548, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 0.3567, "query_norm": 1.3895, "queue_k_norm": 1.4654, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4128, "sent_len_1": 66.8919, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.9062, "stdk": 0.0486, "stdq": 0.0425, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 56400 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.5939, "doc_norm": 1.4633, "encoder_q-embeddings": 12181.5273, "encoder_q-layer.0": 8253.7783, "encoder_q-layer.1": 9017.1738, "encoder_q-layer.10": 13132.418, "encoder_q-layer.11": 29258.7617, "encoder_q-layer.2": 10607.085, "encoder_q-layer.3": 10967.0566, "encoder_q-layer.4": 11214.4941, "encoder_q-layer.5": 11911.8242, "encoder_q-layer.6": 12408.4414, "encoder_q-layer.7": 13520.1084, "encoder_q-layer.8": 14599.9082, "encoder_q-layer.9": 12193.9463, "epoch": 0.25, "inbatch_neg_score": 0.3661, "inbatch_pos_score": 0.9492, "learning_rate": 3.776315789473685e-05, "loss": 3.5939, "norm_diff": 0.0616, "norm_loss": 0.0, "num_token_doc": 66.8494, "num_token_overlap": 11.6665, "num_token_query": 31.3756, "num_token_union": 65.2059, "num_word_context": 202.6001, "num_word_doc": 49.8967, "num_word_query": 23.2961, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20934.7735, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3665, "query_norm": 1.4017, "queue_k_norm": 1.4634, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3756, "sent_len_1": 66.8494, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.03, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56500 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.6115, "doc_norm": 1.4685, "encoder_q-embeddings": 11426.417, "encoder_q-layer.0": 7642.8877, "encoder_q-layer.1": 7908.2407, "encoder_q-layer.10": 12881.1807, "encoder_q-layer.11": 29483.3691, "encoder_q-layer.2": 8565.1025, "encoder_q-layer.3": 8820.085, "encoder_q-layer.4": 9469.335, "encoder_q-layer.5": 9304.3984, "encoder_q-layer.6": 10496.2197, "encoder_q-layer.7": 11661.9775, "encoder_q-layer.8": 13933.5117, "encoder_q-layer.9": 12872.3711, "epoch": 0.25, "inbatch_neg_score": 0.3802, "inbatch_pos_score": 0.9678, "learning_rate": 3.773684210526316e-05, "loss": 3.6115, "norm_diff": 0.0409, "norm_loss": 0.0, "num_token_doc": 66.721, "num_token_overlap": 11.6612, "num_token_query": 31.2924, "num_token_union": 65.0704, "num_word_context": 202.0292, "num_word_doc": 49.7936, "num_word_query": 23.2347, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19464.9109, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3799, "query_norm": 1.4277, "queue_k_norm": 1.4639, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2924, "sent_len_1": 66.721, "sent_len_max_0": 127.9925, "sent_len_max_1": 205.6788, "stdk": 0.0487, "stdq": 0.0438, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56600 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.6009, "doc_norm": 1.4644, "encoder_q-embeddings": 11451.2402, "encoder_q-layer.0": 7597.8545, "encoder_q-layer.1": 8350.4463, "encoder_q-layer.10": 14290.1514, "encoder_q-layer.11": 31601.1465, "encoder_q-layer.2": 9828.1846, "encoder_q-layer.3": 10233.5312, "encoder_q-layer.4": 10690.5586, "encoder_q-layer.5": 10792.8467, "encoder_q-layer.6": 11475.0625, "encoder_q-layer.7": 12795.0537, "encoder_q-layer.8": 15306.8057, "encoder_q-layer.9": 13288.7148, "epoch": 0.25, "inbatch_neg_score": 0.384, "inbatch_pos_score": 0.9707, "learning_rate": 3.771052631578948e-05, "loss": 3.6009, "norm_diff": 0.0564, "norm_loss": 0.0, "num_token_doc": 66.7226, "num_token_overlap": 11.6741, "num_token_query": 31.375, "num_token_union": 65.0929, "num_word_context": 202.2894, "num_word_doc": 49.7936, "num_word_query": 23.293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20411.0363, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3843, "query_norm": 1.4079, "queue_k_norm": 1.4666, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.375, "sent_len_1": 66.7226, "sent_len_max_0": 127.985, "sent_len_max_1": 205.6012, "stdk": 0.0485, "stdq": 0.0435, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56700 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.5911, "doc_norm": 1.4689, "encoder_q-embeddings": 16500.4863, "encoder_q-layer.0": 11649.8418, "encoder_q-layer.1": 12475.6533, "encoder_q-layer.10": 14460.8271, "encoder_q-layer.11": 29651.0234, "encoder_q-layer.2": 14412.1455, "encoder_q-layer.3": 14407.5322, "encoder_q-layer.4": 16185.9053, "encoder_q-layer.5": 15483.1777, "encoder_q-layer.6": 15623.0547, "encoder_q-layer.7": 16122.5137, "encoder_q-layer.8": 16239.9395, "encoder_q-layer.9": 12674.668, "epoch": 0.25, "inbatch_neg_score": 0.3833, "inbatch_pos_score": 0.9634, "learning_rate": 3.768421052631579e-05, "loss": 3.5911, "norm_diff": 0.0962, "norm_loss": 0.0, "num_token_doc": 66.7254, "num_token_overlap": 11.7151, "num_token_query": 31.438, "num_token_union": 65.114, "num_word_context": 201.9775, "num_word_doc": 49.7729, "num_word_query": 23.3288, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24428.0113, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3831, "query_norm": 1.3727, "queue_k_norm": 1.4692, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.438, "sent_len_1": 66.7254, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.4837, "stdk": 0.0486, "stdq": 0.0429, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 56800 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.5987, "doc_norm": 1.462, "encoder_q-embeddings": 9469.4619, "encoder_q-layer.0": 6498.1284, "encoder_q-layer.1": 6764.6865, "encoder_q-layer.10": 12917.124, "encoder_q-layer.11": 30592.7871, "encoder_q-layer.2": 7613.6284, "encoder_q-layer.3": 7878.1108, "encoder_q-layer.4": 8561.0244, "encoder_q-layer.5": 8621.9619, "encoder_q-layer.6": 9676.6777, "encoder_q-layer.7": 11238.0625, "encoder_q-layer.8": 13303.0205, "encoder_q-layer.9": 12075.4326, "epoch": 0.25, "inbatch_neg_score": 0.3815, "inbatch_pos_score": 0.9502, "learning_rate": 3.765789473684211e-05, "loss": 3.5987, "norm_diff": 0.1108, "norm_loss": 0.0, "num_token_doc": 66.7827, "num_token_overlap": 11.7221, "num_token_query": 31.4665, "num_token_union": 65.134, "num_word_context": 202.163, "num_word_doc": 49.8527, "num_word_query": 23.3543, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18357.5615, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3806, "query_norm": 1.3511, "queue_k_norm": 1.469, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4665, "sent_len_1": 66.7827, "sent_len_max_0": 127.9675, "sent_len_max_1": 206.9837, "stdk": 0.0483, "stdq": 0.0425, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 56900 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.5829, "doc_norm": 1.4692, "encoder_q-embeddings": 10809.7412, "encoder_q-layer.0": 7324.979, "encoder_q-layer.1": 8118.2627, "encoder_q-layer.10": 13530.252, "encoder_q-layer.11": 30467.5723, "encoder_q-layer.2": 9322.1416, "encoder_q-layer.3": 9137.29, "encoder_q-layer.4": 9233.5469, "encoder_q-layer.5": 9172.8389, "encoder_q-layer.6": 10512.3789, "encoder_q-layer.7": 11452.1816, "encoder_q-layer.8": 14525.5938, "encoder_q-layer.9": 12867.1152, "epoch": 0.25, "inbatch_neg_score": 0.3783, "inbatch_pos_score": 0.9678, "learning_rate": 3.7631578947368425e-05, "loss": 3.5829, "norm_diff": 0.1166, "norm_loss": 0.0, "num_token_doc": 66.8579, "num_token_overlap": 11.681, "num_token_query": 31.383, "num_token_union": 65.1609, "num_word_context": 202.0856, "num_word_doc": 49.8803, "num_word_query": 23.2943, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19672.1292, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3779, "query_norm": 1.3526, "queue_k_norm": 1.4722, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.383, "sent_len_1": 66.8579, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.4712, "stdk": 0.0485, "stdq": 0.0428, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 57000 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.6069, "doc_norm": 1.4784, "encoder_q-embeddings": 10743.1982, "encoder_q-layer.0": 6860.4941, "encoder_q-layer.1": 7072.6782, "encoder_q-layer.10": 12496.8682, "encoder_q-layer.11": 31820.75, "encoder_q-layer.2": 7949.3452, "encoder_q-layer.3": 8046.2812, "encoder_q-layer.4": 8552.0449, "encoder_q-layer.5": 8719.8418, "encoder_q-layer.6": 9557.0693, "encoder_q-layer.7": 10779.1846, "encoder_q-layer.8": 12969.4629, "encoder_q-layer.9": 12174.9287, "epoch": 0.25, "inbatch_neg_score": 0.3769, "inbatch_pos_score": 0.9668, "learning_rate": 3.760526315789474e-05, "loss": 3.6069, "norm_diff": 0.1139, "norm_loss": 0.0, "num_token_doc": 66.7344, "num_token_overlap": 11.6931, "num_token_query": 31.4558, "num_token_union": 65.162, "num_word_context": 202.5125, "num_word_doc": 49.8239, "num_word_query": 23.3903, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19343.5428, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3765, "query_norm": 1.3644, "queue_k_norm": 1.471, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4558, "sent_len_1": 66.7344, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.7225, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57100 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.5947, "doc_norm": 1.4715, "encoder_q-embeddings": 13060.9316, "encoder_q-layer.0": 9393.2666, "encoder_q-layer.1": 10184.624, "encoder_q-layer.10": 14277.0352, "encoder_q-layer.11": 30302.8848, "encoder_q-layer.2": 11341.9844, "encoder_q-layer.3": 11978.4941, "encoder_q-layer.4": 12481.5605, "encoder_q-layer.5": 12834.373, "encoder_q-layer.6": 13576.293, "encoder_q-layer.7": 13060.9561, "encoder_q-layer.8": 13519.2803, "encoder_q-layer.9": 12686.7129, "epoch": 0.25, "inbatch_neg_score": 0.3738, "inbatch_pos_score": 0.9526, "learning_rate": 3.7578947368421055e-05, "loss": 3.5947, "norm_diff": 0.1282, "norm_loss": 0.0, "num_token_doc": 66.7265, "num_token_overlap": 11.7266, "num_token_query": 31.4339, "num_token_union": 65.107, "num_word_context": 202.1785, "num_word_doc": 49.8141, "num_word_query": 23.3534, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21481.1471, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3745, "query_norm": 1.3434, "queue_k_norm": 1.4715, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4339, "sent_len_1": 66.7265, "sent_len_max_0": 127.985, "sent_len_max_1": 208.5875, "stdk": 0.0486, "stdq": 0.0425, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57200 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.6163, "doc_norm": 1.477, "encoder_q-embeddings": 10411.6172, "encoder_q-layer.0": 6991.3057, "encoder_q-layer.1": 7191.4814, "encoder_q-layer.10": 13561.5732, "encoder_q-layer.11": 30652.3379, "encoder_q-layer.2": 8002.7554, "encoder_q-layer.3": 8140.9927, "encoder_q-layer.4": 8761.2617, "encoder_q-layer.5": 9437.8291, "encoder_q-layer.6": 10347.0459, "encoder_q-layer.7": 11258.3027, "encoder_q-layer.8": 14024.2549, "encoder_q-layer.9": 12610.3818, "epoch": 0.25, "inbatch_neg_score": 0.3722, "inbatch_pos_score": 0.9492, "learning_rate": 3.7552631578947374e-05, "loss": 3.6163, "norm_diff": 0.1233, "norm_loss": 0.0, "num_token_doc": 66.7928, "num_token_overlap": 11.6321, "num_token_query": 31.2847, "num_token_union": 65.1182, "num_word_context": 202.6624, "num_word_doc": 49.8516, "num_word_query": 23.2278, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19528.4765, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3728, "query_norm": 1.3537, "queue_k_norm": 1.4729, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2847, "sent_len_1": 66.7928, "sent_len_max_0": 127.98, "sent_len_max_1": 206.4975, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 57300 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.6088, "doc_norm": 1.4738, "encoder_q-embeddings": 11441.543, "encoder_q-layer.0": 7572.2832, "encoder_q-layer.1": 7982.5161, "encoder_q-layer.10": 14130.0283, "encoder_q-layer.11": 32258.375, "encoder_q-layer.2": 9184.998, "encoder_q-layer.3": 9571.335, "encoder_q-layer.4": 10349.4375, "encoder_q-layer.5": 10597.9121, "encoder_q-layer.6": 11427.5635, "encoder_q-layer.7": 12814.0088, "encoder_q-layer.8": 15233.4072, "encoder_q-layer.9": 13693.4512, "epoch": 0.25, "inbatch_neg_score": 0.3713, "inbatch_pos_score": 0.9634, "learning_rate": 3.7526315789473685e-05, "loss": 3.6088, "norm_diff": 0.1076, "norm_loss": 0.0, "num_token_doc": 66.8528, "num_token_overlap": 11.6776, "num_token_query": 31.3282, "num_token_union": 65.1579, "num_word_context": 202.396, "num_word_doc": 49.9035, "num_word_query": 23.2682, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20824.569, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3713, "query_norm": 1.3662, "queue_k_norm": 1.4719, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3282, "sent_len_1": 66.8528, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.145, "stdk": 0.0486, "stdq": 0.0434, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57400 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5887, "doc_norm": 1.4762, "encoder_q-embeddings": 10793.9502, "encoder_q-layer.0": 7036.0615, "encoder_q-layer.1": 7239.3252, "encoder_q-layer.10": 13992.9531, "encoder_q-layer.11": 30312.1953, "encoder_q-layer.2": 8222.3096, "encoder_q-layer.3": 8388.4355, "encoder_q-layer.4": 8777.8291, "encoder_q-layer.5": 9055.6807, "encoder_q-layer.6": 10134.4561, "encoder_q-layer.7": 11486.0342, "encoder_q-layer.8": 13047.083, "encoder_q-layer.9": 12911.0479, "epoch": 0.25, "inbatch_neg_score": 0.3686, "inbatch_pos_score": 0.9648, "learning_rate": 3.7500000000000003e-05, "loss": 3.5887, "norm_diff": 0.1183, "norm_loss": 0.0, "num_token_doc": 66.7953, "num_token_overlap": 11.7194, "num_token_query": 31.48, "num_token_union": 65.1831, "num_word_context": 202.2454, "num_word_doc": 49.8555, "num_word_query": 23.3723, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19179.975, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3689, "query_norm": 1.3579, "queue_k_norm": 1.4746, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.48, "sent_len_1": 66.7953, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.45, "stdk": 0.0487, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 57500 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.6102, "doc_norm": 1.4775, "encoder_q-embeddings": 12607.7275, "encoder_q-layer.0": 9275.8379, "encoder_q-layer.1": 9988.1045, "encoder_q-layer.10": 12926.4102, "encoder_q-layer.11": 29551.1309, "encoder_q-layer.2": 11439.0791, "encoder_q-layer.3": 11583.1572, "encoder_q-layer.4": 13309.4668, "encoder_q-layer.5": 12666.3262, "encoder_q-layer.6": 13629.9414, "encoder_q-layer.7": 13727.2041, "encoder_q-layer.8": 13788.8555, "encoder_q-layer.9": 12574.8867, "epoch": 0.25, "inbatch_neg_score": 0.3744, "inbatch_pos_score": 0.957, "learning_rate": 3.7473684210526315e-05, "loss": 3.6102, "norm_diff": 0.1173, "norm_loss": 0.0, "num_token_doc": 66.7587, "num_token_overlap": 11.6401, "num_token_query": 31.2521, "num_token_union": 65.057, "num_word_context": 202.2363, "num_word_doc": 49.8024, "num_word_query": 23.1949, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21451.6774, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.374, "query_norm": 1.3602, "queue_k_norm": 1.4733, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2521, "sent_len_1": 66.7587, "sent_len_max_0": 127.975, "sent_len_max_1": 209.5687, "stdk": 0.0487, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57600 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.5944, "doc_norm": 1.4793, "encoder_q-embeddings": 10280.4883, "encoder_q-layer.0": 7197.2324, "encoder_q-layer.1": 7545.9785, "encoder_q-layer.10": 12847.0469, "encoder_q-layer.11": 29316.5977, "encoder_q-layer.2": 8484.7451, "encoder_q-layer.3": 8757.9053, "encoder_q-layer.4": 9114.1943, "encoder_q-layer.5": 9426.2598, "encoder_q-layer.6": 10117.957, "encoder_q-layer.7": 11111.8203, "encoder_q-layer.8": 13373.6875, "encoder_q-layer.9": 12406.8682, "epoch": 0.25, "inbatch_neg_score": 0.3796, "inbatch_pos_score": 0.9653, "learning_rate": 3.7447368421052633e-05, "loss": 3.5944, "norm_diff": 0.1129, "norm_loss": 0.0, "num_token_doc": 66.8711, "num_token_overlap": 11.6774, "num_token_query": 31.3925, "num_token_union": 65.1815, "num_word_context": 202.5649, "num_word_doc": 49.8832, "num_word_query": 23.3238, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19009.9139, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3794, "query_norm": 1.3663, "queue_k_norm": 1.473, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3925, "sent_len_1": 66.8711, "sent_len_max_0": 127.985, "sent_len_max_1": 210.745, "stdk": 0.0488, "stdq": 0.043, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57700 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.5968, "doc_norm": 1.4696, "encoder_q-embeddings": 11522.8174, "encoder_q-layer.0": 7987.0957, "encoder_q-layer.1": 8616.3164, "encoder_q-layer.10": 13162.3691, "encoder_q-layer.11": 28856.0625, "encoder_q-layer.2": 9496.2598, "encoder_q-layer.3": 10355.6367, "encoder_q-layer.4": 10406.8379, "encoder_q-layer.5": 10143.835, "encoder_q-layer.6": 10442.1738, "encoder_q-layer.7": 12122.8379, "encoder_q-layer.8": 13554.7744, "encoder_q-layer.9": 12534.5527, "epoch": 0.25, "inbatch_neg_score": 0.3791, "inbatch_pos_score": 0.9688, "learning_rate": 3.742105263157895e-05, "loss": 3.5968, "norm_diff": 0.0937, "norm_loss": 0.0, "num_token_doc": 66.5995, "num_token_overlap": 11.7229, "num_token_query": 31.4861, "num_token_union": 65.0538, "num_word_context": 202.0401, "num_word_doc": 49.7287, "num_word_query": 23.3872, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19857.769, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3809, "query_norm": 1.3759, "queue_k_norm": 1.4749, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4861, "sent_len_1": 66.5995, "sent_len_max_0": 127.9988, "sent_len_max_1": 205.3338, "stdk": 0.0484, "stdq": 0.0434, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 57800 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.5919, "doc_norm": 1.4744, "encoder_q-embeddings": 11458.8379, "encoder_q-layer.0": 7787.4868, "encoder_q-layer.1": 8306.6152, "encoder_q-layer.10": 14442.1533, "encoder_q-layer.11": 29985.2695, "encoder_q-layer.2": 9225.4668, "encoder_q-layer.3": 9571.1377, "encoder_q-layer.4": 10527.3369, "encoder_q-layer.5": 10567.6904, "encoder_q-layer.6": 11049.8818, "encoder_q-layer.7": 12040.3555, "encoder_q-layer.8": 14288.5547, "encoder_q-layer.9": 13075.7988, "epoch": 0.25, "inbatch_neg_score": 0.3821, "inbatch_pos_score": 0.9648, "learning_rate": 3.739473684210526e-05, "loss": 3.5919, "norm_diff": 0.1026, "norm_loss": 0.0, "num_token_doc": 66.677, "num_token_overlap": 11.7066, "num_token_query": 31.434, "num_token_union": 65.0668, "num_word_context": 202.1828, "num_word_doc": 49.7852, "num_word_query": 23.3521, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20062.556, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3813, "query_norm": 1.3719, "queue_k_norm": 1.4757, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.434, "sent_len_1": 66.677, "sent_len_max_0": 127.9875, "sent_len_max_1": 206.3475, "stdk": 0.0485, "stdq": 0.0432, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 57900 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.5807, "doc_norm": 1.4808, "encoder_q-embeddings": 10497.6523, "encoder_q-layer.0": 7043.9404, "encoder_q-layer.1": 7338.3394, "encoder_q-layer.10": 12559.8018, "encoder_q-layer.11": 29510.1426, "encoder_q-layer.2": 8090.5938, "encoder_q-layer.3": 8226.1416, "encoder_q-layer.4": 8895.793, "encoder_q-layer.5": 8839.5439, "encoder_q-layer.6": 10207.2021, "encoder_q-layer.7": 12047.3066, "encoder_q-layer.8": 13481.7402, "encoder_q-layer.9": 12396.4023, "epoch": 0.25, "inbatch_neg_score": 0.3793, "inbatch_pos_score": 0.9736, "learning_rate": 3.736842105263158e-05, "loss": 3.5807, "norm_diff": 0.1075, "norm_loss": 0.0, "num_token_doc": 66.8267, "num_token_overlap": 11.6888, "num_token_query": 31.373, "num_token_union": 65.1608, "num_word_context": 202.3499, "num_word_doc": 49.8665, "num_word_query": 23.3073, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19247.7219, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.3794, "query_norm": 1.3733, "queue_k_norm": 1.4771, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.373, "sent_len_1": 66.8267, "sent_len_max_0": 127.97, "sent_len_max_1": 210.0687, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 58000 }, { "accuracy": 46.3867, "active_queue_size": 16384.0, "cl_loss": 3.5918, "doc_norm": 1.4771, "encoder_q-embeddings": 5337.5767, "encoder_q-layer.0": 3569.8721, "encoder_q-layer.1": 3751.3086, "encoder_q-layer.10": 6461.4658, "encoder_q-layer.11": 14679.96, "encoder_q-layer.2": 4197.4893, "encoder_q-layer.3": 4255.7163, "encoder_q-layer.4": 4584.7393, "encoder_q-layer.5": 4621.0459, "encoder_q-layer.6": 5278.8945, "encoder_q-layer.7": 5816.397, "encoder_q-layer.8": 7085.3096, "encoder_q-layer.9": 6180.293, "epoch": 0.25, "inbatch_neg_score": 0.3781, "inbatch_pos_score": 0.9438, "learning_rate": 3.734210526315789e-05, "loss": 3.5918, "norm_diff": 0.122, "norm_loss": 0.0, "num_token_doc": 66.9306, "num_token_overlap": 11.7161, "num_token_query": 31.5012, "num_token_union": 65.2454, "num_word_context": 202.5816, "num_word_doc": 49.924, "num_word_query": 23.4, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9700.0315, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3787, "query_norm": 1.3551, "queue_k_norm": 1.4765, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5012, "sent_len_1": 66.9306, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.9613, "stdk": 0.0486, "stdq": 0.0424, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 58100 }, { "accuracy": 47.2168, "active_queue_size": 16384.0, "cl_loss": 3.5823, "doc_norm": 1.4788, "encoder_q-embeddings": 5614.3286, "encoder_q-layer.0": 3801.1562, "encoder_q-layer.1": 4098.877, "encoder_q-layer.10": 6518.1245, "encoder_q-layer.11": 14833.833, "encoder_q-layer.2": 4544.7222, "encoder_q-layer.3": 4714.644, "encoder_q-layer.4": 4877.7559, "encoder_q-layer.5": 4943.7651, "encoder_q-layer.6": 5453.1445, "encoder_q-layer.7": 5788.2744, "encoder_q-layer.8": 7077.2163, "encoder_q-layer.9": 6064.1855, "epoch": 0.25, "inbatch_neg_score": 0.3825, "inbatch_pos_score": 0.9668, "learning_rate": 3.731578947368422e-05, "loss": 3.5823, "norm_diff": 0.0975, "norm_loss": 0.0, "num_token_doc": 66.7475, "num_token_overlap": 11.7007, "num_token_query": 31.4245, "num_token_union": 65.09, "num_word_context": 202.0552, "num_word_doc": 49.7793, "num_word_query": 23.3443, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9871.4366, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3818, "query_norm": 1.3813, "queue_k_norm": 1.4757, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4245, "sent_len_1": 66.7475, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.705, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58200 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.5852, "doc_norm": 1.4763, "encoder_q-embeddings": 6819.8428, "encoder_q-layer.0": 4708.1958, "encoder_q-layer.1": 5445.8252, "encoder_q-layer.10": 6448.6787, "encoder_q-layer.11": 14829.9209, "encoder_q-layer.2": 5970.4224, "encoder_q-layer.3": 6384.7114, "encoder_q-layer.4": 7065.7954, "encoder_q-layer.5": 7098.9556, "encoder_q-layer.6": 7027.3345, "encoder_q-layer.7": 7558.7842, "encoder_q-layer.8": 8069.5205, "encoder_q-layer.9": 6365.1123, "epoch": 0.25, "inbatch_neg_score": 0.3825, "inbatch_pos_score": 0.9897, "learning_rate": 3.728947368421053e-05, "loss": 3.5852, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.6375, "num_token_overlap": 11.6381, "num_token_query": 31.3571, "num_token_union": 65.0465, "num_word_context": 202.3323, "num_word_doc": 49.7479, "num_word_query": 23.2923, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11398.8678, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3826, "query_norm": 1.4042, "queue_k_norm": 1.4766, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3571, "sent_len_1": 66.6375, "sent_len_max_0": 127.97, "sent_len_max_1": 210.0012, "stdk": 0.0486, "stdq": 0.044, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58300 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5935, "doc_norm": 1.4815, "encoder_q-embeddings": 5234.5806, "encoder_q-layer.0": 3517.1326, "encoder_q-layer.1": 3617.1108, "encoder_q-layer.10": 6582.9565, "encoder_q-layer.11": 15033.7666, "encoder_q-layer.2": 3997.0334, "encoder_q-layer.3": 4113.3364, "encoder_q-layer.4": 4407.2954, "encoder_q-layer.5": 4560.5137, "encoder_q-layer.6": 5201.0977, "encoder_q-layer.7": 5836.6587, "encoder_q-layer.8": 7019.082, "encoder_q-layer.9": 6235.8369, "epoch": 0.25, "inbatch_neg_score": 0.3855, "inbatch_pos_score": 0.9678, "learning_rate": 3.726315789473684e-05, "loss": 3.5935, "norm_diff": 0.0967, "norm_loss": 0.0, "num_token_doc": 66.7369, "num_token_overlap": 11.6577, "num_token_query": 31.3413, "num_token_union": 65.1104, "num_word_context": 202.3871, "num_word_doc": 49.8203, "num_word_query": 23.2799, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9669.7709, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.385, "query_norm": 1.3848, "queue_k_norm": 1.4775, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3413, "sent_len_1": 66.7369, "sent_len_max_0": 127.9938, "sent_len_max_1": 206.9475, "stdk": 0.0487, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58400 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.579, "doc_norm": 1.473, "encoder_q-embeddings": 5063.0249, "encoder_q-layer.0": 3346.0317, "encoder_q-layer.1": 3557.3804, "encoder_q-layer.10": 6322.791, "encoder_q-layer.11": 14404.1143, "encoder_q-layer.2": 3920.7935, "encoder_q-layer.3": 3962.2559, "encoder_q-layer.4": 4178.75, "encoder_q-layer.5": 4256.5986, "encoder_q-layer.6": 4787.084, "encoder_q-layer.7": 5248.7153, "encoder_q-layer.8": 6563.4717, "encoder_q-layer.9": 5911.7852, "epoch": 0.25, "inbatch_neg_score": 0.3922, "inbatch_pos_score": 0.9834, "learning_rate": 3.723684210526316e-05, "loss": 3.579, "norm_diff": 0.0652, "norm_loss": 0.0, "num_token_doc": 66.8346, "num_token_overlap": 11.7137, "num_token_query": 31.4496, "num_token_union": 65.1741, "num_word_context": 202.3209, "num_word_doc": 49.9007, "num_word_query": 23.3644, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9186.28, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.3923, "query_norm": 1.4078, "queue_k_norm": 1.4776, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4496, "sent_len_1": 66.8346, "sent_len_max_0": 127.9537, "sent_len_max_1": 207.39, "stdk": 0.0484, "stdq": 0.0433, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58500 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.5738, "doc_norm": 1.4766, "encoder_q-embeddings": 2653.8059, "encoder_q-layer.0": 1763.7408, "encoder_q-layer.1": 1853.2853, "encoder_q-layer.10": 3307.4668, "encoder_q-layer.11": 7382.5483, "encoder_q-layer.2": 2085.8311, "encoder_q-layer.3": 2101.1755, "encoder_q-layer.4": 2237.3171, "encoder_q-layer.5": 2284.2607, "encoder_q-layer.6": 2581.9375, "encoder_q-layer.7": 2911.1707, "encoder_q-layer.8": 3582.2202, "encoder_q-layer.9": 3201.7163, "epoch": 0.25, "inbatch_neg_score": 0.3986, "inbatch_pos_score": 0.9692, "learning_rate": 3.721052631578947e-05, "loss": 3.5738, "norm_diff": 0.0614, "norm_loss": 0.0, "num_token_doc": 66.9276, "num_token_overlap": 11.6587, "num_token_query": 31.3582, "num_token_union": 65.2227, "num_word_context": 202.4212, "num_word_doc": 49.9436, "num_word_query": 23.2929, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4795.1481, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.3977, "query_norm": 1.4153, "queue_k_norm": 1.478, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3582, "sent_len_1": 66.9276, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.8137, "stdk": 0.0485, "stdq": 0.0429, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58600 }, { "accuracy": 47.5098, "active_queue_size": 16384.0, "cl_loss": 3.5891, "doc_norm": 1.4818, "encoder_q-embeddings": 2725.2271, "encoder_q-layer.0": 1858.4545, "encoder_q-layer.1": 1931.9386, "encoder_q-layer.10": 3665.8369, "encoder_q-layer.11": 8388.5254, "encoder_q-layer.2": 2095.4648, "encoder_q-layer.3": 2217.0977, "encoder_q-layer.4": 2360.0149, "encoder_q-layer.5": 2394.0823, "encoder_q-layer.6": 2725.3696, "encoder_q-layer.7": 3088.6543, "encoder_q-layer.8": 3925.3752, "encoder_q-layer.9": 3308.1702, "epoch": 0.25, "inbatch_neg_score": 0.4077, "inbatch_pos_score": 0.9873, "learning_rate": 3.71842105263158e-05, "loss": 3.5891, "norm_diff": 0.0551, "norm_loss": 0.0, "num_token_doc": 66.7706, "num_token_overlap": 11.693, "num_token_query": 31.3826, "num_token_union": 65.1226, "num_word_context": 202.4777, "num_word_doc": 49.8143, "num_word_query": 23.3044, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5305.5709, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4072, "query_norm": 1.4268, "queue_k_norm": 1.4805, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3826, "sent_len_1": 66.7706, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.765, "stdk": 0.0487, "stdq": 0.0428, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 58700 }, { "accuracy": 46.0938, "active_queue_size": 16384.0, "cl_loss": 3.5829, "doc_norm": 1.4814, "encoder_q-embeddings": 2579.6255, "encoder_q-layer.0": 1735.6843, "encoder_q-layer.1": 1818.1855, "encoder_q-layer.10": 3160.8784, "encoder_q-layer.11": 7410.3828, "encoder_q-layer.2": 2037.7435, "encoder_q-layer.3": 2100.3879, "encoder_q-layer.4": 2259.9663, "encoder_q-layer.5": 2264.9785, "encoder_q-layer.6": 2498.4744, "encoder_q-layer.7": 2745.7693, "encoder_q-layer.8": 3339.5305, "encoder_q-layer.9": 3073.8574, "epoch": 0.26, "inbatch_neg_score": 0.422, "inbatch_pos_score": 0.9893, "learning_rate": 3.715789473684211e-05, "loss": 3.5829, "norm_diff": 0.0479, "norm_loss": 0.0, "num_token_doc": 66.6401, "num_token_overlap": 11.6595, "num_token_query": 31.3489, "num_token_union": 65.0398, "num_word_context": 202.0007, "num_word_doc": 49.7377, "num_word_query": 23.287, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4778.6684, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4219, "query_norm": 1.4335, "queue_k_norm": 1.4808, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3489, "sent_len_1": 66.6401, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.3088, "stdk": 0.0486, "stdq": 0.0428, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 58800 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.5766, "doc_norm": 1.4847, "encoder_q-embeddings": 2849.6396, "encoder_q-layer.0": 1941.9227, "encoder_q-layer.1": 2099.7612, "encoder_q-layer.10": 3549.2949, "encoder_q-layer.11": 7844.3252, "encoder_q-layer.2": 2366.7136, "encoder_q-layer.3": 2506.3604, "encoder_q-layer.4": 2684.2214, "encoder_q-layer.5": 2717.1843, "encoder_q-layer.6": 2931.5088, "encoder_q-layer.7": 3173.9729, "encoder_q-layer.8": 3631.8049, "encoder_q-layer.9": 3395.0771, "epoch": 0.26, "inbatch_neg_score": 0.4356, "inbatch_pos_score": 1.0283, "learning_rate": 3.713157894736842e-05, "loss": 3.5766, "norm_diff": 0.0278, "norm_loss": 0.0, "num_token_doc": 66.7337, "num_token_overlap": 11.6727, "num_token_query": 31.3698, "num_token_union": 65.072, "num_word_context": 202.2593, "num_word_doc": 49.7596, "num_word_query": 23.3032, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5222.4991, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4355, "query_norm": 1.4569, "queue_k_norm": 1.483, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3698, "sent_len_1": 66.7337, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.7937, "stdk": 0.0487, "stdq": 0.0437, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 58900 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5792, "doc_norm": 1.4836, "encoder_q-embeddings": 2971.3435, "encoder_q-layer.0": 1947.9521, "encoder_q-layer.1": 2121.6821, "encoder_q-layer.10": 3104.4731, "encoder_q-layer.11": 7493.3213, "encoder_q-layer.2": 2411.8276, "encoder_q-layer.3": 2624.6775, "encoder_q-layer.4": 2771.3325, "encoder_q-layer.5": 2892.561, "encoder_q-layer.6": 2945.1182, "encoder_q-layer.7": 3251.0618, "encoder_q-layer.8": 3598.8601, "encoder_q-layer.9": 3065.0916, "epoch": 0.26, "inbatch_neg_score": 0.448, "inbatch_pos_score": 1.0449, "learning_rate": 3.710526315789474e-05, "loss": 3.5792, "norm_diff": 0.0455, "norm_loss": 0.0, "num_token_doc": 66.7915, "num_token_overlap": 11.635, "num_token_query": 31.1769, "num_token_union": 65.0498, "num_word_context": 202.3372, "num_word_doc": 49.8581, "num_word_query": 23.1302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5071.8875, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.448, "query_norm": 1.4381, "queue_k_norm": 1.4854, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.1769, "sent_len_1": 66.7915, "sent_len_max_0": 127.965, "sent_len_max_1": 207.875, "stdk": 0.0486, "stdq": 0.0433, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59000 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.5826, "doc_norm": 1.4884, "encoder_q-embeddings": 3865.8396, "encoder_q-layer.0": 2821.8308, "encoder_q-layer.1": 2932.1904, "encoder_q-layer.10": 3518.7317, "encoder_q-layer.11": 8150.0298, "encoder_q-layer.2": 3229.3235, "encoder_q-layer.3": 3412.7368, "encoder_q-layer.4": 3592.8269, "encoder_q-layer.5": 3730.2178, "encoder_q-layer.6": 3656.0459, "encoder_q-layer.7": 3916.5586, "encoder_q-layer.8": 4286.1094, "encoder_q-layer.9": 3424.772, "epoch": 0.26, "inbatch_neg_score": 0.4495, "inbatch_pos_score": 1.0352, "learning_rate": 3.707894736842105e-05, "loss": 3.5826, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.8453, "num_token_overlap": 11.6904, "num_token_query": 31.3622, "num_token_union": 65.1122, "num_word_context": 202.541, "num_word_doc": 49.859, "num_word_query": 23.3135, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6147.2712, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4485, "query_norm": 1.4109, "queue_k_norm": 1.486, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3622, "sent_len_1": 66.8453, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.0962, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59100 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.5897, "doc_norm": 1.4894, "encoder_q-embeddings": 2817.0488, "encoder_q-layer.0": 1793.4618, "encoder_q-layer.1": 1890.9624, "encoder_q-layer.10": 3393.4634, "encoder_q-layer.11": 7320.4849, "encoder_q-layer.2": 2120.1804, "encoder_q-layer.3": 2168.834, "encoder_q-layer.4": 2348.1426, "encoder_q-layer.5": 2383.8716, "encoder_q-layer.6": 2553.4346, "encoder_q-layer.7": 2796.8982, "encoder_q-layer.8": 3435.3315, "encoder_q-layer.9": 3184.9102, "epoch": 0.26, "inbatch_neg_score": 0.4516, "inbatch_pos_score": 1.0449, "learning_rate": 3.7052631578947375e-05, "loss": 3.5897, "norm_diff": 0.0676, "norm_loss": 0.0, "num_token_doc": 66.7683, "num_token_overlap": 11.6666, "num_token_query": 31.3482, "num_token_union": 65.0962, "num_word_context": 202.146, "num_word_doc": 49.818, "num_word_query": 23.2861, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4824.0276, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4521, "query_norm": 1.4217, "queue_k_norm": 1.4898, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3482, "sent_len_1": 66.7683, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.34, "stdk": 0.0487, "stdq": 0.0437, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59200 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.5783, "doc_norm": 1.4976, "encoder_q-embeddings": 2550.2119, "encoder_q-layer.0": 1708.2249, "encoder_q-layer.1": 1819.5153, "encoder_q-layer.10": 3838.4019, "encoder_q-layer.11": 8400.6357, "encoder_q-layer.2": 1985.0502, "encoder_q-layer.3": 2082.7322, "encoder_q-layer.4": 2198.9099, "encoder_q-layer.5": 2322.011, "encoder_q-layer.6": 2566.595, "encoder_q-layer.7": 2895.387, "encoder_q-layer.8": 3853.428, "encoder_q-layer.9": 3580.7629, "epoch": 0.26, "inbatch_neg_score": 0.4537, "inbatch_pos_score": 1.041, "learning_rate": 3.7026315789473687e-05, "loss": 3.5783, "norm_diff": 0.1123, "norm_loss": 0.0, "num_token_doc": 66.9668, "num_token_overlap": 11.7031, "num_token_query": 31.3902, "num_token_union": 65.234, "num_word_context": 202.4089, "num_word_doc": 49.948, "num_word_query": 23.321, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5053.2906, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4536, "query_norm": 1.3853, "queue_k_norm": 1.4915, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3902, "sent_len_1": 66.9668, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.0675, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59300 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.5796, "doc_norm": 1.4883, "encoder_q-embeddings": 9879.1465, "encoder_q-layer.0": 8211.3682, "encoder_q-layer.1": 8788.3398, "encoder_q-layer.10": 3124.7073, "encoder_q-layer.11": 7458.8086, "encoder_q-layer.2": 10735.0762, "encoder_q-layer.3": 10820.7109, "encoder_q-layer.4": 13199.1855, "encoder_q-layer.5": 12018.4404, "encoder_q-layer.6": 10869.4902, "encoder_q-layer.7": 9608.459, "encoder_q-layer.8": 6076.0566, "encoder_q-layer.9": 3379.1523, "epoch": 0.26, "inbatch_neg_score": 0.4494, "inbatch_pos_score": 1.0273, "learning_rate": 3.7e-05, "loss": 3.5796, "norm_diff": 0.1071, "norm_loss": 0.0, "num_token_doc": 66.8663, "num_token_overlap": 11.6843, "num_token_query": 31.4474, "num_token_union": 65.2165, "num_word_context": 202.5101, "num_word_doc": 49.8633, "num_word_query": 23.3583, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13639.3718, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.45, "query_norm": 1.3812, "queue_k_norm": 1.4932, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4474, "sent_len_1": 66.8663, "sent_len_max_0": 127.9688, "sent_len_max_1": 210.3063, "stdk": 0.0485, "stdq": 0.0425, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59400 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5745, "doc_norm": 1.4958, "encoder_q-embeddings": 2614.1062, "encoder_q-layer.0": 1716.0674, "encoder_q-layer.1": 1799.2628, "encoder_q-layer.10": 3204.4036, "encoder_q-layer.11": 7237.3296, "encoder_q-layer.2": 2013.4491, "encoder_q-layer.3": 2103.7178, "encoder_q-layer.4": 2134.52, "encoder_q-layer.5": 2157.2507, "encoder_q-layer.6": 2589.9487, "encoder_q-layer.7": 2899.7075, "encoder_q-layer.8": 3503.5071, "encoder_q-layer.9": 3096.6494, "epoch": 0.26, "inbatch_neg_score": 0.4499, "inbatch_pos_score": 1.0498, "learning_rate": 3.6973684210526316e-05, "loss": 3.5745, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.6087, "num_token_overlap": 11.6835, "num_token_query": 31.3945, "num_token_union": 65.0156, "num_word_context": 202.1838, "num_word_doc": 49.7167, "num_word_query": 23.3109, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4786.8717, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4509, "query_norm": 1.4157, "queue_k_norm": 1.4952, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3945, "sent_len_1": 66.6087, "sent_len_max_0": 127.975, "sent_len_max_1": 208.8262, "stdk": 0.0487, "stdq": 0.0441, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 59500 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.5747, "doc_norm": 1.4989, "encoder_q-embeddings": 2902.7446, "encoder_q-layer.0": 1906.2803, "encoder_q-layer.1": 2004.4282, "encoder_q-layer.10": 3282.1577, "encoder_q-layer.11": 7760.6611, "encoder_q-layer.2": 2271.2676, "encoder_q-layer.3": 2440.5803, "encoder_q-layer.4": 2528.4302, "encoder_q-layer.5": 2626.7173, "encoder_q-layer.6": 2847.9258, "encoder_q-layer.7": 3169.3994, "encoder_q-layer.8": 3713.3259, "encoder_q-layer.9": 3242.147, "epoch": 0.26, "inbatch_neg_score": 0.4459, "inbatch_pos_score": 1.0391, "learning_rate": 3.6947368421052635e-05, "loss": 3.5747, "norm_diff": 0.1062, "norm_loss": 0.0, "num_token_doc": 66.8158, "num_token_overlap": 11.7123, "num_token_query": 31.4422, "num_token_union": 65.1722, "num_word_context": 202.3756, "num_word_doc": 49.8527, "num_word_query": 23.3569, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5151.9214, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4465, "query_norm": 1.3927, "queue_k_norm": 1.4944, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4422, "sent_len_1": 66.8158, "sent_len_max_0": 127.96, "sent_len_max_1": 207.34, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59600 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.5713, "doc_norm": 1.4992, "encoder_q-embeddings": 2681.0264, "encoder_q-layer.0": 1753.8335, "encoder_q-layer.1": 1837.3859, "encoder_q-layer.10": 3147.781, "encoder_q-layer.11": 7331.5469, "encoder_q-layer.2": 2020.2113, "encoder_q-layer.3": 2111.0051, "encoder_q-layer.4": 2208.9856, "encoder_q-layer.5": 2232.5574, "encoder_q-layer.6": 2542.1404, "encoder_q-layer.7": 2947.9148, "encoder_q-layer.8": 3457.6455, "encoder_q-layer.9": 3048.9163, "epoch": 0.26, "inbatch_neg_score": 0.4467, "inbatch_pos_score": 1.0342, "learning_rate": 3.692105263157895e-05, "loss": 3.5713, "norm_diff": 0.1029, "norm_loss": 0.0, "num_token_doc": 66.8173, "num_token_overlap": 11.6833, "num_token_query": 31.4594, "num_token_union": 65.1683, "num_word_context": 202.429, "num_word_doc": 49.8471, "num_word_query": 23.3622, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4826.127, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4468, "query_norm": 1.3962, "queue_k_norm": 1.4969, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4594, "sent_len_1": 66.8173, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.5838, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 59700 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.5739, "doc_norm": 1.4921, "encoder_q-embeddings": 2583.3201, "encoder_q-layer.0": 1833.2949, "encoder_q-layer.1": 1968.2892, "encoder_q-layer.10": 3148.3926, "encoder_q-layer.11": 7603.4907, "encoder_q-layer.2": 2166.0591, "encoder_q-layer.3": 2301.6558, "encoder_q-layer.4": 2339.7671, "encoder_q-layer.5": 2488.2437, "encoder_q-layer.6": 2773.8962, "encoder_q-layer.7": 2834.8105, "encoder_q-layer.8": 3411.7703, "encoder_q-layer.9": 3090.6272, "epoch": 0.26, "inbatch_neg_score": 0.4507, "inbatch_pos_score": 1.0342, "learning_rate": 3.6894736842105265e-05, "loss": 3.5739, "norm_diff": 0.1101, "norm_loss": 0.0, "num_token_doc": 66.6856, "num_token_overlap": 11.6798, "num_token_query": 31.4045, "num_token_union": 65.0807, "num_word_context": 202.1758, "num_word_doc": 49.7589, "num_word_query": 23.3322, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4832.4769, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4502, "query_norm": 1.382, "queue_k_norm": 1.4941, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4045, "sent_len_1": 66.6856, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.2975, "stdk": 0.0485, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 59800 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.5921, "doc_norm": 1.4979, "encoder_q-embeddings": 2598.051, "encoder_q-layer.0": 1737.061, "encoder_q-layer.1": 1834.6069, "encoder_q-layer.10": 3335.2053, "encoder_q-layer.11": 7569.5005, "encoder_q-layer.2": 2043.1753, "encoder_q-layer.3": 2117.2549, "encoder_q-layer.4": 2217.0645, "encoder_q-layer.5": 2220.5122, "encoder_q-layer.6": 2500.2869, "encoder_q-layer.7": 2828.0908, "encoder_q-layer.8": 3397.8833, "encoder_q-layer.9": 3141.3372, "epoch": 0.26, "inbatch_neg_score": 0.4482, "inbatch_pos_score": 1.0322, "learning_rate": 3.686842105263158e-05, "loss": 3.5921, "norm_diff": 0.112, "norm_loss": 0.0, "num_token_doc": 66.6967, "num_token_overlap": 11.6566, "num_token_query": 31.3254, "num_token_union": 65.0325, "num_word_context": 202.2989, "num_word_doc": 49.7634, "num_word_query": 23.2727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4831.7176, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.448, "query_norm": 1.3859, "queue_k_norm": 1.4991, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3254, "sent_len_1": 66.6967, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.9525, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 59900 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.5774, "doc_norm": 1.4955, "encoder_q-embeddings": 2806.0498, "encoder_q-layer.0": 1782.8628, "encoder_q-layer.1": 1917.9309, "encoder_q-layer.10": 3129.937, "encoder_q-layer.11": 7170.5771, "encoder_q-layer.2": 2134.0381, "encoder_q-layer.3": 2183.7266, "encoder_q-layer.4": 2259.6047, "encoder_q-layer.5": 2363.4419, "encoder_q-layer.6": 2628.6748, "encoder_q-layer.7": 2819.7627, "encoder_q-layer.8": 3401.5112, "encoder_q-layer.9": 3232.3452, "epoch": 0.26, "inbatch_neg_score": 0.4413, "inbatch_pos_score": 1.0332, "learning_rate": 3.6842105263157895e-05, "loss": 3.5774, "norm_diff": 0.1078, "norm_loss": 0.0, "num_token_doc": 66.8599, "num_token_overlap": 11.697, "num_token_query": 31.3494, "num_token_union": 65.1507, "num_word_context": 202.4562, "num_word_doc": 49.8882, "num_word_query": 23.2806, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4795.255, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4419, "query_norm": 1.3877, "queue_k_norm": 1.4981, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3494, "sent_len_1": 66.8599, "sent_len_max_0": 127.9688, "sent_len_max_1": 208.38, "stdk": 0.0486, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60000 }, { "dev_runtime": 28.2775, "dev_samples_per_second": 1.132, "dev_steps_per_second": 0.035, "epoch": 0.26, "step": 60000, "test_accuracy": 93.65234375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3680509328842163, "test_doc_norm": 1.4676713943481445, "test_inbatch_neg_score": 0.7163667678833008, "test_inbatch_pos_score": 1.7080531120300293, "test_loss": 0.3680509328842163, "test_loss_align": 1.0572949647903442, "test_loss_unif": 3.6132149696350098, "test_loss_unif_q@queue": 3.6132149696350098, "test_norm_diff": 0.046140965074300766, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.4373435080051422, "test_query_norm": 1.5138123035430908, "test_queue_k_norm": 1.4978995323181152, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04342241585254669, "test_stdq": 0.043796107172966, "test_stdqueue_k": 0.04870619624853134, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.2775, "dev_samples_per_second": 1.132, "dev_steps_per_second": 0.035, "epoch": 0.26, "eval_beir-arguana_ndcg@10": 0.35617, "eval_beir-arguana_recall@10": 0.60669, "eval_beir-arguana_recall@100": 0.9175, "eval_beir-arguana_recall@20": 0.74324, "eval_beir-avg_ndcg@10": 0.39074558333333337, "eval_beir-avg_recall@10": 0.4575633333333333, "eval_beir-avg_recall@100": 0.6421240000000001, "eval_beir-avg_recall@20": 0.5203675000000001, "eval_beir-cqadupstack_ndcg@10": 0.2714258333333333, "eval_beir-cqadupstack_recall@10": 0.36738333333333334, "eval_beir-cqadupstack_recall@100": 0.6027600000000001, "eval_beir-cqadupstack_recall@20": 0.43490500000000004, "eval_beir-fiqa_ndcg@10": 0.25844, "eval_beir-fiqa_recall@10": 0.31906, "eval_beir-fiqa_recall@100": 0.59439, "eval_beir-fiqa_recall@20": 0.39215, "eval_beir-nfcorpus_ndcg@10": 0.30936, "eval_beir-nfcorpus_recall@10": 0.14909, "eval_beir-nfcorpus_recall@100": 0.29042, "eval_beir-nfcorpus_recall@20": 0.18914, "eval_beir-nq_ndcg@10": 0.3121, "eval_beir-nq_recall@10": 0.5015, "eval_beir-nq_recall@100": 0.82887, "eval_beir-nq_recall@20": 0.61648, "eval_beir-quora_ndcg@10": 0.78251, "eval_beir-quora_recall@10": 0.89091, "eval_beir-quora_recall@100": 0.97813, "eval_beir-quora_recall@20": 0.93219, "eval_beir-scidocs_ndcg@10": 0.1551, "eval_beir-scidocs_recall@10": 0.16297, "eval_beir-scidocs_recall@100": 0.3704, "eval_beir-scidocs_recall@20": 0.22037, "eval_beir-scifact_ndcg@10": 0.65433, "eval_beir-scifact_recall@10": 0.79111, "eval_beir-scifact_recall@100": 0.91933, "eval_beir-scifact_recall@20": 0.84078, "eval_beir-trec-covid_ndcg@10": 0.60488, "eval_beir-trec-covid_recall@10": 0.648, "eval_beir-trec-covid_recall@100": 0.4778, "eval_beir-trec-covid_recall@20": 0.632, "eval_beir-webis-touche2020_ndcg@10": 0.20314, "eval_beir-webis-touche2020_recall@10": 0.13892, "eval_beir-webis-touche2020_recall@100": 0.44164, "eval_beir-webis-touche2020_recall@20": 0.20242, "eval_senteval-avg_sts": 0.75589917638254, "eval_senteval-sickr_spearman": 0.7304507466276304, "eval_senteval-stsb_spearman": 0.7813476061374496, "step": 60000, "test_accuracy": 93.65234375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3680509328842163, "test_doc_norm": 1.4676713943481445, "test_inbatch_neg_score": 0.7163667678833008, "test_inbatch_pos_score": 1.7080531120300293, "test_loss": 0.3680509328842163, "test_loss_align": 1.0572949647903442, "test_loss_unif": 3.6132149696350098, "test_loss_unif_q@queue": 3.6132149696350098, "test_norm_diff": 0.046140965074300766, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.4373435080051422, "test_query_norm": 1.5138123035430908, "test_queue_k_norm": 1.4978995323181152, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04342241585254669, "test_stdq": 0.043796107172966, "test_stdqueue_k": 0.04870619624853134, "test_stdqueue_q": 0.0 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.5742, "doc_norm": 1.5001, "encoder_q-embeddings": 3108.665, "encoder_q-layer.0": 2051.8293, "encoder_q-layer.1": 2140.8784, "encoder_q-layer.10": 3468.2388, "encoder_q-layer.11": 7993.3828, "encoder_q-layer.2": 2400.0752, "encoder_q-layer.3": 2449.2852, "encoder_q-layer.4": 2599.0466, "encoder_q-layer.5": 2605.4841, "encoder_q-layer.6": 2762.2427, "encoder_q-layer.7": 3206.7441, "encoder_q-layer.8": 3832.1428, "encoder_q-layer.9": 3341.9023, "epoch": 0.26, "inbatch_neg_score": 0.4447, "inbatch_pos_score": 1.0332, "learning_rate": 3.681578947368421e-05, "loss": 3.5742, "norm_diff": 0.1055, "norm_loss": 0.0, "num_token_doc": 66.6872, "num_token_overlap": 11.671, "num_token_query": 31.3449, "num_token_union": 65.0879, "num_word_context": 202.2456, "num_word_doc": 49.7507, "num_word_query": 23.2717, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5310.1954, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4451, "query_norm": 1.3946, "queue_k_norm": 1.4979, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3449, "sent_len_1": 66.6872, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.0225, "stdk": 0.0487, "stdq": 0.0433, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60100 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.5687, "doc_norm": 1.4973, "encoder_q-embeddings": 2608.0757, "encoder_q-layer.0": 1676.8024, "encoder_q-layer.1": 1791.6467, "encoder_q-layer.10": 3174.6636, "encoder_q-layer.11": 7488.7363, "encoder_q-layer.2": 1979.9449, "encoder_q-layer.3": 1987.3595, "encoder_q-layer.4": 2151.9663, "encoder_q-layer.5": 2221.5283, "encoder_q-layer.6": 2435.2373, "encoder_q-layer.7": 2849.7671, "encoder_q-layer.8": 3538.5491, "encoder_q-layer.9": 3107.22, "epoch": 0.26, "inbatch_neg_score": 0.4458, "inbatch_pos_score": 1.0244, "learning_rate": 3.678947368421053e-05, "loss": 3.5687, "norm_diff": 0.114, "norm_loss": 0.0, "num_token_doc": 66.7107, "num_token_overlap": 11.6881, "num_token_query": 31.4156, "num_token_union": 65.0987, "num_word_context": 202.2283, "num_word_doc": 49.7978, "num_word_query": 23.3431, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4820.6404, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4451, "query_norm": 1.3833, "queue_k_norm": 1.5003, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4156, "sent_len_1": 66.7107, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.3638, "stdk": 0.0486, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 60200 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.566, "doc_norm": 1.504, "encoder_q-embeddings": 2540.9963, "encoder_q-layer.0": 1752.8041, "encoder_q-layer.1": 1842.989, "encoder_q-layer.10": 3381.9768, "encoder_q-layer.11": 7676.8213, "encoder_q-layer.2": 2019.5707, "encoder_q-layer.3": 2059.25, "encoder_q-layer.4": 2172.5706, "encoder_q-layer.5": 2175.7302, "encoder_q-layer.6": 2516.7129, "encoder_q-layer.7": 3011.8645, "encoder_q-layer.8": 3529.9077, "encoder_q-layer.9": 3320.9814, "epoch": 0.26, "inbatch_neg_score": 0.4444, "inbatch_pos_score": 1.0439, "learning_rate": 3.676315789473684e-05, "loss": 3.566, "norm_diff": 0.1175, "norm_loss": 0.0, "num_token_doc": 66.6791, "num_token_overlap": 11.7063, "num_token_query": 31.4711, "num_token_union": 65.1364, "num_word_context": 202.1222, "num_word_doc": 49.7895, "num_word_query": 23.3829, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4849.6694, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4453, "query_norm": 1.3865, "queue_k_norm": 1.4997, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4711, "sent_len_1": 66.6791, "sent_len_max_0": 127.975, "sent_len_max_1": 206.3325, "stdk": 0.0488, "stdq": 0.043, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60300 }, { "accuracy": 46.3379, "active_queue_size": 16384.0, "cl_loss": 3.589, "doc_norm": 1.5004, "encoder_q-embeddings": 2908.5676, "encoder_q-layer.0": 1965.6376, "encoder_q-layer.1": 2170.3486, "encoder_q-layer.10": 3497.978, "encoder_q-layer.11": 7676.314, "encoder_q-layer.2": 2498.8782, "encoder_q-layer.3": 2613.7778, "encoder_q-layer.4": 2875.366, "encoder_q-layer.5": 2843.3452, "encoder_q-layer.6": 2991.2346, "encoder_q-layer.7": 3082.6653, "encoder_q-layer.8": 3528.0366, "encoder_q-layer.9": 3452.2485, "epoch": 0.26, "inbatch_neg_score": 0.4471, "inbatch_pos_score": 1.0186, "learning_rate": 3.673684210526316e-05, "loss": 3.589, "norm_diff": 0.1183, "norm_loss": 0.0, "num_token_doc": 66.6579, "num_token_overlap": 11.6457, "num_token_query": 31.2632, "num_token_union": 65.0201, "num_word_context": 202.183, "num_word_doc": 49.7632, "num_word_query": 23.2228, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5182.8129, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4478, "query_norm": 1.3821, "queue_k_norm": 1.4983, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2632, "sent_len_1": 66.6579, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.2837, "stdk": 0.0487, "stdq": 0.0427, "stdqueue_k": 0.0486, "stdqueue_q": 0.0, "step": 60400 }, { "accuracy": 46.8262, "active_queue_size": 16384.0, "cl_loss": 3.5733, "doc_norm": 1.4982, "encoder_q-embeddings": 2743.3369, "encoder_q-layer.0": 1826.2408, "encoder_q-layer.1": 1931.8781, "encoder_q-layer.10": 3012.1814, "encoder_q-layer.11": 7209.9399, "encoder_q-layer.2": 2239.6897, "encoder_q-layer.3": 2256.0479, "encoder_q-layer.4": 2382.0923, "encoder_q-layer.5": 2515.5542, "encoder_q-layer.6": 2708.7034, "encoder_q-layer.7": 2895.7444, "encoder_q-layer.8": 3137.8682, "encoder_q-layer.9": 3037.9883, "epoch": 0.26, "inbatch_neg_score": 0.4482, "inbatch_pos_score": 1.0176, "learning_rate": 3.671052631578947e-05, "loss": 3.5733, "norm_diff": 0.117, "norm_loss": 0.0, "num_token_doc": 66.6738, "num_token_overlap": 11.7236, "num_token_query": 31.4575, "num_token_union": 65.072, "num_word_context": 202.3054, "num_word_doc": 49.7667, "num_word_query": 23.3517, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4783.166, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4482, "query_norm": 1.3812, "queue_k_norm": 1.5004, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4575, "sent_len_1": 66.6738, "sent_len_max_0": 128.0, "sent_len_max_1": 209.5925, "stdk": 0.0486, "stdq": 0.0426, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60500 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.5715, "doc_norm": 1.5029, "encoder_q-embeddings": 5238.7793, "encoder_q-layer.0": 3442.6406, "encoder_q-layer.1": 3701.5256, "encoder_q-layer.10": 6403.9395, "encoder_q-layer.11": 15229.79, "encoder_q-layer.2": 4241.1187, "encoder_q-layer.3": 4267.9478, "encoder_q-layer.4": 4496.5757, "encoder_q-layer.5": 4705.8164, "encoder_q-layer.6": 5495.6606, "encoder_q-layer.7": 5964.7095, "encoder_q-layer.8": 7015.4458, "encoder_q-layer.9": 6054.1831, "epoch": 0.26, "inbatch_neg_score": 0.4489, "inbatch_pos_score": 1.0293, "learning_rate": 3.668421052631579e-05, "loss": 3.5715, "norm_diff": 0.1186, "norm_loss": 0.0, "num_token_doc": 66.9083, "num_token_overlap": 11.7321, "num_token_query": 31.5175, "num_token_union": 65.2407, "num_word_context": 202.5318, "num_word_doc": 49.9392, "num_word_query": 23.4156, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9968.4146, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4492, "query_norm": 1.3843, "queue_k_norm": 1.5014, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5175, "sent_len_1": 66.9083, "sent_len_max_0": 127.995, "sent_len_max_1": 209.7413, "stdk": 0.0487, "stdq": 0.0428, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60600 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.5774, "doc_norm": 1.5045, "encoder_q-embeddings": 5999.1001, "encoder_q-layer.0": 4003.9712, "encoder_q-layer.1": 4548.2412, "encoder_q-layer.10": 6604.1118, "encoder_q-layer.11": 14683.334, "encoder_q-layer.2": 5148.2949, "encoder_q-layer.3": 5429.145, "encoder_q-layer.4": 5710.2734, "encoder_q-layer.5": 5942.0474, "encoder_q-layer.6": 6563.6963, "encoder_q-layer.7": 6637.3975, "encoder_q-layer.8": 7111.8433, "encoder_q-layer.9": 6172.8872, "epoch": 0.26, "inbatch_neg_score": 0.4469, "inbatch_pos_score": 1.04, "learning_rate": 3.665789473684211e-05, "loss": 3.5774, "norm_diff": 0.1103, "norm_loss": 0.0, "num_token_doc": 66.6031, "num_token_overlap": 11.653, "num_token_query": 31.3077, "num_token_union": 65.0107, "num_word_context": 202.1573, "num_word_doc": 49.6828, "num_word_query": 23.2471, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10368.7528, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4473, "query_norm": 1.3942, "queue_k_norm": 1.5017, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3077, "sent_len_1": 66.6031, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.785, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60700 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.5667, "doc_norm": 1.5063, "encoder_q-embeddings": 5355.4219, "encoder_q-layer.0": 3618.3086, "encoder_q-layer.1": 3903.5073, "encoder_q-layer.10": 6385.9229, "encoder_q-layer.11": 14871.2344, "encoder_q-layer.2": 4366.6904, "encoder_q-layer.3": 4494.3047, "encoder_q-layer.4": 4827.4561, "encoder_q-layer.5": 5096.2588, "encoder_q-layer.6": 5754.5195, "encoder_q-layer.7": 6271.0811, "encoder_q-layer.8": 7174.9116, "encoder_q-layer.9": 6243.2314, "epoch": 0.26, "inbatch_neg_score": 0.4501, "inbatch_pos_score": 1.0332, "learning_rate": 3.663157894736842e-05, "loss": 3.5667, "norm_diff": 0.1172, "norm_loss": 0.0, "num_token_doc": 66.9496, "num_token_overlap": 11.6943, "num_token_query": 31.5048, "num_token_union": 65.307, "num_word_context": 202.4882, "num_word_doc": 49.9767, "num_word_query": 23.4276, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9984.4687, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4502, "query_norm": 1.389, "queue_k_norm": 1.504, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5048, "sent_len_1": 66.9496, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.17, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 60800 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.5637, "doc_norm": 1.5073, "encoder_q-embeddings": 5956.4121, "encoder_q-layer.0": 3942.0088, "encoder_q-layer.1": 4181.7485, "encoder_q-layer.10": 6339.2119, "encoder_q-layer.11": 14871.8711, "encoder_q-layer.2": 4755.4395, "encoder_q-layer.3": 4969.2212, "encoder_q-layer.4": 5466.3008, "encoder_q-layer.5": 5470.5811, "encoder_q-layer.6": 5844.1055, "encoder_q-layer.7": 6654.2119, "encoder_q-layer.8": 7029.4956, "encoder_q-layer.9": 6499.689, "epoch": 0.26, "inbatch_neg_score": 0.4584, "inbatch_pos_score": 1.0459, "learning_rate": 3.660526315789474e-05, "loss": 3.5637, "norm_diff": 0.1076, "norm_loss": 0.0, "num_token_doc": 66.7303, "num_token_overlap": 11.6703, "num_token_query": 31.3443, "num_token_union": 65.07, "num_word_context": 202.328, "num_word_doc": 49.8004, "num_word_query": 23.2888, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10180.3407, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4578, "query_norm": 1.3996, "queue_k_norm": 1.5031, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3443, "sent_len_1": 66.7303, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.8225, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 60900 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.5737, "doc_norm": 1.4965, "encoder_q-embeddings": 5648.021, "encoder_q-layer.0": 3724.4668, "encoder_q-layer.1": 3935.9216, "encoder_q-layer.10": 6540.5151, "encoder_q-layer.11": 14435.7773, "encoder_q-layer.2": 4371.5044, "encoder_q-layer.3": 4422.5981, "encoder_q-layer.4": 4611.1958, "encoder_q-layer.5": 4613.752, "encoder_q-layer.6": 5008.4121, "encoder_q-layer.7": 5709.2632, "encoder_q-layer.8": 6994.6509, "encoder_q-layer.9": 6325.0078, "epoch": 0.26, "inbatch_neg_score": 0.4556, "inbatch_pos_score": 1.042, "learning_rate": 3.657894736842106e-05, "loss": 3.5737, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.6814, "num_token_overlap": 11.664, "num_token_query": 31.4067, "num_token_union": 65.1078, "num_word_context": 202.3297, "num_word_doc": 49.762, "num_word_query": 23.3224, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9568.0212, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4563, "query_norm": 1.4156, "queue_k_norm": 1.5031, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4067, "sent_len_1": 66.6814, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.9638, "stdk": 0.0484, "stdq": 0.0437, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61000 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.5768, "doc_norm": 1.4986, "encoder_q-embeddings": 5283.3267, "encoder_q-layer.0": 3438.5488, "encoder_q-layer.1": 3700.9827, "encoder_q-layer.10": 6880.0854, "encoder_q-layer.11": 14939.3047, "encoder_q-layer.2": 4131.375, "encoder_q-layer.3": 4431.04, "encoder_q-layer.4": 4705.5278, "encoder_q-layer.5": 4698.9438, "encoder_q-layer.6": 5395.4956, "encoder_q-layer.7": 6012.269, "encoder_q-layer.8": 7428.8784, "encoder_q-layer.9": 6524.4268, "epoch": 0.27, "inbatch_neg_score": 0.4614, "inbatch_pos_score": 1.0498, "learning_rate": 3.655263157894737e-05, "loss": 3.5768, "norm_diff": 0.0848, "norm_loss": 0.0, "num_token_doc": 66.7731, "num_token_overlap": 11.6765, "num_token_query": 31.4373, "num_token_union": 65.1652, "num_word_context": 202.4188, "num_word_doc": 49.8093, "num_word_query": 23.3608, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9654.715, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4619, "query_norm": 1.4138, "queue_k_norm": 1.5048, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4373, "sent_len_1": 66.7731, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.7188, "stdk": 0.0485, "stdq": 0.0434, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61100 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5678, "doc_norm": 1.5065, "encoder_q-embeddings": 5341.5317, "encoder_q-layer.0": 3694.0527, "encoder_q-layer.1": 3952.9736, "encoder_q-layer.10": 6722.0522, "encoder_q-layer.11": 14625.1768, "encoder_q-layer.2": 4478.8691, "encoder_q-layer.3": 4580.9785, "encoder_q-layer.4": 4946.6353, "encoder_q-layer.5": 5111.0786, "encoder_q-layer.6": 5162.1108, "encoder_q-layer.7": 5767.8418, "encoder_q-layer.8": 6698.5869, "encoder_q-layer.9": 6149.9404, "epoch": 0.27, "inbatch_neg_score": 0.4661, "inbatch_pos_score": 1.0547, "learning_rate": 3.652631578947369e-05, "loss": 3.5678, "norm_diff": 0.0983, "norm_loss": 0.0, "num_token_doc": 66.7816, "num_token_overlap": 11.7092, "num_token_query": 31.426, "num_token_union": 65.108, "num_word_context": 202.3526, "num_word_doc": 49.811, "num_word_query": 23.3409, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9573.7587, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4656, "query_norm": 1.4083, "queue_k_norm": 1.5053, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.426, "sent_len_1": 66.7816, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.24, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61200 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.5734, "doc_norm": 1.5082, "encoder_q-embeddings": 5129.522, "encoder_q-layer.0": 3405.3777, "encoder_q-layer.1": 3609.5779, "encoder_q-layer.10": 7044.6147, "encoder_q-layer.11": 14611.4004, "encoder_q-layer.2": 3958.1599, "encoder_q-layer.3": 4081.3657, "encoder_q-layer.4": 4328.2251, "encoder_q-layer.5": 4536.3228, "encoder_q-layer.6": 5213.7222, "encoder_q-layer.7": 5635.3032, "encoder_q-layer.8": 6913.5532, "encoder_q-layer.9": 6395.5557, "epoch": 0.27, "inbatch_neg_score": 0.4733, "inbatch_pos_score": 1.0527, "learning_rate": 3.65e-05, "loss": 3.5734, "norm_diff": 0.0909, "norm_loss": 0.0, "num_token_doc": 66.7969, "num_token_overlap": 11.6051, "num_token_query": 31.2877, "num_token_union": 65.1178, "num_word_context": 202.3729, "num_word_doc": 49.8393, "num_word_query": 23.2316, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9529.7985, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.4734, "query_norm": 1.4174, "queue_k_norm": 1.5054, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2877, "sent_len_1": 66.7969, "sent_len_max_0": 127.98, "sent_len_max_1": 208.7425, "stdk": 0.0488, "stdq": 0.0428, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61300 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.5658, "doc_norm": 1.5102, "encoder_q-embeddings": 2614.9741, "encoder_q-layer.0": 1763.2834, "encoder_q-layer.1": 1861.6155, "encoder_q-layer.10": 3352.2622, "encoder_q-layer.11": 7577.269, "encoder_q-layer.2": 2101.0391, "encoder_q-layer.3": 2102.9971, "encoder_q-layer.4": 2267.5112, "encoder_q-layer.5": 2321.2009, "encoder_q-layer.6": 2555.8071, "encoder_q-layer.7": 2984.8364, "encoder_q-layer.8": 3615.396, "encoder_q-layer.9": 3192.5308, "epoch": 0.27, "inbatch_neg_score": 0.481, "inbatch_pos_score": 1.0801, "learning_rate": 3.647368421052632e-05, "loss": 3.5658, "norm_diff": 0.0626, "norm_loss": 0.0, "num_token_doc": 66.6874, "num_token_overlap": 11.6923, "num_token_query": 31.4927, "num_token_union": 65.1057, "num_word_context": 202.3265, "num_word_doc": 49.7531, "num_word_query": 23.3865, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4825.9182, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4807, "query_norm": 1.4476, "queue_k_norm": 1.5089, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4927, "sent_len_1": 66.6874, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.5637, "stdk": 0.0488, "stdq": 0.0439, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 61400 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.5732, "doc_norm": 1.5061, "encoder_q-embeddings": 3289.3696, "encoder_q-layer.0": 2217.5195, "encoder_q-layer.1": 2590.626, "encoder_q-layer.10": 3306.4675, "encoder_q-layer.11": 7137.1333, "encoder_q-layer.2": 2886.4905, "encoder_q-layer.3": 2820.7844, "encoder_q-layer.4": 2998.8003, "encoder_q-layer.5": 3034.7544, "encoder_q-layer.6": 3165.4131, "encoder_q-layer.7": 3247.4099, "encoder_q-layer.8": 3492.4805, "encoder_q-layer.9": 3091.3608, "epoch": 0.27, "inbatch_neg_score": 0.4873, "inbatch_pos_score": 1.082, "learning_rate": 3.6447368421052636e-05, "loss": 3.5732, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.6003, "num_token_overlap": 11.6569, "num_token_query": 31.4494, "num_token_union": 65.0719, "num_word_context": 202.149, "num_word_doc": 49.7075, "num_word_query": 23.3594, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5265.5646, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.4873, "query_norm": 1.4362, "queue_k_norm": 1.5074, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4494, "sent_len_1": 66.6003, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.9625, "stdk": 0.0486, "stdq": 0.0431, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61500 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.5812, "doc_norm": 1.5064, "encoder_q-embeddings": 1327.2595, "encoder_q-layer.0": 879.7543, "encoder_q-layer.1": 926.7469, "encoder_q-layer.10": 1720.3232, "encoder_q-layer.11": 3946.9099, "encoder_q-layer.2": 1064.3921, "encoder_q-layer.3": 1094.5742, "encoder_q-layer.4": 1154.0344, "encoder_q-layer.5": 1187.6558, "encoder_q-layer.6": 1366.0554, "encoder_q-layer.7": 1415.5411, "encoder_q-layer.8": 1760.8315, "encoder_q-layer.9": 1566.2655, "epoch": 0.27, "inbatch_neg_score": 0.5009, "inbatch_pos_score": 1.0811, "learning_rate": 3.642105263157895e-05, "loss": 3.5812, "norm_diff": 0.0604, "norm_loss": 0.0, "num_token_doc": 66.6403, "num_token_overlap": 11.6594, "num_token_query": 31.3944, "num_token_union": 65.0717, "num_word_context": 202.0509, "num_word_doc": 49.7184, "num_word_query": 23.3055, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2479.9645, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5005, "query_norm": 1.4461, "queue_k_norm": 1.509, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3944, "sent_len_1": 66.6403, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.1, "stdk": 0.0486, "stdq": 0.0432, "stdqueue_k": 0.0487, "stdqueue_q": 0.0, "step": 61600 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.576, "doc_norm": 1.5117, "encoder_q-embeddings": 1264.9022, "encoder_q-layer.0": 881.4426, "encoder_q-layer.1": 942.2673, "encoder_q-layer.10": 1852.7848, "encoder_q-layer.11": 3794.9465, "encoder_q-layer.2": 1042.8273, "encoder_q-layer.3": 1085.7938, "encoder_q-layer.4": 1171.2125, "encoder_q-layer.5": 1270.1619, "encoder_q-layer.6": 1357.5406, "encoder_q-layer.7": 1451.567, "encoder_q-layer.8": 1723.6956, "encoder_q-layer.9": 1618.1576, "epoch": 0.27, "inbatch_neg_score": 0.5057, "inbatch_pos_score": 1.0791, "learning_rate": 3.6394736842105266e-05, "loss": 3.576, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.7759, "num_token_overlap": 11.6357, "num_token_query": 31.3073, "num_token_union": 65.1047, "num_word_context": 202.4378, "num_word_doc": 49.8004, "num_word_query": 23.2518, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2432.5119, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5059, "query_norm": 1.4348, "queue_k_norm": 1.5111, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3073, "sent_len_1": 66.7759, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.6738, "stdk": 0.0487, "stdq": 0.043, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 61700 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.5623, "doc_norm": 1.5132, "encoder_q-embeddings": 1401.5872, "encoder_q-layer.0": 940.7556, "encoder_q-layer.1": 1022.0374, "encoder_q-layer.10": 1644.2058, "encoder_q-layer.11": 3763.9387, "encoder_q-layer.2": 1129.8928, "encoder_q-layer.3": 1146.5828, "encoder_q-layer.4": 1198.4298, "encoder_q-layer.5": 1246.77, "encoder_q-layer.6": 1395.0793, "encoder_q-layer.7": 1684.1967, "encoder_q-layer.8": 1892.2865, "encoder_q-layer.9": 1620.0177, "epoch": 0.27, "inbatch_neg_score": 0.5122, "inbatch_pos_score": 1.0918, "learning_rate": 3.636842105263158e-05, "loss": 3.5623, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.8452, "num_token_overlap": 11.6924, "num_token_query": 31.3284, "num_token_union": 65.1381, "num_word_context": 202.226, "num_word_doc": 49.8629, "num_word_query": 23.2628, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2551.3592, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5127, "query_norm": 1.4283, "queue_k_norm": 1.5126, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3284, "sent_len_1": 66.8452, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.9625, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 61800 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.5678, "doc_norm": 1.5137, "encoder_q-embeddings": 1440.0896, "encoder_q-layer.0": 954.6337, "encoder_q-layer.1": 1015.6183, "encoder_q-layer.10": 1642.7455, "encoder_q-layer.11": 3807.875, "encoder_q-layer.2": 1135.1594, "encoder_q-layer.3": 1157.2059, "encoder_q-layer.4": 1224.7373, "encoder_q-layer.5": 1235.0444, "encoder_q-layer.6": 1363.244, "encoder_q-layer.7": 1488.7803, "encoder_q-layer.8": 1850.8953, "encoder_q-layer.9": 1636.4492, "epoch": 0.27, "inbatch_neg_score": 0.5129, "inbatch_pos_score": 1.0996, "learning_rate": 3.6342105263157896e-05, "loss": 3.5678, "norm_diff": 0.0904, "norm_loss": 0.0, "num_token_doc": 66.9313, "num_token_overlap": 11.6597, "num_token_query": 31.344, "num_token_union": 65.1998, "num_word_context": 202.3573, "num_word_doc": 49.9461, "num_word_query": 23.2801, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2544.1565, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5132, "query_norm": 1.4233, "queue_k_norm": 1.5138, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.344, "sent_len_1": 66.9313, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.7325, "stdk": 0.0487, "stdq": 0.0431, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 61900 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.56, "doc_norm": 1.5183, "encoder_q-embeddings": 1368.3226, "encoder_q-layer.0": 932.5521, "encoder_q-layer.1": 970.9312, "encoder_q-layer.10": 1572.3121, "encoder_q-layer.11": 3663.9868, "encoder_q-layer.2": 1063.2113, "encoder_q-layer.3": 1085.8169, "encoder_q-layer.4": 1190.8179, "encoder_q-layer.5": 1201.1198, "encoder_q-layer.6": 1306.83, "encoder_q-layer.7": 1481.6592, "encoder_q-layer.8": 1733.4833, "encoder_q-layer.9": 1630.9797, "epoch": 0.27, "inbatch_neg_score": 0.5138, "inbatch_pos_score": 1.1191, "learning_rate": 3.6315789473684214e-05, "loss": 3.56, "norm_diff": 0.0939, "norm_loss": 0.0, "num_token_doc": 66.7242, "num_token_overlap": 11.6726, "num_token_query": 31.2926, "num_token_union": 65.0701, "num_word_context": 202.2972, "num_word_doc": 49.8045, "num_word_query": 23.2291, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2404.5049, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5146, "query_norm": 1.4244, "queue_k_norm": 1.5175, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2926, "sent_len_1": 66.7242, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.7713, "stdk": 0.0489, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 62000 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.5898, "doc_norm": 1.517, "encoder_q-embeddings": 1251.8635, "encoder_q-layer.0": 853.2339, "encoder_q-layer.1": 906.912, "encoder_q-layer.10": 1610.0823, "encoder_q-layer.11": 3704.4968, "encoder_q-layer.2": 994.2833, "encoder_q-layer.3": 1022.2183, "encoder_q-layer.4": 1082.7734, "encoder_q-layer.5": 1102.5983, "encoder_q-layer.6": 1252.577, "encoder_q-layer.7": 1439.277, "encoder_q-layer.8": 1670.2297, "encoder_q-layer.9": 1507.1692, "epoch": 0.27, "inbatch_neg_score": 0.5169, "inbatch_pos_score": 1.1133, "learning_rate": 3.6289473684210526e-05, "loss": 3.5898, "norm_diff": 0.0944, "norm_loss": 0.0, "num_token_doc": 66.603, "num_token_overlap": 11.6097, "num_token_query": 31.3033, "num_token_union": 65.0271, "num_word_context": 202.0992, "num_word_doc": 49.6904, "num_word_query": 23.2412, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2365.6895, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5176, "query_norm": 1.4226, "queue_k_norm": 1.5171, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3033, "sent_len_1": 66.603, "sent_len_max_0": 127.97, "sent_len_max_1": 207.6987, "stdk": 0.0487, "stdq": 0.0434, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62100 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.5669, "doc_norm": 1.5198, "encoder_q-embeddings": 1351.4907, "encoder_q-layer.0": 912.3354, "encoder_q-layer.1": 949.8045, "encoder_q-layer.10": 1615.3403, "encoder_q-layer.11": 3757.4146, "encoder_q-layer.2": 1064.889, "encoder_q-layer.3": 1094.2104, "encoder_q-layer.4": 1140.2717, "encoder_q-layer.5": 1122.8268, "encoder_q-layer.6": 1272.8843, "encoder_q-layer.7": 1440.4995, "encoder_q-layer.8": 1751.2317, "encoder_q-layer.9": 1629.2534, "epoch": 0.27, "inbatch_neg_score": 0.5149, "inbatch_pos_score": 1.1064, "learning_rate": 3.6263157894736844e-05, "loss": 3.5669, "norm_diff": 0.1074, "norm_loss": 0.0, "num_token_doc": 66.8901, "num_token_overlap": 11.6528, "num_token_query": 31.3363, "num_token_union": 65.1739, "num_word_context": 202.4295, "num_word_doc": 49.8897, "num_word_query": 23.2567, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2447.3377, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5166, "query_norm": 1.4123, "queue_k_norm": 1.5186, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3363, "sent_len_1": 66.8901, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.5637, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62200 }, { "accuracy": 47.2168, "active_queue_size": 16384.0, "cl_loss": 3.5688, "doc_norm": 1.5197, "encoder_q-embeddings": 1361.3186, "encoder_q-layer.0": 892.2312, "encoder_q-layer.1": 940.3934, "encoder_q-layer.10": 1861.4548, "encoder_q-layer.11": 4027.3452, "encoder_q-layer.2": 1020.9283, "encoder_q-layer.3": 1082.7915, "encoder_q-layer.4": 1130.9827, "encoder_q-layer.5": 1161.6926, "encoder_q-layer.6": 1295.8287, "encoder_q-layer.7": 1599.4458, "encoder_q-layer.8": 1930.9955, "encoder_q-layer.9": 1738.542, "epoch": 0.27, "inbatch_neg_score": 0.5171, "inbatch_pos_score": 1.0908, "learning_rate": 3.6236842105263156e-05, "loss": 3.5688, "norm_diff": 0.1065, "norm_loss": 0.0, "num_token_doc": 66.9174, "num_token_overlap": 11.689, "num_token_query": 31.4253, "num_token_union": 65.2544, "num_word_context": 202.6282, "num_word_doc": 49.9285, "num_word_query": 23.344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2593.4118, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5166, "query_norm": 1.4132, "queue_k_norm": 1.5197, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4253, "sent_len_1": 66.9174, "sent_len_max_0": 127.995, "sent_len_max_1": 208.5062, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62300 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.5697, "doc_norm": 1.5198, "encoder_q-embeddings": 1274.8463, "encoder_q-layer.0": 867.8116, "encoder_q-layer.1": 914.1247, "encoder_q-layer.10": 1627.0546, "encoder_q-layer.11": 3804.658, "encoder_q-layer.2": 977.4592, "encoder_q-layer.3": 981.7233, "encoder_q-layer.4": 1042.704, "encoder_q-layer.5": 1103.3669, "encoder_q-layer.6": 1241.1118, "encoder_q-layer.7": 1435.9432, "encoder_q-layer.8": 1686.3013, "encoder_q-layer.9": 1611.4388, "epoch": 0.27, "inbatch_neg_score": 0.52, "inbatch_pos_score": 1.1143, "learning_rate": 3.621052631578948e-05, "loss": 3.5697, "norm_diff": 0.0989, "norm_loss": 0.0, "num_token_doc": 66.9006, "num_token_overlap": 11.6877, "num_token_query": 31.3745, "num_token_union": 65.1897, "num_word_context": 202.4338, "num_word_doc": 49.944, "num_word_query": 23.3151, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2384.9449, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5205, "query_norm": 1.4209, "queue_k_norm": 1.5204, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3745, "sent_len_1": 66.9006, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.4875, "stdk": 0.0487, "stdq": 0.0431, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62400 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.5611, "doc_norm": 1.5277, "encoder_q-embeddings": 1359.132, "encoder_q-layer.0": 890.2991, "encoder_q-layer.1": 938.4019, "encoder_q-layer.10": 1668.4415, "encoder_q-layer.11": 3801.5076, "encoder_q-layer.2": 1018.7057, "encoder_q-layer.3": 1041.7446, "encoder_q-layer.4": 1075.4102, "encoder_q-layer.5": 1110.4845, "encoder_q-layer.6": 1275.8256, "encoder_q-layer.7": 1430.0597, "encoder_q-layer.8": 1730.0027, "encoder_q-layer.9": 1516.3207, "epoch": 0.27, "inbatch_neg_score": 0.5105, "inbatch_pos_score": 1.0928, "learning_rate": 3.618421052631579e-05, "loss": 3.5611, "norm_diff": 0.1177, "norm_loss": 0.0, "num_token_doc": 67.043, "num_token_overlap": 11.6807, "num_token_query": 31.3929, "num_token_union": 65.3006, "num_word_context": 202.5595, "num_word_doc": 50.03, "num_word_query": 23.3195, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2411.4092, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5103, "query_norm": 1.41, "queue_k_norm": 1.5223, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3929, "sent_len_1": 67.043, "sent_len_max_0": 127.995, "sent_len_max_1": 210.65, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62500 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.5653, "doc_norm": 1.5228, "encoder_q-embeddings": 1370.7837, "encoder_q-layer.0": 934.1733, "encoder_q-layer.1": 994.2086, "encoder_q-layer.10": 1607.6395, "encoder_q-layer.11": 3707.6504, "encoder_q-layer.2": 1146.6094, "encoder_q-layer.3": 1169.1431, "encoder_q-layer.4": 1186.5582, "encoder_q-layer.5": 1247.7577, "encoder_q-layer.6": 1343.3199, "encoder_q-layer.7": 1498.0992, "encoder_q-layer.8": 1797.151, "encoder_q-layer.9": 1539.3011, "epoch": 0.27, "inbatch_neg_score": 0.5179, "inbatch_pos_score": 1.0977, "learning_rate": 3.6157894736842104e-05, "loss": 3.5653, "norm_diff": 0.0997, "norm_loss": 0.0, "num_token_doc": 66.7313, "num_token_overlap": 11.6714, "num_token_query": 31.3027, "num_token_union": 65.0417, "num_word_context": 202.052, "num_word_doc": 49.7845, "num_word_query": 23.2457, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2461.4166, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5176, "query_norm": 1.4232, "queue_k_norm": 1.5228, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3027, "sent_len_1": 66.7313, "sent_len_max_0": 127.9525, "sent_len_max_1": 207.1475, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62600 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.5673, "doc_norm": 1.5261, "encoder_q-embeddings": 1373.2526, "encoder_q-layer.0": 879.8579, "encoder_q-layer.1": 940.9661, "encoder_q-layer.10": 1548.3351, "encoder_q-layer.11": 3599.2046, "encoder_q-layer.2": 1059.6873, "encoder_q-layer.3": 1119.3668, "encoder_q-layer.4": 1188.3406, "encoder_q-layer.5": 1306.4165, "encoder_q-layer.6": 1387.0729, "encoder_q-layer.7": 1487.7511, "encoder_q-layer.8": 1754.2963, "encoder_q-layer.9": 1560.0247, "epoch": 0.27, "inbatch_neg_score": 0.5193, "inbatch_pos_score": 1.1182, "learning_rate": 3.613157894736842e-05, "loss": 3.5673, "norm_diff": 0.1042, "norm_loss": 0.0, "num_token_doc": 66.6971, "num_token_overlap": 11.638, "num_token_query": 31.2485, "num_token_union": 65.0206, "num_word_context": 202.1227, "num_word_doc": 49.7318, "num_word_query": 23.2168, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2412.2223, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.519, "query_norm": 1.4219, "queue_k_norm": 1.5229, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2485, "sent_len_1": 66.6971, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.0863, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62700 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5613, "doc_norm": 1.5226, "encoder_q-embeddings": 1306.5905, "encoder_q-layer.0": 840.8758, "encoder_q-layer.1": 891.6948, "encoder_q-layer.10": 1656.6006, "encoder_q-layer.11": 3851.6863, "encoder_q-layer.2": 1013.807, "encoder_q-layer.3": 1039.731, "encoder_q-layer.4": 1112.0682, "encoder_q-layer.5": 1138.1864, "encoder_q-layer.6": 1331.3756, "encoder_q-layer.7": 1570.0125, "encoder_q-layer.8": 1995.4261, "encoder_q-layer.9": 1674.035, "epoch": 0.27, "inbatch_neg_score": 0.5181, "inbatch_pos_score": 1.1064, "learning_rate": 3.6105263157894734e-05, "loss": 3.5613, "norm_diff": 0.0911, "norm_loss": 0.0, "num_token_doc": 66.8591, "num_token_overlap": 11.6647, "num_token_query": 31.31, "num_token_union": 65.142, "num_word_context": 202.7008, "num_word_doc": 49.9151, "num_word_query": 23.2634, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2519.049, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.519, "query_norm": 1.4314, "queue_k_norm": 1.5244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.31, "sent_len_1": 66.8591, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.74, "stdk": 0.0487, "stdq": 0.0435, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 62800 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.5808, "doc_norm": 1.5205, "encoder_q-embeddings": 1367.8909, "encoder_q-layer.0": 905.308, "encoder_q-layer.1": 1015.4255, "encoder_q-layer.10": 1588.0762, "encoder_q-layer.11": 3671.6484, "encoder_q-layer.2": 1133.9161, "encoder_q-layer.3": 1181.8481, "encoder_q-layer.4": 1257.1229, "encoder_q-layer.5": 1278.0428, "encoder_q-layer.6": 1412.877, "encoder_q-layer.7": 1525.4625, "encoder_q-layer.8": 1791.1788, "encoder_q-layer.9": 1548.5424, "epoch": 0.27, "inbatch_neg_score": 0.519, "inbatch_pos_score": 1.0898, "learning_rate": 3.607894736842106e-05, "loss": 3.5808, "norm_diff": 0.118, "norm_loss": 0.0, "num_token_doc": 66.8979, "num_token_overlap": 11.6813, "num_token_query": 31.3791, "num_token_union": 65.1862, "num_word_context": 202.3163, "num_word_doc": 49.9256, "num_word_query": 23.3137, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2455.2213, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5195, "query_norm": 1.4025, "queue_k_norm": 1.5259, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3791, "sent_len_1": 66.8979, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.8162, "stdk": 0.0486, "stdq": 0.0422, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 62900 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.5625, "doc_norm": 1.5265, "encoder_q-embeddings": 1422.7615, "encoder_q-layer.0": 947.1016, "encoder_q-layer.1": 1036.0959, "encoder_q-layer.10": 1761.9478, "encoder_q-layer.11": 3835.4988, "encoder_q-layer.2": 1149.7681, "encoder_q-layer.3": 1191.9132, "encoder_q-layer.4": 1292.7695, "encoder_q-layer.5": 1302.6606, "encoder_q-layer.6": 1517.4862, "encoder_q-layer.7": 1663.6938, "encoder_q-layer.8": 1906.1115, "encoder_q-layer.9": 1662.9005, "epoch": 0.27, "inbatch_neg_score": 0.5214, "inbatch_pos_score": 1.1113, "learning_rate": 3.605263157894737e-05, "loss": 3.5625, "norm_diff": 0.0921, "norm_loss": 0.0, "num_token_doc": 66.767, "num_token_overlap": 11.634, "num_token_query": 31.3284, "num_token_union": 65.1497, "num_word_context": 202.3476, "num_word_doc": 49.8048, "num_word_query": 23.2591, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2558.3416, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5215, "query_norm": 1.4344, "queue_k_norm": 1.5272, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3284, "sent_len_1": 66.767, "sent_len_max_0": 127.975, "sent_len_max_1": 209.3638, "stdk": 0.0488, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 63000 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.5659, "doc_norm": 1.5227, "encoder_q-embeddings": 1314.1791, "encoder_q-layer.0": 889.17, "encoder_q-layer.1": 946.058, "encoder_q-layer.10": 1762.3442, "encoder_q-layer.11": 3977.5435, "encoder_q-layer.2": 1047.1141, "encoder_q-layer.3": 1084.7694, "encoder_q-layer.4": 1181.2772, "encoder_q-layer.5": 1218.0511, "encoder_q-layer.6": 1400.1711, "encoder_q-layer.7": 1497.6251, "encoder_q-layer.8": 1757.8229, "encoder_q-layer.9": 1690.453, "epoch": 0.27, "inbatch_neg_score": 0.5243, "inbatch_pos_score": 1.1045, "learning_rate": 3.602631578947368e-05, "loss": 3.5659, "norm_diff": 0.1091, "norm_loss": 0.0, "num_token_doc": 66.8847, "num_token_overlap": 11.691, "num_token_query": 31.4031, "num_token_union": 65.2124, "num_word_context": 202.1996, "num_word_doc": 49.891, "num_word_query": 23.323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2555.5455, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5244, "query_norm": 1.4136, "queue_k_norm": 1.5274, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4031, "sent_len_1": 66.8847, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.4712, "stdk": 0.0486, "stdq": 0.0424, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 63100 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.553, "doc_norm": 1.5305, "encoder_q-embeddings": 1219.5786, "encoder_q-layer.0": 843.7901, "encoder_q-layer.1": 896.5472, "encoder_q-layer.10": 1559.8325, "encoder_q-layer.11": 3668.1543, "encoder_q-layer.2": 981.4562, "encoder_q-layer.3": 992.7021, "encoder_q-layer.4": 1050.7612, "encoder_q-layer.5": 1083.9568, "encoder_q-layer.6": 1213.2456, "encoder_q-layer.7": 1368.6279, "encoder_q-layer.8": 1612.7588, "encoder_q-layer.9": 1533.0836, "epoch": 0.27, "inbatch_neg_score": 0.5238, "inbatch_pos_score": 1.127, "learning_rate": 3.6e-05, "loss": 3.553, "norm_diff": 0.087, "norm_loss": 0.0, "num_token_doc": 66.8761, "num_token_overlap": 11.6595, "num_token_query": 31.3364, "num_token_union": 65.1793, "num_word_context": 202.2912, "num_word_doc": 49.9086, "num_word_query": 23.274, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2354.5851, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5239, "query_norm": 1.4435, "queue_k_norm": 1.529, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3364, "sent_len_1": 66.8761, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.4062, "stdk": 0.0489, "stdq": 0.0436, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 63200 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.5676, "doc_norm": 1.5262, "encoder_q-embeddings": 1328.8237, "encoder_q-layer.0": 871.4651, "encoder_q-layer.1": 923.6547, "encoder_q-layer.10": 1625.7727, "encoder_q-layer.11": 3727.49, "encoder_q-layer.2": 1036.0923, "encoder_q-layer.3": 1065.312, "encoder_q-layer.4": 1128.6166, "encoder_q-layer.5": 1137.1949, "encoder_q-layer.6": 1279.4274, "encoder_q-layer.7": 1484.6028, "encoder_q-layer.8": 1650.7006, "encoder_q-layer.9": 1563.5511, "epoch": 0.27, "inbatch_neg_score": 0.5313, "inbatch_pos_score": 1.1182, "learning_rate": 3.597368421052631e-05, "loss": 3.5676, "norm_diff": 0.084, "norm_loss": 0.0, "num_token_doc": 66.7935, "num_token_overlap": 11.6801, "num_token_query": 31.3644, "num_token_union": 65.1034, "num_word_context": 202.3092, "num_word_doc": 49.8282, "num_word_query": 23.2902, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2395.9152, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5312, "query_norm": 1.4423, "queue_k_norm": 1.5268, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3644, "sent_len_1": 66.7935, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.2075, "stdk": 0.0487, "stdq": 0.0431, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 63300 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5564, "doc_norm": 1.5269, "encoder_q-embeddings": 1362.0483, "encoder_q-layer.0": 908.0782, "encoder_q-layer.1": 945.8298, "encoder_q-layer.10": 1585.9014, "encoder_q-layer.11": 3580.0508, "encoder_q-layer.2": 1045.0874, "encoder_q-layer.3": 1110.4015, "encoder_q-layer.4": 1183.6672, "encoder_q-layer.5": 1167.7522, "encoder_q-layer.6": 1289.2659, "encoder_q-layer.7": 1572.6442, "encoder_q-layer.8": 1755.0989, "encoder_q-layer.9": 1515.3977, "epoch": 0.28, "inbatch_neg_score": 0.5364, "inbatch_pos_score": 1.1279, "learning_rate": 3.594736842105264e-05, "loss": 3.5564, "norm_diff": 0.091, "norm_loss": 0.0, "num_token_doc": 66.6983, "num_token_overlap": 11.6641, "num_token_query": 31.4847, "num_token_union": 65.146, "num_word_context": 202.47, "num_word_doc": 49.7786, "num_word_query": 23.3922, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2383.8607, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5366, "query_norm": 1.4359, "queue_k_norm": 1.5289, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4847, "sent_len_1": 66.6983, "sent_len_max_0": 127.995, "sent_len_max_1": 207.26, "stdk": 0.0487, "stdq": 0.0425, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 63400 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.548, "doc_norm": 1.5285, "encoder_q-embeddings": 1406.6567, "encoder_q-layer.0": 908.7577, "encoder_q-layer.1": 931.5321, "encoder_q-layer.10": 1652.7639, "encoder_q-layer.11": 3890.5586, "encoder_q-layer.2": 1045.7175, "encoder_q-layer.3": 1113.67, "encoder_q-layer.4": 1149.5151, "encoder_q-layer.5": 1186.4884, "encoder_q-layer.6": 1337.8632, "encoder_q-layer.7": 1494.5146, "encoder_q-layer.8": 1789.4645, "encoder_q-layer.9": 1638.2118, "epoch": 0.28, "inbatch_neg_score": 0.5396, "inbatch_pos_score": 1.1309, "learning_rate": 3.592105263157895e-05, "loss": 3.548, "norm_diff": 0.0624, "norm_loss": 0.0, "num_token_doc": 66.8628, "num_token_overlap": 11.6878, "num_token_query": 31.3943, "num_token_union": 65.1783, "num_word_context": 202.3141, "num_word_doc": 49.8817, "num_word_query": 23.3104, "postclip_grad_norm": 1.0, "preclip_grad_norm": 2512.3243, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.54, "query_norm": 1.4661, "queue_k_norm": 1.5289, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3943, "sent_len_1": 66.8628, "sent_len_max_0": 127.98, "sent_len_max_1": 208.4313, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 63500 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.5544, "doc_norm": 1.5272, "encoder_q-embeddings": 3383.7334, "encoder_q-layer.0": 2298.822, "encoder_q-layer.1": 2498.0261, "encoder_q-layer.10": 3346.793, "encoder_q-layer.11": 7473.8203, "encoder_q-layer.2": 2912.3464, "encoder_q-layer.3": 3103.2874, "encoder_q-layer.4": 3273.374, "encoder_q-layer.5": 3531.0813, "encoder_q-layer.6": 3405.6113, "encoder_q-layer.7": 3536.0911, "encoder_q-layer.8": 3745.1179, "encoder_q-layer.9": 3205.1218, "epoch": 0.28, "inbatch_neg_score": 0.5518, "inbatch_pos_score": 1.1475, "learning_rate": 3.589473684210527e-05, "loss": 3.5544, "norm_diff": 0.0437, "norm_loss": 0.0, "num_token_doc": 66.7167, "num_token_overlap": 11.703, "num_token_query": 31.4183, "num_token_union": 65.0957, "num_word_context": 202.2617, "num_word_doc": 49.7733, "num_word_query": 23.3321, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5515.5052, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.5518, "query_norm": 1.4835, "queue_k_norm": 1.5316, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4183, "sent_len_1": 66.7167, "sent_len_max_0": 127.99, "sent_len_max_1": 208.8837, "stdk": 0.0487, "stdq": 0.0438, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 63600 }, { "accuracy": 46.2891, "active_queue_size": 16384.0, "cl_loss": 3.5451, "doc_norm": 1.5269, "encoder_q-embeddings": 2859.002, "encoder_q-layer.0": 1855.8026, "encoder_q-layer.1": 1996.6282, "encoder_q-layer.10": 3834.6846, "encoder_q-layer.11": 7802.7964, "encoder_q-layer.2": 2300.2949, "encoder_q-layer.3": 2398.4365, "encoder_q-layer.4": 2567.8064, "encoder_q-layer.5": 2595.0654, "encoder_q-layer.6": 2835.0083, "encoder_q-layer.7": 3224.2283, "encoder_q-layer.8": 4102.0771, "encoder_q-layer.9": 3624.2761, "epoch": 0.28, "inbatch_neg_score": 0.5646, "inbatch_pos_score": 1.1377, "learning_rate": 3.586842105263158e-05, "loss": 3.5451, "norm_diff": 0.0484, "norm_loss": 0.0, "num_token_doc": 66.7337, "num_token_overlap": 11.7023, "num_token_query": 31.4556, "num_token_union": 65.1365, "num_word_context": 202.2697, "num_word_doc": 49.8106, "num_word_query": 23.3805, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5196.7956, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5635, "query_norm": 1.4785, "queue_k_norm": 1.5332, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4556, "sent_len_1": 66.7337, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.3275, "stdk": 0.0486, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 63700 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.558, "doc_norm": 1.5291, "encoder_q-embeddings": 3031.7622, "encoder_q-layer.0": 2107.4736, "encoder_q-layer.1": 2341.3374, "encoder_q-layer.10": 3685.3535, "encoder_q-layer.11": 7709.9624, "encoder_q-layer.2": 2687.4436, "encoder_q-layer.3": 2735.4888, "encoder_q-layer.4": 2796.7297, "encoder_q-layer.5": 2640.1577, "encoder_q-layer.6": 2872.6653, "encoder_q-layer.7": 3030.8569, "encoder_q-layer.8": 3898.5581, "encoder_q-layer.9": 3456.1096, "epoch": 0.28, "inbatch_neg_score": 0.5734, "inbatch_pos_score": 1.1768, "learning_rate": 3.58421052631579e-05, "loss": 3.558, "norm_diff": 0.0377, "norm_loss": 0.0, "num_token_doc": 66.8259, "num_token_overlap": 11.6961, "num_token_query": 31.4323, "num_token_union": 65.1774, "num_word_context": 202.4244, "num_word_doc": 49.8824, "num_word_query": 23.3498, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5214.3977, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5728, "query_norm": 1.4914, "queue_k_norm": 1.5332, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4323, "sent_len_1": 66.8259, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.0687, "stdk": 0.0486, "stdq": 0.0439, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 63800 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.5669, "doc_norm": 1.5315, "encoder_q-embeddings": 2733.499, "encoder_q-layer.0": 1813.7882, "encoder_q-layer.1": 1921.6217, "encoder_q-layer.10": 3251.5076, "encoder_q-layer.11": 7606.2329, "encoder_q-layer.2": 2122.8511, "encoder_q-layer.3": 2193.2866, "encoder_q-layer.4": 2347.3125, "encoder_q-layer.5": 2329.0791, "encoder_q-layer.6": 2614.752, "encoder_q-layer.7": 2806.6399, "encoder_q-layer.8": 3466.7671, "encoder_q-layer.9": 3107.334, "epoch": 0.28, "inbatch_neg_score": 0.581, "inbatch_pos_score": 1.1514, "learning_rate": 3.5815789473684216e-05, "loss": 3.5669, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.6968, "num_token_overlap": 11.6759, "num_token_query": 31.3872, "num_token_union": 65.0655, "num_word_context": 202.4391, "num_word_doc": 49.7588, "num_word_query": 23.2922, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4828.1742, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5801, "query_norm": 1.4576, "queue_k_norm": 1.5334, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3872, "sent_len_1": 66.6968, "sent_len_max_0": 127.97, "sent_len_max_1": 209.95, "stdk": 0.0487, "stdq": 0.0428, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 63900 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.5616, "doc_norm": 1.5369, "encoder_q-embeddings": 2755.4653, "encoder_q-layer.0": 1931.8081, "encoder_q-layer.1": 2082.1985, "encoder_q-layer.10": 2958.8945, "encoder_q-layer.11": 7446.1333, "encoder_q-layer.2": 2371.166, "encoder_q-layer.3": 2348.439, "encoder_q-layer.4": 2558.0303, "encoder_q-layer.5": 2655.991, "encoder_q-layer.6": 2788.1594, "encoder_q-layer.7": 3045.998, "encoder_q-layer.8": 3274.5332, "encoder_q-layer.9": 2964.8188, "epoch": 0.28, "inbatch_neg_score": 0.5793, "inbatch_pos_score": 1.1729, "learning_rate": 3.578947368421053e-05, "loss": 3.5616, "norm_diff": 0.0937, "norm_loss": 0.0, "num_token_doc": 66.894, "num_token_overlap": 11.707, "num_token_query": 31.4553, "num_token_union": 65.1965, "num_word_context": 202.6138, "num_word_doc": 49.9364, "num_word_query": 23.3573, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4865.9597, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5791, "query_norm": 1.4432, "queue_k_norm": 1.5368, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4553, "sent_len_1": 66.894, "sent_len_max_0": 127.995, "sent_len_max_1": 207.3675, "stdk": 0.0488, "stdq": 0.0427, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 64000 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.5444, "doc_norm": 1.5393, "encoder_q-embeddings": 2500.1653, "encoder_q-layer.0": 1711.504, "encoder_q-layer.1": 1789.4259, "encoder_q-layer.10": 3331.6382, "encoder_q-layer.11": 7630.3755, "encoder_q-layer.2": 1979.999, "encoder_q-layer.3": 2081.3376, "encoder_q-layer.4": 2224.2634, "encoder_q-layer.5": 2321.324, "encoder_q-layer.6": 2617.1184, "encoder_q-layer.7": 2840.2795, "encoder_q-layer.8": 3511.0486, "encoder_q-layer.9": 3195.459, "epoch": 0.28, "inbatch_neg_score": 0.5841, "inbatch_pos_score": 1.1689, "learning_rate": 3.5763157894736846e-05, "loss": 3.5444, "norm_diff": 0.1072, "norm_loss": 0.0, "num_token_doc": 66.7643, "num_token_overlap": 11.6752, "num_token_query": 31.4427, "num_token_union": 65.1147, "num_word_context": 202.2564, "num_word_doc": 49.7814, "num_word_query": 23.3755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4757.9223, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5835, "query_norm": 1.4321, "queue_k_norm": 1.5367, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4427, "sent_len_1": 66.7643, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.92, "stdk": 0.0488, "stdq": 0.0425, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 64100 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.564, "doc_norm": 1.5356, "encoder_q-embeddings": 2547.7361, "encoder_q-layer.0": 1716.0261, "encoder_q-layer.1": 1836.1576, "encoder_q-layer.10": 3116.0105, "encoder_q-layer.11": 7673.0552, "encoder_q-layer.2": 2038.9449, "encoder_q-layer.3": 2094.8123, "encoder_q-layer.4": 2150.7627, "encoder_q-layer.5": 2201.7344, "encoder_q-layer.6": 2387.1155, "encoder_q-layer.7": 2890.4873, "encoder_q-layer.8": 3424.5356, "encoder_q-layer.9": 3183.7839, "epoch": 0.28, "inbatch_neg_score": 0.5858, "inbatch_pos_score": 1.1787, "learning_rate": 3.573684210526316e-05, "loss": 3.564, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.7394, "num_token_overlap": 11.6144, "num_token_query": 31.251, "num_token_union": 65.0468, "num_word_context": 202.2646, "num_word_doc": 49.7731, "num_word_query": 23.1873, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4706.0164, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5854, "query_norm": 1.4562, "queue_k_norm": 1.5388, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.251, "sent_len_1": 66.7394, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.2663, "stdk": 0.0486, "stdq": 0.0437, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 64200 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5714, "doc_norm": 1.5434, "encoder_q-embeddings": 2636.1279, "encoder_q-layer.0": 1758.418, "encoder_q-layer.1": 1842.186, "encoder_q-layer.10": 3999.7817, "encoder_q-layer.11": 8261.1719, "encoder_q-layer.2": 2023.1211, "encoder_q-layer.3": 2138.7893, "encoder_q-layer.4": 2315.1506, "encoder_q-layer.5": 2414.2844, "encoder_q-layer.6": 2729.6509, "encoder_q-layer.7": 3259.9053, "encoder_q-layer.8": 4134.8618, "encoder_q-layer.9": 3743.1846, "epoch": 0.28, "inbatch_neg_score": 0.5831, "inbatch_pos_score": 1.1709, "learning_rate": 3.5710526315789476e-05, "loss": 3.5714, "norm_diff": 0.0868, "norm_loss": 0.0, "num_token_doc": 66.7265, "num_token_overlap": 11.6694, "num_token_query": 31.3251, "num_token_union": 65.0496, "num_word_context": 202.2582, "num_word_doc": 49.8149, "num_word_query": 23.2765, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5200.5458, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.582, "query_norm": 1.4566, "queue_k_norm": 1.5394, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3251, "sent_len_1": 66.7265, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.44, "stdk": 0.0489, "stdq": 0.0439, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 64300 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.5698, "doc_norm": 1.5394, "encoder_q-embeddings": 2615.6455, "encoder_q-layer.0": 1767.4554, "encoder_q-layer.1": 1843.5802, "encoder_q-layer.10": 3242.7708, "encoder_q-layer.11": 7479.1143, "encoder_q-layer.2": 2039.4108, "encoder_q-layer.3": 2121.5569, "encoder_q-layer.4": 2272.8037, "encoder_q-layer.5": 2270.3904, "encoder_q-layer.6": 2552.0159, "encoder_q-layer.7": 3025.6865, "encoder_q-layer.8": 3578.5225, "encoder_q-layer.9": 3190.1753, "epoch": 0.28, "inbatch_neg_score": 0.5776, "inbatch_pos_score": 1.165, "learning_rate": 3.5684210526315794e-05, "loss": 3.5698, "norm_diff": 0.1093, "norm_loss": 0.0, "num_token_doc": 66.7712, "num_token_overlap": 11.6649, "num_token_query": 31.3441, "num_token_union": 65.1003, "num_word_context": 202.3791, "num_word_doc": 49.8464, "num_word_query": 23.2833, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4837.5482, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5771, "query_norm": 1.4301, "queue_k_norm": 1.5423, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3441, "sent_len_1": 66.7712, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.1687, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 64400 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.5637, "doc_norm": 1.5399, "encoder_q-embeddings": 2626.8618, "encoder_q-layer.0": 1836.2122, "encoder_q-layer.1": 2014.5912, "encoder_q-layer.10": 3295.8823, "encoder_q-layer.11": 7776.5142, "encoder_q-layer.2": 2194.4414, "encoder_q-layer.3": 2281.2012, "encoder_q-layer.4": 2304.658, "encoder_q-layer.5": 2414.2271, "encoder_q-layer.6": 2706.6785, "encoder_q-layer.7": 3050.1748, "encoder_q-layer.8": 3698.5508, "encoder_q-layer.9": 3209.1025, "epoch": 0.28, "inbatch_neg_score": 0.5758, "inbatch_pos_score": 1.1465, "learning_rate": 3.5657894736842106e-05, "loss": 3.5637, "norm_diff": 0.1135, "norm_loss": 0.0, "num_token_doc": 66.9532, "num_token_overlap": 11.6433, "num_token_query": 31.2448, "num_token_union": 65.164, "num_word_context": 202.6386, "num_word_doc": 49.9624, "num_word_query": 23.1966, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4927.0762, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5752, "query_norm": 1.4265, "queue_k_norm": 1.5427, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2448, "sent_len_1": 66.9532, "sent_len_max_0": 127.985, "sent_len_max_1": 209.5563, "stdk": 0.0487, "stdq": 0.0426, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 64500 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.5619, "doc_norm": 1.5417, "encoder_q-embeddings": 2594.9441, "encoder_q-layer.0": 1713.9655, "encoder_q-layer.1": 1776.2336, "encoder_q-layer.10": 3296.2786, "encoder_q-layer.11": 7573.4502, "encoder_q-layer.2": 1950.3958, "encoder_q-layer.3": 2018.4327, "encoder_q-layer.4": 2143.1531, "encoder_q-layer.5": 2210.5586, "encoder_q-layer.6": 2530.2261, "encoder_q-layer.7": 2915.551, "encoder_q-layer.8": 3497.7161, "encoder_q-layer.9": 3282.1206, "epoch": 0.28, "inbatch_neg_score": 0.5742, "inbatch_pos_score": 1.1592, "learning_rate": 3.5631578947368424e-05, "loss": 3.5619, "norm_diff": 0.1195, "norm_loss": 0.0, "num_token_doc": 66.9386, "num_token_overlap": 11.6821, "num_token_query": 31.3132, "num_token_union": 65.1578, "num_word_context": 202.541, "num_word_doc": 49.9181, "num_word_query": 23.2648, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4854.646, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5752, "query_norm": 1.4221, "queue_k_norm": 1.5426, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3132, "sent_len_1": 66.9386, "sent_len_max_0": 127.98, "sent_len_max_1": 210.7788, "stdk": 0.0487, "stdq": 0.0425, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 64600 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.5526, "doc_norm": 1.5403, "encoder_q-embeddings": 2650.8762, "encoder_q-layer.0": 1792.4552, "encoder_q-layer.1": 1897.2623, "encoder_q-layer.10": 3185.8743, "encoder_q-layer.11": 7751.8628, "encoder_q-layer.2": 2156.1187, "encoder_q-layer.3": 2200.3057, "encoder_q-layer.4": 2500.5745, "encoder_q-layer.5": 2463.491, "encoder_q-layer.6": 2716.813, "encoder_q-layer.7": 3190.3604, "encoder_q-layer.8": 3389.8586, "encoder_q-layer.9": 3045.8926, "epoch": 0.28, "inbatch_neg_score": 0.5771, "inbatch_pos_score": 1.1504, "learning_rate": 3.5605263157894736e-05, "loss": 3.5526, "norm_diff": 0.1157, "norm_loss": 0.0, "num_token_doc": 66.9109, "num_token_overlap": 11.6718, "num_token_query": 31.334, "num_token_union": 65.1794, "num_word_context": 202.2085, "num_word_doc": 49.8979, "num_word_query": 23.274, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4934.3565, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5771, "query_norm": 1.4246, "queue_k_norm": 1.5454, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.334, "sent_len_1": 66.9109, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.62, "stdk": 0.0486, "stdq": 0.0425, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 64700 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.5603, "doc_norm": 1.5408, "encoder_q-embeddings": 2879.5029, "encoder_q-layer.0": 2052.8396, "encoder_q-layer.1": 2177.3398, "encoder_q-layer.10": 3191.6062, "encoder_q-layer.11": 7566.6514, "encoder_q-layer.2": 2494.5017, "encoder_q-layer.3": 2488.7778, "encoder_q-layer.4": 2771.1731, "encoder_q-layer.5": 2624.1021, "encoder_q-layer.6": 2929.6819, "encoder_q-layer.7": 3030.7759, "encoder_q-layer.8": 3352.8596, "encoder_q-layer.9": 2994.8711, "epoch": 0.28, "inbatch_neg_score": 0.5815, "inbatch_pos_score": 1.1553, "learning_rate": 3.5578947368421054e-05, "loss": 3.5603, "norm_diff": 0.1134, "norm_loss": 0.0, "num_token_doc": 66.6553, "num_token_overlap": 11.683, "num_token_query": 31.3818, "num_token_union": 65.0721, "num_word_context": 202.0951, "num_word_doc": 49.7555, "num_word_query": 23.2983, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5056.8242, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5815, "query_norm": 1.4274, "queue_k_norm": 1.5437, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3818, "sent_len_1": 66.6553, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.9825, "stdk": 0.0486, "stdq": 0.0425, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 64800 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.5549, "doc_norm": 1.5494, "encoder_q-embeddings": 2571.9763, "encoder_q-layer.0": 1756.2141, "encoder_q-layer.1": 1858.1389, "encoder_q-layer.10": 3480.0454, "encoder_q-layer.11": 7534.7827, "encoder_q-layer.2": 2068.9775, "encoder_q-layer.3": 2167.6956, "encoder_q-layer.4": 2298.3962, "encoder_q-layer.5": 2243.3499, "encoder_q-layer.6": 2503.5889, "encoder_q-layer.7": 2847.3486, "encoder_q-layer.8": 3426.2246, "encoder_q-layer.9": 3203.261, "epoch": 0.28, "inbatch_neg_score": 0.5824, "inbatch_pos_score": 1.1758, "learning_rate": 3.555263157894737e-05, "loss": 3.5549, "norm_diff": 0.0999, "norm_loss": 0.0, "num_token_doc": 66.7166, "num_token_overlap": 11.6751, "num_token_query": 31.3495, "num_token_union": 65.0681, "num_word_context": 202.3485, "num_word_doc": 49.7438, "num_word_query": 23.2924, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4782.1801, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5825, "query_norm": 1.4494, "queue_k_norm": 1.5456, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3495, "sent_len_1": 66.7166, "sent_len_max_0": 127.9663, "sent_len_max_1": 211.68, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 64900 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.5516, "doc_norm": 1.5506, "encoder_q-embeddings": 2667.1406, "encoder_q-layer.0": 1753.0917, "encoder_q-layer.1": 1841.8439, "encoder_q-layer.10": 3189.8726, "encoder_q-layer.11": 7283.2896, "encoder_q-layer.2": 2097.1467, "encoder_q-layer.3": 2159.2097, "encoder_q-layer.4": 2294.5366, "encoder_q-layer.5": 2319.3113, "encoder_q-layer.6": 2580.7373, "encoder_q-layer.7": 2905.6255, "encoder_q-layer.8": 3495.1958, "encoder_q-layer.9": 3123.6948, "epoch": 0.28, "inbatch_neg_score": 0.582, "inbatch_pos_score": 1.1758, "learning_rate": 3.5526315789473684e-05, "loss": 3.5516, "norm_diff": 0.1094, "norm_loss": 0.0, "num_token_doc": 66.7042, "num_token_overlap": 11.6126, "num_token_query": 31.2687, "num_token_union": 65.0651, "num_word_context": 202.1823, "num_word_doc": 49.7781, "num_word_query": 23.222, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4782.8039, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.583, "query_norm": 1.4412, "queue_k_norm": 1.5467, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2687, "sent_len_1": 66.7042, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.4775, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65000 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.5553, "doc_norm": 1.5517, "encoder_q-embeddings": 2694.697, "encoder_q-layer.0": 1789.824, "encoder_q-layer.1": 1892.6711, "encoder_q-layer.10": 3268.0923, "encoder_q-layer.11": 7376.3296, "encoder_q-layer.2": 2189.1589, "encoder_q-layer.3": 2207.4558, "encoder_q-layer.4": 2253.2488, "encoder_q-layer.5": 2286.3381, "encoder_q-layer.6": 2538.2698, "encoder_q-layer.7": 2982.1238, "encoder_q-layer.8": 3507.7803, "encoder_q-layer.9": 3133.2112, "epoch": 0.28, "inbatch_neg_score": 0.5859, "inbatch_pos_score": 1.1807, "learning_rate": 3.55e-05, "loss": 3.5553, "norm_diff": 0.0984, "norm_loss": 0.0, "num_token_doc": 66.8741, "num_token_overlap": 11.6779, "num_token_query": 31.33, "num_token_union": 65.1606, "num_word_context": 202.2803, "num_word_doc": 49.9013, "num_word_query": 23.2703, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4890.5491, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5854, "query_norm": 1.4534, "queue_k_norm": 1.5478, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.33, "sent_len_1": 66.8741, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.27, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65100 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.5408, "doc_norm": 1.5463, "encoder_q-embeddings": 3226.8167, "encoder_q-layer.0": 2254.718, "encoder_q-layer.1": 2408.981, "encoder_q-layer.10": 3394.8184, "encoder_q-layer.11": 7667.0664, "encoder_q-layer.2": 2625.9338, "encoder_q-layer.3": 2585.4768, "encoder_q-layer.4": 2650.8887, "encoder_q-layer.5": 2744.0464, "encoder_q-layer.6": 2815.8123, "encoder_q-layer.7": 2941.3447, "encoder_q-layer.8": 3409.9575, "encoder_q-layer.9": 3163.6262, "epoch": 0.28, "inbatch_neg_score": 0.5939, "inbatch_pos_score": 1.1787, "learning_rate": 3.547368421052632e-05, "loss": 3.5408, "norm_diff": 0.1013, "norm_loss": 0.0, "num_token_doc": 66.7802, "num_token_overlap": 11.6821, "num_token_query": 31.3344, "num_token_union": 65.0845, "num_word_context": 202.3634, "num_word_doc": 49.8427, "num_word_query": 23.2734, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5193.4398, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5928, "query_norm": 1.4451, "queue_k_norm": 1.5474, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3344, "sent_len_1": 66.7802, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.895, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 65200 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.5443, "doc_norm": 1.5492, "encoder_q-embeddings": 2757.5227, "encoder_q-layer.0": 1851.8365, "encoder_q-layer.1": 1999.5222, "encoder_q-layer.10": 3060.1389, "encoder_q-layer.11": 7052.3057, "encoder_q-layer.2": 2197.4089, "encoder_q-layer.3": 2253.0117, "encoder_q-layer.4": 2360.0503, "encoder_q-layer.5": 2456.958, "encoder_q-layer.6": 2695.2864, "encoder_q-layer.7": 2876.6104, "encoder_q-layer.8": 3482.7432, "encoder_q-layer.9": 3016.2771, "epoch": 0.28, "inbatch_neg_score": 0.5923, "inbatch_pos_score": 1.1953, "learning_rate": 3.544736842105263e-05, "loss": 3.5443, "norm_diff": 0.093, "norm_loss": 0.0, "num_token_doc": 66.8599, "num_token_overlap": 11.7271, "num_token_query": 31.491, "num_token_union": 65.1805, "num_word_context": 202.3189, "num_word_doc": 49.8757, "num_word_query": 23.3912, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4802.3861, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5913, "query_norm": 1.4562, "queue_k_norm": 1.5487, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.491, "sent_len_1": 66.8599, "sent_len_max_0": 127.985, "sent_len_max_1": 210.015, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65300 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.555, "doc_norm": 1.5479, "encoder_q-embeddings": 2640.2744, "encoder_q-layer.0": 1790.0743, "encoder_q-layer.1": 1964.266, "encoder_q-layer.10": 3313.9448, "encoder_q-layer.11": 7532.8145, "encoder_q-layer.2": 2128.3542, "encoder_q-layer.3": 2205.6995, "encoder_q-layer.4": 2323.4116, "encoder_q-layer.5": 2460.1401, "encoder_q-layer.6": 2647.3376, "encoder_q-layer.7": 2863.0308, "encoder_q-layer.8": 3403.8318, "encoder_q-layer.9": 3118.5745, "epoch": 0.28, "inbatch_neg_score": 0.5945, "inbatch_pos_score": 1.1738, "learning_rate": 3.542105263157895e-05, "loss": 3.555, "norm_diff": 0.0915, "norm_loss": 0.0, "num_token_doc": 66.8025, "num_token_overlap": 11.6627, "num_token_query": 31.3328, "num_token_union": 65.1205, "num_word_context": 202.1574, "num_word_doc": 49.8764, "num_word_query": 23.2763, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4906.1894, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5952, "query_norm": 1.4564, "queue_k_norm": 1.5498, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3328, "sent_len_1": 66.8025, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.835, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65400 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.5588, "doc_norm": 1.5415, "encoder_q-embeddings": 3168.4795, "encoder_q-layer.0": 2190.4727, "encoder_q-layer.1": 2404.2209, "encoder_q-layer.10": 3356.5303, "encoder_q-layer.11": 7909.2192, "encoder_q-layer.2": 2709.697, "encoder_q-layer.3": 2792.3074, "encoder_q-layer.4": 2886.9597, "encoder_q-layer.5": 2989.0098, "encoder_q-layer.6": 3200.8965, "encoder_q-layer.7": 3272.4739, "encoder_q-layer.8": 3551.7217, "encoder_q-layer.9": 3272.1206, "epoch": 0.28, "inbatch_neg_score": 0.5954, "inbatch_pos_score": 1.1846, "learning_rate": 3.539473684210526e-05, "loss": 3.5588, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.7783, "num_token_overlap": 11.6759, "num_token_query": 31.3305, "num_token_union": 65.0854, "num_word_context": 202.3462, "num_word_doc": 49.8362, "num_word_query": 23.2733, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5397.9541, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.5957, "query_norm": 1.4622, "queue_k_norm": 1.551, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3305, "sent_len_1": 66.7783, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.8063, "stdk": 0.0485, "stdq": 0.0433, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65500 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.557, "doc_norm": 1.5519, "encoder_q-embeddings": 5466.2197, "encoder_q-layer.0": 3576.1792, "encoder_q-layer.1": 3811.1135, "encoder_q-layer.10": 7193.0884, "encoder_q-layer.11": 16260.5537, "encoder_q-layer.2": 4343.582, "encoder_q-layer.3": 4543.8779, "encoder_q-layer.4": 4867.1694, "encoder_q-layer.5": 5058.644, "encoder_q-layer.6": 5630.2373, "encoder_q-layer.7": 6276.1782, "encoder_q-layer.8": 7965.7998, "encoder_q-layer.9": 7094.7383, "epoch": 0.28, "inbatch_neg_score": 0.6017, "inbatch_pos_score": 1.1846, "learning_rate": 3.536842105263158e-05, "loss": 3.557, "norm_diff": 0.101, "norm_loss": 0.0, "num_token_doc": 66.7319, "num_token_overlap": 11.6815, "num_token_query": 31.3544, "num_token_union": 65.1122, "num_word_context": 202.1545, "num_word_doc": 49.7777, "num_word_query": 23.2983, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10389.4762, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6006, "query_norm": 1.451, "queue_k_norm": 1.5511, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3544, "sent_len_1": 66.7319, "sent_len_max_0": 127.98, "sent_len_max_1": 210.6825, "stdk": 0.0489, "stdq": 0.0426, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65600 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.5619, "doc_norm": 1.5523, "encoder_q-embeddings": 4992.9438, "encoder_q-layer.0": 3437.3066, "encoder_q-layer.1": 3630.0681, "encoder_q-layer.10": 6579.2319, "encoder_q-layer.11": 14878.7012, "encoder_q-layer.2": 3938.667, "encoder_q-layer.3": 4152.1392, "encoder_q-layer.4": 4474.6968, "encoder_q-layer.5": 4386.3047, "encoder_q-layer.6": 4877.0591, "encoder_q-layer.7": 5535.9067, "encoder_q-layer.8": 6578.5835, "encoder_q-layer.9": 6069.7461, "epoch": 0.29, "inbatch_neg_score": 0.6092, "inbatch_pos_score": 1.1895, "learning_rate": 3.53421052631579e-05, "loss": 3.5619, "norm_diff": 0.0947, "norm_loss": 0.0, "num_token_doc": 66.8495, "num_token_overlap": 11.6945, "num_token_query": 31.3771, "num_token_union": 65.1578, "num_word_context": 202.5577, "num_word_doc": 49.9083, "num_word_query": 23.3106, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9388.1595, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6099, "query_norm": 1.4576, "queue_k_norm": 1.5512, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3771, "sent_len_1": 66.8495, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.2413, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65700 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.5632, "doc_norm": 1.5482, "encoder_q-embeddings": 5221.2983, "encoder_q-layer.0": 3421.948, "encoder_q-layer.1": 3532.5066, "encoder_q-layer.10": 6327.998, "encoder_q-layer.11": 14429.9141, "encoder_q-layer.2": 3971.5457, "encoder_q-layer.3": 4178.9883, "encoder_q-layer.4": 4629.5996, "encoder_q-layer.5": 4544.9048, "encoder_q-layer.6": 5064.9844, "encoder_q-layer.7": 5617.6753, "encoder_q-layer.8": 6652.4688, "encoder_q-layer.9": 6259.5508, "epoch": 0.29, "inbatch_neg_score": 0.6143, "inbatch_pos_score": 1.1934, "learning_rate": 3.531578947368421e-05, "loss": 3.5632, "norm_diff": 0.0676, "norm_loss": 0.0, "num_token_doc": 66.9698, "num_token_overlap": 11.7229, "num_token_query": 31.4951, "num_token_union": 65.2746, "num_word_context": 202.6298, "num_word_doc": 49.9983, "num_word_query": 23.3958, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9294.988, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6143, "query_norm": 1.4807, "queue_k_norm": 1.5524, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4951, "sent_len_1": 66.9698, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.0863, "stdk": 0.0486, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 65800 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.5622, "doc_norm": 1.5522, "encoder_q-embeddings": 5418.5415, "encoder_q-layer.0": 3755.6892, "encoder_q-layer.1": 3896.8108, "encoder_q-layer.10": 6809.4761, "encoder_q-layer.11": 15742.7109, "encoder_q-layer.2": 4443.7856, "encoder_q-layer.3": 4461.1528, "encoder_q-layer.4": 4924.7432, "encoder_q-layer.5": 4886.6304, "encoder_q-layer.6": 5370.7358, "encoder_q-layer.7": 5879.9653, "encoder_q-layer.8": 6641.7046, "encoder_q-layer.9": 6242.458, "epoch": 0.29, "inbatch_neg_score": 0.627, "inbatch_pos_score": 1.21, "learning_rate": 3.528947368421053e-05, "loss": 3.5622, "norm_diff": 0.0637, "norm_loss": 0.0, "num_token_doc": 66.6933, "num_token_overlap": 11.6216, "num_token_query": 31.184, "num_token_union": 64.9803, "num_word_context": 202.1253, "num_word_doc": 49.7455, "num_word_query": 23.1439, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9864.9191, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.626, "query_norm": 1.4885, "queue_k_norm": 1.5526, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.184, "sent_len_1": 66.6933, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.5137, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 65900 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.5339, "doc_norm": 1.5529, "encoder_q-embeddings": 5204.2168, "encoder_q-layer.0": 3643.6482, "encoder_q-layer.1": 3789.0066, "encoder_q-layer.10": 6520.8569, "encoder_q-layer.11": 14373.9297, "encoder_q-layer.2": 4234.7705, "encoder_q-layer.3": 4407.7993, "encoder_q-layer.4": 4580.5068, "encoder_q-layer.5": 4690.8018, "encoder_q-layer.6": 5122.2764, "encoder_q-layer.7": 5651.0283, "encoder_q-layer.8": 6457.0957, "encoder_q-layer.9": 5832.1479, "epoch": 0.29, "inbatch_neg_score": 0.6394, "inbatch_pos_score": 1.2432, "learning_rate": 3.526315789473684e-05, "loss": 3.5339, "norm_diff": 0.0518, "norm_loss": 0.0, "num_token_doc": 66.7189, "num_token_overlap": 11.7105, "num_token_query": 31.4082, "num_token_union": 65.0841, "num_word_context": 202.2497, "num_word_doc": 49.7806, "num_word_query": 23.3229, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9400.8794, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6382, "query_norm": 1.5011, "queue_k_norm": 1.5556, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4082, "sent_len_1": 66.7189, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.0425, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66000 }, { "accuracy": 46.5332, "active_queue_size": 16384.0, "cl_loss": 3.5398, "doc_norm": 1.5584, "encoder_q-embeddings": 15662.4385, "encoder_q-layer.0": 10417.1094, "encoder_q-layer.1": 10245.3789, "encoder_q-layer.10": 8835.5088, "encoder_q-layer.11": 16921.6895, "encoder_q-layer.2": 11486.9717, "encoder_q-layer.3": 12320.9482, "encoder_q-layer.4": 13157.3193, "encoder_q-layer.5": 16517.5137, "encoder_q-layer.6": 16040.9795, "encoder_q-layer.7": 14818.5156, "encoder_q-layer.8": 12170.8223, "encoder_q-layer.9": 8822.5625, "epoch": 0.29, "inbatch_neg_score": 0.651, "inbatch_pos_score": 1.2373, "learning_rate": 3.523684210526316e-05, "loss": 3.5398, "norm_diff": 0.0363, "norm_loss": 0.0, "num_token_doc": 66.7951, "num_token_overlap": 11.6888, "num_token_query": 31.4456, "num_token_union": 65.1506, "num_word_context": 202.3542, "num_word_doc": 49.8495, "num_word_query": 23.3565, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19789.9668, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.6499, "query_norm": 1.5221, "queue_k_norm": 1.5555, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4456, "sent_len_1": 66.7951, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.1675, "stdk": 0.0489, "stdq": 0.044, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 66100 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.5307, "doc_norm": 1.5585, "encoder_q-embeddings": 6983.6499, "encoder_q-layer.0": 4872.2603, "encoder_q-layer.1": 5576.7969, "encoder_q-layer.10": 7038.2134, "encoder_q-layer.11": 15260.0107, "encoder_q-layer.2": 6205.8188, "encoder_q-layer.3": 5870.7744, "encoder_q-layer.4": 6353.71, "encoder_q-layer.5": 6278.7837, "encoder_q-layer.6": 6719.1626, "encoder_q-layer.7": 8088.8579, "encoder_q-layer.8": 8102.1021, "encoder_q-layer.9": 6688.0625, "epoch": 0.29, "inbatch_neg_score": 0.6637, "inbatch_pos_score": 1.2607, "learning_rate": 3.521052631578948e-05, "loss": 3.5307, "norm_diff": 0.0377, "norm_loss": 0.0, "num_token_doc": 66.8524, "num_token_overlap": 11.6846, "num_token_query": 31.3949, "num_token_union": 65.1612, "num_word_context": 202.3557, "num_word_doc": 49.8772, "num_word_query": 23.3178, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11428.9857, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6636, "query_norm": 1.5208, "queue_k_norm": 1.5581, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3949, "sent_len_1": 66.8524, "sent_len_max_0": 127.9463, "sent_len_max_1": 207.905, "stdk": 0.0488, "stdq": 0.0437, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66200 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.5528, "doc_norm": 1.5622, "encoder_q-embeddings": 5328.9238, "encoder_q-layer.0": 3559.2754, "encoder_q-layer.1": 3752.7666, "encoder_q-layer.10": 6501.3486, "encoder_q-layer.11": 15084.2432, "encoder_q-layer.2": 4046.8838, "encoder_q-layer.3": 4304.9819, "encoder_q-layer.4": 4549.9829, "encoder_q-layer.5": 4683.0469, "encoder_q-layer.6": 5219.2969, "encoder_q-layer.7": 5857.0708, "encoder_q-layer.8": 7077.4038, "encoder_q-layer.9": 6331.9111, "epoch": 0.29, "inbatch_neg_score": 0.6704, "inbatch_pos_score": 1.2686, "learning_rate": 3.518421052631579e-05, "loss": 3.5528, "norm_diff": 0.0544, "norm_loss": 0.0, "num_token_doc": 66.7802, "num_token_overlap": 11.6676, "num_token_query": 31.3736, "num_token_union": 65.1214, "num_word_context": 202.2614, "num_word_doc": 49.7974, "num_word_query": 23.3021, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9802.756, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6709, "query_norm": 1.5077, "queue_k_norm": 1.5607, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3736, "sent_len_1": 66.7802, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.1962, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66300 }, { "accuracy": 47.5586, "active_queue_size": 16384.0, "cl_loss": 3.5656, "doc_norm": 1.5636, "encoder_q-embeddings": 5171.7559, "encoder_q-layer.0": 3537.6707, "encoder_q-layer.1": 3693.199, "encoder_q-layer.10": 6215.9697, "encoder_q-layer.11": 14526.4189, "encoder_q-layer.2": 4215.9712, "encoder_q-layer.3": 4344.5566, "encoder_q-layer.4": 4646.0205, "encoder_q-layer.5": 4588.7017, "encoder_q-layer.6": 5167.7935, "encoder_q-layer.7": 5374.2593, "encoder_q-layer.8": 6142.439, "encoder_q-layer.9": 5689.0322, "epoch": 0.29, "inbatch_neg_score": 0.6785, "inbatch_pos_score": 1.2666, "learning_rate": 3.515789473684211e-05, "loss": 3.5656, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 66.6729, "num_token_overlap": 11.6396, "num_token_query": 31.2573, "num_token_union": 65.0149, "num_word_context": 202.3132, "num_word_doc": 49.7721, "num_word_query": 23.2049, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9318.945, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6782, "query_norm": 1.4866, "queue_k_norm": 1.563, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2573, "sent_len_1": 66.6729, "sent_len_max_0": 127.9875, "sent_len_max_1": 206.87, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66400 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.5647, "doc_norm": 1.5661, "encoder_q-embeddings": 6163.7241, "encoder_q-layer.0": 4299.001, "encoder_q-layer.1": 4734.9541, "encoder_q-layer.10": 6924.8276, "encoder_q-layer.11": 15037.7031, "encoder_q-layer.2": 5728.3105, "encoder_q-layer.3": 5979.0166, "encoder_q-layer.4": 6793.1045, "encoder_q-layer.5": 6756.3018, "encoder_q-layer.6": 6716.7544, "encoder_q-layer.7": 6136.6401, "encoder_q-layer.8": 6980.189, "encoder_q-layer.9": 6451.3604, "epoch": 0.29, "inbatch_neg_score": 0.6811, "inbatch_pos_score": 1.2627, "learning_rate": 3.513157894736842e-05, "loss": 3.5647, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.6986, "num_token_overlap": 11.6785, "num_token_query": 31.4279, "num_token_union": 65.0961, "num_word_context": 202.3578, "num_word_doc": 49.7928, "num_word_query": 23.3492, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10642.2664, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6807, "query_norm": 1.4837, "queue_k_norm": 1.5638, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4279, "sent_len_1": 66.6986, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.0325, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 66500 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.5418, "doc_norm": 1.5686, "encoder_q-embeddings": 5016.2051, "encoder_q-layer.0": 3404.9443, "encoder_q-layer.1": 3644.375, "encoder_q-layer.10": 6921.0933, "encoder_q-layer.11": 15717.2617, "encoder_q-layer.2": 4005.3545, "encoder_q-layer.3": 4219.7554, "encoder_q-layer.4": 4371.1465, "encoder_q-layer.5": 4555.6694, "encoder_q-layer.6": 5253.5469, "encoder_q-layer.7": 5801.416, "encoder_q-layer.8": 7082.4189, "encoder_q-layer.9": 6535.8047, "epoch": 0.29, "inbatch_neg_score": 0.6809, "inbatch_pos_score": 1.2744, "learning_rate": 3.5105263157894744e-05, "loss": 3.5418, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7717, "num_token_overlap": 11.6978, "num_token_query": 31.3577, "num_token_union": 65.0855, "num_word_context": 202.3315, "num_word_doc": 49.8093, "num_word_query": 23.2965, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9640.345, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6797, "query_norm": 1.4924, "queue_k_norm": 1.5677, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3577, "sent_len_1": 66.7717, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.67, "stdk": 0.0489, "stdq": 0.0437, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66600 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.5375, "doc_norm": 1.5703, "encoder_q-embeddings": 5845.0518, "encoder_q-layer.0": 3986.6851, "encoder_q-layer.1": 4261.9307, "encoder_q-layer.10": 6517.2466, "encoder_q-layer.11": 15228.3721, "encoder_q-layer.2": 4967.3193, "encoder_q-layer.3": 4960.7959, "encoder_q-layer.4": 5353.5049, "encoder_q-layer.5": 5706.2886, "encoder_q-layer.6": 6041.0034, "encoder_q-layer.7": 6358.2246, "encoder_q-layer.8": 6673.9907, "encoder_q-layer.9": 6104.645, "epoch": 0.29, "inbatch_neg_score": 0.6803, "inbatch_pos_score": 1.2666, "learning_rate": 3.5078947368421055e-05, "loss": 3.5375, "norm_diff": 0.0897, "norm_loss": 0.0, "num_token_doc": 66.9285, "num_token_overlap": 11.7188, "num_token_query": 31.4737, "num_token_union": 65.2431, "num_word_context": 202.6673, "num_word_doc": 49.9455, "num_word_query": 23.3893, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10191.5063, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6807, "query_norm": 1.4806, "queue_k_norm": 1.5694, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4737, "sent_len_1": 66.9285, "sent_len_max_0": 128.0, "sent_len_max_1": 210.025, "stdk": 0.0489, "stdq": 0.0433, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66700 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.5577, "doc_norm": 1.5704, "encoder_q-embeddings": 5621.5664, "encoder_q-layer.0": 3631.5715, "encoder_q-layer.1": 3730.855, "encoder_q-layer.10": 6944.6479, "encoder_q-layer.11": 15830.2441, "encoder_q-layer.2": 4207.1885, "encoder_q-layer.3": 4290.2339, "encoder_q-layer.4": 4565.0562, "encoder_q-layer.5": 4636.8857, "encoder_q-layer.6": 5405.251, "encoder_q-layer.7": 5851.7925, "encoder_q-layer.8": 6993.3105, "encoder_q-layer.9": 6638.0522, "epoch": 0.29, "inbatch_neg_score": 0.6836, "inbatch_pos_score": 1.2598, "learning_rate": 3.505263157894737e-05, "loss": 3.5577, "norm_diff": 0.104, "norm_loss": 0.0, "num_token_doc": 66.818, "num_token_overlap": 11.6549, "num_token_query": 31.3346, "num_token_union": 65.1371, "num_word_context": 202.4351, "num_word_doc": 49.8405, "num_word_query": 23.2586, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9978.4184, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6826, "query_norm": 1.4664, "queue_k_norm": 1.5715, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3346, "sent_len_1": 66.818, "sent_len_max_0": 127.975, "sent_len_max_1": 209.9187, "stdk": 0.0488, "stdq": 0.0427, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66800 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.552, "doc_norm": 1.5753, "encoder_q-embeddings": 5519.0396, "encoder_q-layer.0": 3814.9275, "encoder_q-layer.1": 4215.8379, "encoder_q-layer.10": 6695.1323, "encoder_q-layer.11": 14599.8389, "encoder_q-layer.2": 4864.4502, "encoder_q-layer.3": 4870.1406, "encoder_q-layer.4": 5049.2095, "encoder_q-layer.5": 5131.0625, "encoder_q-layer.6": 5467.3516, "encoder_q-layer.7": 5685.4717, "encoder_q-layer.8": 6633.3931, "encoder_q-layer.9": 6203.2617, "epoch": 0.29, "inbatch_neg_score": 0.6799, "inbatch_pos_score": 1.2656, "learning_rate": 3.5026315789473685e-05, "loss": 3.552, "norm_diff": 0.111, "norm_loss": 0.0, "num_token_doc": 66.7506, "num_token_overlap": 11.6749, "num_token_query": 31.3641, "num_token_union": 65.1031, "num_word_context": 202.3704, "num_word_doc": 49.8031, "num_word_query": 23.2935, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9873.8186, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6807, "query_norm": 1.4642, "queue_k_norm": 1.5733, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3641, "sent_len_1": 66.7506, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.2138, "stdk": 0.049, "stdq": 0.0426, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 66900 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5525, "doc_norm": 1.5718, "encoder_q-embeddings": 5390.4585, "encoder_q-layer.0": 3762.0632, "encoder_q-layer.1": 3992.0029, "encoder_q-layer.10": 6813.0195, "encoder_q-layer.11": 15100.709, "encoder_q-layer.2": 4491.2827, "encoder_q-layer.3": 4591.7495, "encoder_q-layer.4": 4870.7207, "encoder_q-layer.5": 4945.4419, "encoder_q-layer.6": 5572.4038, "encoder_q-layer.7": 6181.0098, "encoder_q-layer.8": 7273.8135, "encoder_q-layer.9": 6621.8687, "epoch": 0.29, "inbatch_neg_score": 0.679, "inbatch_pos_score": 1.2764, "learning_rate": 3.5e-05, "loss": 3.5525, "norm_diff": 0.0893, "norm_loss": 0.0, "num_token_doc": 66.7876, "num_token_overlap": 11.6805, "num_token_query": 31.3635, "num_token_union": 65.1163, "num_word_context": 202.3554, "num_word_doc": 49.8216, "num_word_query": 23.2926, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9902.9493, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6782, "query_norm": 1.4825, "queue_k_norm": 1.5741, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3635, "sent_len_1": 66.7876, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.6025, "stdk": 0.0488, "stdq": 0.0436, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67000 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.5465, "doc_norm": 1.5717, "encoder_q-embeddings": 5153.3828, "encoder_q-layer.0": 3429.9536, "encoder_q-layer.1": 3712.675, "encoder_q-layer.10": 6384.5815, "encoder_q-layer.11": 14743.7236, "encoder_q-layer.2": 4037.8752, "encoder_q-layer.3": 4277.3911, "encoder_q-layer.4": 4432.0396, "encoder_q-layer.5": 4672.2661, "encoder_q-layer.6": 5079.8306, "encoder_q-layer.7": 5647.0747, "encoder_q-layer.8": 6736.0083, "encoder_q-layer.9": 6052.4038, "epoch": 0.29, "inbatch_neg_score": 0.6845, "inbatch_pos_score": 1.2656, "learning_rate": 3.497368421052632e-05, "loss": 3.5465, "norm_diff": 0.1123, "norm_loss": 0.0, "num_token_doc": 66.6286, "num_token_overlap": 11.6924, "num_token_query": 31.3888, "num_token_union": 65.0318, "num_word_context": 202.0371, "num_word_doc": 49.7243, "num_word_query": 23.3188, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9480.4451, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6846, "query_norm": 1.4593, "queue_k_norm": 1.5727, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3888, "sent_len_1": 66.6286, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.365, "stdk": 0.0487, "stdq": 0.0424, "stdqueue_k": 0.0488, "stdqueue_q": 0.0, "step": 67100 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.5586, "doc_norm": 1.5738, "encoder_q-embeddings": 8586.1963, "encoder_q-layer.0": 6426.5063, "encoder_q-layer.1": 7296.0044, "encoder_q-layer.10": 6665.6138, "encoder_q-layer.11": 14966.6191, "encoder_q-layer.2": 7731.7188, "encoder_q-layer.3": 7434.0586, "encoder_q-layer.4": 7955.4619, "encoder_q-layer.5": 7046.186, "encoder_q-layer.6": 7779.5288, "encoder_q-layer.7": 7722.876, "encoder_q-layer.8": 7162.3584, "encoder_q-layer.9": 6227.8589, "epoch": 0.29, "inbatch_neg_score": 0.6797, "inbatch_pos_score": 1.2666, "learning_rate": 3.4947368421052634e-05, "loss": 3.5586, "norm_diff": 0.1033, "norm_loss": 0.0, "num_token_doc": 66.741, "num_token_overlap": 11.6683, "num_token_query": 31.38, "num_token_union": 65.1321, "num_word_context": 202.17, "num_word_doc": 49.8142, "num_word_query": 23.3173, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12235.2252, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6807, "query_norm": 1.4704, "queue_k_norm": 1.5749, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.38, "sent_len_1": 66.741, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.9625, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67200 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5562, "doc_norm": 1.5835, "encoder_q-embeddings": 15271.0732, "encoder_q-layer.0": 12318.2598, "encoder_q-layer.1": 14180.8887, "encoder_q-layer.10": 6264.3179, "encoder_q-layer.11": 14794.7695, "encoder_q-layer.2": 16201.7334, "encoder_q-layer.3": 15019.999, "encoder_q-layer.4": 14894.2109, "encoder_q-layer.5": 14466.3066, "encoder_q-layer.6": 14372.3428, "encoder_q-layer.7": 11213.0967, "encoder_q-layer.8": 8233.3271, "encoder_q-layer.9": 6395.6323, "epoch": 0.29, "inbatch_neg_score": 0.6827, "inbatch_pos_score": 1.2822, "learning_rate": 3.492105263157895e-05, "loss": 3.5562, "norm_diff": 0.1015, "norm_loss": 0.0, "num_token_doc": 66.7064, "num_token_overlap": 11.6684, "num_token_query": 31.324, "num_token_union": 65.0398, "num_word_context": 202.0902, "num_word_doc": 49.7876, "num_word_query": 23.2652, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19588.1567, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.6836, "query_norm": 1.482, "queue_k_norm": 1.5761, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.324, "sent_len_1": 66.7064, "sent_len_max_0": 127.99, "sent_len_max_1": 208.7262, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67300 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.5533, "doc_norm": 1.5756, "encoder_q-embeddings": 8722.7617, "encoder_q-layer.0": 6385.1284, "encoder_q-layer.1": 5948.7964, "encoder_q-layer.10": 6441.8018, "encoder_q-layer.11": 14933.125, "encoder_q-layer.2": 7406.9092, "encoder_q-layer.3": 6377.3926, "encoder_q-layer.4": 6459.2202, "encoder_q-layer.5": 6304.4243, "encoder_q-layer.6": 6560.2817, "encoder_q-layer.7": 6443.4434, "encoder_q-layer.8": 6715.916, "encoder_q-layer.9": 6235.8281, "epoch": 0.29, "inbatch_neg_score": 0.6832, "inbatch_pos_score": 1.2871, "learning_rate": 3.4894736842105264e-05, "loss": 3.5533, "norm_diff": 0.1074, "norm_loss": 0.0, "num_token_doc": 66.9074, "num_token_overlap": 11.6685, "num_token_query": 31.3593, "num_token_union": 65.2193, "num_word_context": 202.4836, "num_word_doc": 49.9505, "num_word_query": 23.2902, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11726.9643, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6841, "query_norm": 1.4682, "queue_k_norm": 1.5776, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3593, "sent_len_1": 66.9074, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.14, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67400 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.5532, "doc_norm": 1.5785, "encoder_q-embeddings": 7453.0396, "encoder_q-layer.0": 5288.7056, "encoder_q-layer.1": 5736.0303, "encoder_q-layer.10": 6749.4854, "encoder_q-layer.11": 15168.21, "encoder_q-layer.2": 6679.7793, "encoder_q-layer.3": 7460.2515, "encoder_q-layer.4": 7326.6299, "encoder_q-layer.5": 7297.7598, "encoder_q-layer.6": 8758.3408, "encoder_q-layer.7": 8095.4688, "encoder_q-layer.8": 7167.6011, "encoder_q-layer.9": 6623.623, "epoch": 0.29, "inbatch_neg_score": 0.6847, "inbatch_pos_score": 1.2842, "learning_rate": 3.4868421052631575e-05, "loss": 3.5532, "norm_diff": 0.106, "norm_loss": 0.0, "num_token_doc": 66.9879, "num_token_overlap": 11.6661, "num_token_query": 31.301, "num_token_union": 65.1888, "num_word_context": 202.7849, "num_word_doc": 49.9754, "num_word_query": 23.2323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11828.6814, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6855, "query_norm": 1.4725, "queue_k_norm": 1.5786, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.301, "sent_len_1": 66.9879, "sent_len_max_0": 127.9163, "sent_len_max_1": 210.3187, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67500 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.5544, "doc_norm": 1.5816, "encoder_q-embeddings": 10610.3408, "encoder_q-layer.0": 7116.7222, "encoder_q-layer.1": 7591.8276, "encoder_q-layer.10": 13688.251, "encoder_q-layer.11": 29413.2441, "encoder_q-layer.2": 8761.4404, "encoder_q-layer.3": 8721.7422, "encoder_q-layer.4": 9258.333, "encoder_q-layer.5": 9507.8965, "encoder_q-layer.6": 10635.6553, "encoder_q-layer.7": 11618.9619, "encoder_q-layer.8": 13568.8047, "encoder_q-layer.9": 12467.625, "epoch": 0.29, "inbatch_neg_score": 0.6799, "inbatch_pos_score": 1.2803, "learning_rate": 3.48421052631579e-05, "loss": 3.5544, "norm_diff": 0.1095, "norm_loss": 0.0, "num_token_doc": 66.7944, "num_token_overlap": 11.7153, "num_token_query": 31.5055, "num_token_union": 65.203, "num_word_context": 202.3175, "num_word_doc": 49.8438, "num_word_query": 23.3992, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19394.8411, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.6807, "query_norm": 1.4721, "queue_k_norm": 1.5781, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5055, "sent_len_1": 66.7944, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.8963, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67600 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.5521, "doc_norm": 1.5824, "encoder_q-embeddings": 5098.543, "encoder_q-layer.0": 3446.6531, "encoder_q-layer.1": 3551.4924, "encoder_q-layer.10": 6264.1768, "encoder_q-layer.11": 14699.2734, "encoder_q-layer.2": 4095.9612, "encoder_q-layer.3": 4202.8613, "encoder_q-layer.4": 4572.5137, "encoder_q-layer.5": 4679.4604, "encoder_q-layer.6": 5164.5586, "encoder_q-layer.7": 5543.1123, "encoder_q-layer.8": 6422.9976, "encoder_q-layer.9": 6101.8662, "epoch": 0.29, "inbatch_neg_score": 0.6831, "inbatch_pos_score": 1.2764, "learning_rate": 3.481578947368421e-05, "loss": 3.5521, "norm_diff": 0.1173, "norm_loss": 0.0, "num_token_doc": 66.746, "num_token_overlap": 11.6562, "num_token_query": 31.3518, "num_token_union": 65.1053, "num_word_context": 202.1623, "num_word_doc": 49.7983, "num_word_query": 23.2847, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9457.6728, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6831, "query_norm": 1.465, "queue_k_norm": 1.58, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3518, "sent_len_1": 66.746, "sent_len_max_0": 127.9513, "sent_len_max_1": 211.0575, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67700 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.564, "doc_norm": 1.5828, "encoder_q-embeddings": 5770.9434, "encoder_q-layer.0": 3992.1257, "encoder_q-layer.1": 4285.043, "encoder_q-layer.10": 7740.7197, "encoder_q-layer.11": 16897.2871, "encoder_q-layer.2": 4789.7622, "encoder_q-layer.3": 4785.8511, "encoder_q-layer.4": 4955.4312, "encoder_q-layer.5": 5101.0059, "encoder_q-layer.6": 5557.9072, "encoder_q-layer.7": 6486.4658, "encoder_q-layer.8": 7677.2476, "encoder_q-layer.9": 7335.4268, "epoch": 0.29, "inbatch_neg_score": 0.6845, "inbatch_pos_score": 1.2842, "learning_rate": 3.478947368421053e-05, "loss": 3.564, "norm_diff": 0.0955, "norm_loss": 0.0, "num_token_doc": 66.7627, "num_token_overlap": 11.6925, "num_token_query": 31.419, "num_token_union": 65.1406, "num_word_context": 202.4432, "num_word_doc": 49.8153, "num_word_query": 23.336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10577.5741, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6855, "query_norm": 1.4873, "queue_k_norm": 1.5798, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.419, "sent_len_1": 66.7627, "sent_len_max_0": 127.98, "sent_len_max_1": 207.9437, "stdk": 0.049, "stdq": 0.0438, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67800 }, { "accuracy": 46.4844, "active_queue_size": 16384.0, "cl_loss": 3.5639, "doc_norm": 1.5795, "encoder_q-embeddings": 5970.4771, "encoder_q-layer.0": 3957.9485, "encoder_q-layer.1": 4166.5967, "encoder_q-layer.10": 7554.105, "encoder_q-layer.11": 16263.9805, "encoder_q-layer.2": 4790.6943, "encoder_q-layer.3": 4904.4082, "encoder_q-layer.4": 5310.0391, "encoder_q-layer.5": 5186.6235, "encoder_q-layer.6": 5664.8496, "encoder_q-layer.7": 6071.4746, "encoder_q-layer.8": 7562.7417, "encoder_q-layer.9": 6934.6797, "epoch": 0.29, "inbatch_neg_score": 0.684, "inbatch_pos_score": 1.2695, "learning_rate": 3.476315789473684e-05, "loss": 3.5639, "norm_diff": 0.1009, "norm_loss": 0.0, "num_token_doc": 66.7802, "num_token_overlap": 11.6359, "num_token_query": 31.2991, "num_token_union": 65.1342, "num_word_context": 202.247, "num_word_doc": 49.8369, "num_word_query": 23.2455, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10587.6862, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6846, "query_norm": 1.4786, "queue_k_norm": 1.5807, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2991, "sent_len_1": 66.7802, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.3288, "stdk": 0.0488, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 67900 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.5358, "doc_norm": 1.5771, "encoder_q-embeddings": 5023.7534, "encoder_q-layer.0": 3482.0332, "encoder_q-layer.1": 3649.1274, "encoder_q-layer.10": 6871.856, "encoder_q-layer.11": 15453.2871, "encoder_q-layer.2": 4127.2017, "encoder_q-layer.3": 4237.001, "encoder_q-layer.4": 4357.6641, "encoder_q-layer.5": 4608.4292, "encoder_q-layer.6": 5146.4731, "encoder_q-layer.7": 5686.4565, "encoder_q-layer.8": 7193.5786, "encoder_q-layer.9": 6517.417, "epoch": 0.3, "inbatch_neg_score": 0.6846, "inbatch_pos_score": 1.2627, "learning_rate": 3.473684210526316e-05, "loss": 3.5358, "norm_diff": 0.1049, "norm_loss": 0.0, "num_token_doc": 66.6599, "num_token_overlap": 11.6729, "num_token_query": 31.4176, "num_token_union": 65.0884, "num_word_context": 202.305, "num_word_doc": 49.729, "num_word_query": 23.3306, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9685.8863, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6831, "query_norm": 1.4722, "queue_k_norm": 1.5803, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4176, "sent_len_1": 66.6599, "sent_len_max_0": 127.99, "sent_len_max_1": 208.8775, "stdk": 0.0487, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68000 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.5473, "doc_norm": 1.5827, "encoder_q-embeddings": 6226.1113, "encoder_q-layer.0": 4283.4292, "encoder_q-layer.1": 4599.0957, "encoder_q-layer.10": 6846.1436, "encoder_q-layer.11": 14671.8467, "encoder_q-layer.2": 5266.7744, "encoder_q-layer.3": 5512.9517, "encoder_q-layer.4": 5909.7939, "encoder_q-layer.5": 6086.1514, "encoder_q-layer.6": 6118.7954, "encoder_q-layer.7": 6250.3472, "encoder_q-layer.8": 6679.2393, "encoder_q-layer.9": 6261.8745, "epoch": 0.3, "inbatch_neg_score": 0.687, "inbatch_pos_score": 1.2793, "learning_rate": 3.471052631578948e-05, "loss": 3.5473, "norm_diff": 0.1096, "norm_loss": 0.0, "num_token_doc": 66.7908, "num_token_overlap": 11.6706, "num_token_query": 31.334, "num_token_union": 65.0991, "num_word_context": 202.2703, "num_word_doc": 49.8451, "num_word_query": 23.2668, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10347.1447, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6865, "query_norm": 1.4731, "queue_k_norm": 1.5826, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.334, "sent_len_1": 66.7908, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.21, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68100 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.5374, "doc_norm": 1.5844, "encoder_q-embeddings": 7287.0107, "encoder_q-layer.0": 5355.1572, "encoder_q-layer.1": 5268.6426, "encoder_q-layer.10": 6465.2334, "encoder_q-layer.11": 15038.7227, "encoder_q-layer.2": 6169.9717, "encoder_q-layer.3": 6027.2349, "encoder_q-layer.4": 6027.5283, "encoder_q-layer.5": 6082.4341, "encoder_q-layer.6": 6610.6968, "encoder_q-layer.7": 6940.8628, "encoder_q-layer.8": 6991.9644, "encoder_q-layer.9": 6229.4775, "epoch": 0.3, "inbatch_neg_score": 0.6843, "inbatch_pos_score": 1.2666, "learning_rate": 3.468421052631579e-05, "loss": 3.5374, "norm_diff": 0.1155, "norm_loss": 0.0, "num_token_doc": 66.9017, "num_token_overlap": 11.7223, "num_token_query": 31.4908, "num_token_union": 65.252, "num_word_context": 202.5662, "num_word_doc": 49.9345, "num_word_query": 23.3852, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11005.2037, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6846, "query_norm": 1.4689, "queue_k_norm": 1.5837, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4908, "sent_len_1": 66.9017, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.26, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 68200 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.5401, "doc_norm": 1.5811, "encoder_q-embeddings": 5083.0835, "encoder_q-layer.0": 3455.1458, "encoder_q-layer.1": 3740.0081, "encoder_q-layer.10": 6066.292, "encoder_q-layer.11": 14365.418, "encoder_q-layer.2": 4205.2803, "encoder_q-layer.3": 4282.1978, "encoder_q-layer.4": 4487.3286, "encoder_q-layer.5": 4602.4546, "encoder_q-layer.6": 5069.9487, "encoder_q-layer.7": 5875.4961, "encoder_q-layer.8": 6577.5859, "encoder_q-layer.9": 6086.3828, "epoch": 0.3, "inbatch_neg_score": 0.6883, "inbatch_pos_score": 1.2764, "learning_rate": 3.465789473684211e-05, "loss": 3.5401, "norm_diff": 0.1182, "norm_loss": 0.0, "num_token_doc": 66.795, "num_token_overlap": 11.7293, "num_token_query": 31.4175, "num_token_union": 65.1397, "num_word_context": 202.2387, "num_word_doc": 49.8378, "num_word_query": 23.3341, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9428.8388, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.687, "query_norm": 1.4629, "queue_k_norm": 1.5841, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4175, "sent_len_1": 66.795, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.4913, "stdk": 0.0488, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 68300 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.5519, "doc_norm": 1.5798, "encoder_q-embeddings": 4939.4067, "encoder_q-layer.0": 3633.5278, "encoder_q-layer.1": 3953.78, "encoder_q-layer.10": 6551.0205, "encoder_q-layer.11": 15323.4248, "encoder_q-layer.2": 4388.2476, "encoder_q-layer.3": 4506.1748, "encoder_q-layer.4": 4697.8374, "encoder_q-layer.5": 4914.4551, "encoder_q-layer.6": 5229.604, "encoder_q-layer.7": 5919.0249, "encoder_q-layer.8": 6369.7832, "encoder_q-layer.9": 6061.5659, "epoch": 0.3, "inbatch_neg_score": 0.6896, "inbatch_pos_score": 1.2715, "learning_rate": 3.463157894736842e-05, "loss": 3.5519, "norm_diff": 0.1235, "norm_loss": 0.0, "num_token_doc": 66.7695, "num_token_overlap": 11.6753, "num_token_query": 31.4523, "num_token_union": 65.1466, "num_word_context": 202.0875, "num_word_doc": 49.8253, "num_word_query": 23.356, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9552.4353, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.689, "query_norm": 1.4563, "queue_k_norm": 1.583, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4523, "sent_len_1": 66.7695, "sent_len_max_0": 127.965, "sent_len_max_1": 207.925, "stdk": 0.0487, "stdq": 0.0425, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68400 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5559, "doc_norm": 1.5858, "encoder_q-embeddings": 6262.6396, "encoder_q-layer.0": 4325.9629, "encoder_q-layer.1": 4341.6367, "encoder_q-layer.10": 6611.1538, "encoder_q-layer.11": 15050.4229, "encoder_q-layer.2": 4855.0259, "encoder_q-layer.3": 5002.3486, "encoder_q-layer.4": 5315.7432, "encoder_q-layer.5": 5408.5181, "encoder_q-layer.6": 5849.2505, "encoder_q-layer.7": 6491.1655, "encoder_q-layer.8": 6942.2163, "encoder_q-layer.9": 6147.3286, "epoch": 0.3, "inbatch_neg_score": 0.6858, "inbatch_pos_score": 1.2627, "learning_rate": 3.460526315789474e-05, "loss": 3.5559, "norm_diff": 0.1236, "norm_loss": 0.0, "num_token_doc": 66.6859, "num_token_overlap": 11.6627, "num_token_query": 31.3545, "num_token_union": 65.0971, "num_word_context": 202.2701, "num_word_doc": 49.7636, "num_word_query": 23.3022, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10356.9969, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.686, "query_norm": 1.4622, "queue_k_norm": 1.5857, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3545, "sent_len_1": 66.6859, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.2713, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 68500 }, { "accuracy": 47.9492, "active_queue_size": 16384.0, "cl_loss": 3.5571, "doc_norm": 1.5783, "encoder_q-embeddings": 5300.2227, "encoder_q-layer.0": 3699.282, "encoder_q-layer.1": 3992.3823, "encoder_q-layer.10": 6449.6704, "encoder_q-layer.11": 14725.5605, "encoder_q-layer.2": 4423.4619, "encoder_q-layer.3": 4552.4746, "encoder_q-layer.4": 4987.8647, "encoder_q-layer.5": 5283.6035, "encoder_q-layer.6": 6130.3472, "encoder_q-layer.7": 7025.8877, "encoder_q-layer.8": 7967.5293, "encoder_q-layer.9": 6468.541, "epoch": 0.3, "inbatch_neg_score": 0.6842, "inbatch_pos_score": 1.2715, "learning_rate": 3.457894736842106e-05, "loss": 3.5571, "norm_diff": 0.1111, "norm_loss": 0.0, "num_token_doc": 66.6001, "num_token_overlap": 11.6835, "num_token_query": 31.3475, "num_token_union": 64.9896, "num_word_context": 202.1442, "num_word_doc": 49.7015, "num_word_query": 23.2788, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10081.349, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6846, "query_norm": 1.4671, "queue_k_norm": 1.5855, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3475, "sent_len_1": 66.6001, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.07, "stdk": 0.0486, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 68600 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5413, "doc_norm": 1.582, "encoder_q-embeddings": 5748.04, "encoder_q-layer.0": 3953.4536, "encoder_q-layer.1": 4218.7007, "encoder_q-layer.10": 6531.7227, "encoder_q-layer.11": 14667.9414, "encoder_q-layer.2": 4810.9316, "encoder_q-layer.3": 4980.1543, "encoder_q-layer.4": 5183.5469, "encoder_q-layer.5": 5410.7109, "encoder_q-layer.6": 5608.8027, "encoder_q-layer.7": 5965.9312, "encoder_q-layer.8": 6946.9277, "encoder_q-layer.9": 6363.335, "epoch": 0.3, "inbatch_neg_score": 0.6863, "inbatch_pos_score": 1.2852, "learning_rate": 3.455263157894737e-05, "loss": 3.5413, "norm_diff": 0.1062, "norm_loss": 0.0, "num_token_doc": 66.6569, "num_token_overlap": 11.6831, "num_token_query": 31.3662, "num_token_union": 65.0436, "num_word_context": 202.3569, "num_word_doc": 49.7803, "num_word_query": 23.315, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10052.5535, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.686, "query_norm": 1.4758, "queue_k_norm": 1.5837, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3662, "sent_len_1": 66.6569, "sent_len_max_0": 127.99, "sent_len_max_1": 210.4638, "stdk": 0.0487, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68700 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.5507, "doc_norm": 1.5829, "encoder_q-embeddings": 5055.9287, "encoder_q-layer.0": 3430.6116, "encoder_q-layer.1": 3669.2751, "encoder_q-layer.10": 7031.1343, "encoder_q-layer.11": 14666.8799, "encoder_q-layer.2": 3990.5117, "encoder_q-layer.3": 4162.8975, "encoder_q-layer.4": 4349.0439, "encoder_q-layer.5": 4510.9448, "encoder_q-layer.6": 5251.4731, "encoder_q-layer.7": 5922.2808, "encoder_q-layer.8": 7344.4092, "encoder_q-layer.9": 6778.0259, "epoch": 0.3, "inbatch_neg_score": 0.6876, "inbatch_pos_score": 1.2803, "learning_rate": 3.452631578947369e-05, "loss": 3.5507, "norm_diff": 0.1131, "norm_loss": 0.0, "num_token_doc": 66.6575, "num_token_overlap": 11.6791, "num_token_query": 31.3947, "num_token_union": 65.0517, "num_word_context": 202.1572, "num_word_doc": 49.7362, "num_word_query": 23.3142, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9712.0407, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.688, "query_norm": 1.4698, "queue_k_norm": 1.5855, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3947, "sent_len_1": 66.6575, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.9638, "stdk": 0.0488, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68800 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.5541, "doc_norm": 1.5857, "encoder_q-embeddings": 6071.6968, "encoder_q-layer.0": 4411.27, "encoder_q-layer.1": 4730.9067, "encoder_q-layer.10": 6098.9805, "encoder_q-layer.11": 14552.1426, "encoder_q-layer.2": 5316.7939, "encoder_q-layer.3": 5131.356, "encoder_q-layer.4": 5244.1968, "encoder_q-layer.5": 4911.4814, "encoder_q-layer.6": 5189.8804, "encoder_q-layer.7": 6119.8491, "encoder_q-layer.8": 6830.0088, "encoder_q-layer.9": 5981.5288, "epoch": 0.3, "inbatch_neg_score": 0.6908, "inbatch_pos_score": 1.2891, "learning_rate": 3.45e-05, "loss": 3.5541, "norm_diff": 0.1094, "norm_loss": 0.0, "num_token_doc": 66.8696, "num_token_overlap": 11.6674, "num_token_query": 31.3155, "num_token_union": 65.1116, "num_word_context": 202.4244, "num_word_doc": 49.9048, "num_word_query": 23.2484, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9970.1131, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6904, "query_norm": 1.4763, "queue_k_norm": 1.5852, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3155, "sent_len_1": 66.8696, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7388, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 68900 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.5484, "doc_norm": 1.5862, "encoder_q-embeddings": 5913.8887, "encoder_q-layer.0": 3903.6165, "encoder_q-layer.1": 4002.1667, "encoder_q-layer.10": 6556.1499, "encoder_q-layer.11": 15085.3857, "encoder_q-layer.2": 4120.7788, "encoder_q-layer.3": 4101.98, "encoder_q-layer.4": 4492.1172, "encoder_q-layer.5": 4677.0386, "encoder_q-layer.6": 5342.4609, "encoder_q-layer.7": 6382.7983, "encoder_q-layer.8": 7403.7607, "encoder_q-layer.9": 6612.019, "epoch": 0.3, "inbatch_neg_score": 0.6928, "inbatch_pos_score": 1.2852, "learning_rate": 3.447368421052632e-05, "loss": 3.5484, "norm_diff": 0.1138, "norm_loss": 0.0, "num_token_doc": 66.8643, "num_token_overlap": 11.6816, "num_token_query": 31.3854, "num_token_union": 65.1814, "num_word_context": 202.5346, "num_word_doc": 49.9198, "num_word_query": 23.2917, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9948.4463, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6934, "query_norm": 1.4723, "queue_k_norm": 1.5857, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3854, "sent_len_1": 66.8643, "sent_len_max_0": 127.99, "sent_len_max_1": 209.0838, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69000 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.5624, "doc_norm": 1.584, "encoder_q-embeddings": 3119.9551, "encoder_q-layer.0": 2280.7993, "encoder_q-layer.1": 2427.0273, "encoder_q-layer.10": 3331.2822, "encoder_q-layer.11": 7779.3223, "encoder_q-layer.2": 2803.4126, "encoder_q-layer.3": 2986.9106, "encoder_q-layer.4": 3344.8013, "encoder_q-layer.5": 3374.123, "encoder_q-layer.6": 3660.8394, "encoder_q-layer.7": 3529.7307, "encoder_q-layer.8": 3814.1047, "encoder_q-layer.9": 3278.155, "epoch": 0.3, "inbatch_neg_score": 0.6931, "inbatch_pos_score": 1.2666, "learning_rate": 3.4447368421052635e-05, "loss": 3.5624, "norm_diff": 0.127, "norm_loss": 0.0, "num_token_doc": 66.6796, "num_token_overlap": 11.6406, "num_token_query": 31.3648, "num_token_union": 65.0807, "num_word_context": 202.3308, "num_word_doc": 49.7716, "num_word_query": 23.3182, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5635.1022, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6934, "query_norm": 1.457, "queue_k_norm": 1.5858, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3648, "sent_len_1": 66.6796, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.1675, "stdk": 0.0488, "stdq": 0.0424, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69100 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.5442, "doc_norm": 1.5865, "encoder_q-embeddings": 2634.2075, "encoder_q-layer.0": 1748.818, "encoder_q-layer.1": 1915.814, "encoder_q-layer.10": 3363.1111, "encoder_q-layer.11": 7343.0498, "encoder_q-layer.2": 2067.2205, "encoder_q-layer.3": 2133.4517, "encoder_q-layer.4": 2310.447, "encoder_q-layer.5": 2299.6394, "encoder_q-layer.6": 2523.9954, "encoder_q-layer.7": 2838.2654, "encoder_q-layer.8": 3378.8337, "encoder_q-layer.9": 3121.1523, "epoch": 0.3, "inbatch_neg_score": 0.6924, "inbatch_pos_score": 1.293, "learning_rate": 3.442105263157895e-05, "loss": 3.5442, "norm_diff": 0.1097, "norm_loss": 0.0, "num_token_doc": 66.7859, "num_token_overlap": 11.6897, "num_token_query": 31.3915, "num_token_union": 65.1424, "num_word_context": 202.1803, "num_word_doc": 49.8663, "num_word_query": 23.3095, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4725.973, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6929, "query_norm": 1.4768, "queue_k_norm": 1.5868, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3915, "sent_len_1": 66.7859, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.1662, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69200 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.552, "doc_norm": 1.5879, "encoder_q-embeddings": 2780.6057, "encoder_q-layer.0": 1974.5381, "encoder_q-layer.1": 2060.7131, "encoder_q-layer.10": 3170.7068, "encoder_q-layer.11": 7306.2803, "encoder_q-layer.2": 2299.9873, "encoder_q-layer.3": 2324.8992, "encoder_q-layer.4": 2486.4114, "encoder_q-layer.5": 2462.8008, "encoder_q-layer.6": 2706.426, "encoder_q-layer.7": 2982.6545, "encoder_q-layer.8": 3269.4509, "encoder_q-layer.9": 2941.6262, "epoch": 0.3, "inbatch_neg_score": 0.6925, "inbatch_pos_score": 1.2783, "learning_rate": 3.4394736842105265e-05, "loss": 3.552, "norm_diff": 0.1208, "norm_loss": 0.0, "num_token_doc": 66.6695, "num_token_overlap": 11.6782, "num_token_query": 31.4221, "num_token_union": 65.0764, "num_word_context": 202.4682, "num_word_doc": 49.7739, "num_word_query": 23.3458, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4899.6578, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6924, "query_norm": 1.4671, "queue_k_norm": 1.5857, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4221, "sent_len_1": 66.6695, "sent_len_max_0": 128.0, "sent_len_max_1": 208.4925, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69300 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.5458, "doc_norm": 1.5843, "encoder_q-embeddings": 3950.8376, "encoder_q-layer.0": 2928.5857, "encoder_q-layer.1": 3195.5637, "encoder_q-layer.10": 3273.6289, "encoder_q-layer.11": 7229.7271, "encoder_q-layer.2": 3575.9832, "encoder_q-layer.3": 3647.0425, "encoder_q-layer.4": 3794.384, "encoder_q-layer.5": 3672.384, "encoder_q-layer.6": 3773.1316, "encoder_q-layer.7": 3732.175, "encoder_q-layer.8": 3456.8086, "encoder_q-layer.9": 2971.2732, "epoch": 0.3, "inbatch_neg_score": 0.6948, "inbatch_pos_score": 1.2812, "learning_rate": 3.436842105263158e-05, "loss": 3.5458, "norm_diff": 0.1156, "norm_loss": 0.0, "num_token_doc": 66.8532, "num_token_overlap": 11.6738, "num_token_query": 31.3772, "num_token_union": 65.1744, "num_word_context": 202.6464, "num_word_doc": 49.9104, "num_word_query": 23.3126, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5871.5838, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.6943, "query_norm": 1.4687, "queue_k_norm": 1.5853, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3772, "sent_len_1": 66.8532, "sent_len_max_0": 127.965, "sent_len_max_1": 208.9787, "stdk": 0.0488, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69400 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.5398, "doc_norm": 1.5837, "encoder_q-embeddings": 3111.4414, "encoder_q-layer.0": 2174.6265, "encoder_q-layer.1": 2379.7358, "encoder_q-layer.10": 3335.0557, "encoder_q-layer.11": 7196.0225, "encoder_q-layer.2": 2730.1716, "encoder_q-layer.3": 2929.6499, "encoder_q-layer.4": 3097.9392, "encoder_q-layer.5": 2862.7092, "encoder_q-layer.6": 3251.1519, "encoder_q-layer.7": 3314.2756, "encoder_q-layer.8": 3557.8926, "encoder_q-layer.9": 3172.375, "epoch": 0.3, "inbatch_neg_score": 0.6899, "inbatch_pos_score": 1.2793, "learning_rate": 3.4342105263157895e-05, "loss": 3.5398, "norm_diff": 0.1182, "norm_loss": 0.0, "num_token_doc": 66.8541, "num_token_overlap": 11.6805, "num_token_query": 31.3851, "num_token_union": 65.1752, "num_word_context": 202.1871, "num_word_doc": 49.8823, "num_word_query": 23.3006, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5195.3525, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6899, "query_norm": 1.4655, "queue_k_norm": 1.5882, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3851, "sent_len_1": 66.8541, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.7113, "stdk": 0.0487, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 69500 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.5475, "doc_norm": 1.5923, "encoder_q-embeddings": 2663.8186, "encoder_q-layer.0": 1807.392, "encoder_q-layer.1": 1842.9659, "encoder_q-layer.10": 3341.5906, "encoder_q-layer.11": 7408.0137, "encoder_q-layer.2": 2057.9636, "encoder_q-layer.3": 2088.4465, "encoder_q-layer.4": 2190.4187, "encoder_q-layer.5": 2193.1445, "encoder_q-layer.6": 2471.353, "encoder_q-layer.7": 2736.9089, "encoder_q-layer.8": 3480.074, "encoder_q-layer.9": 3091.374, "epoch": 0.3, "inbatch_neg_score": 0.6906, "inbatch_pos_score": 1.2939, "learning_rate": 3.431578947368421e-05, "loss": 3.5475, "norm_diff": 0.1141, "norm_loss": 0.0, "num_token_doc": 66.7655, "num_token_overlap": 11.6622, "num_token_query": 31.3576, "num_token_union": 65.1235, "num_word_context": 202.1046, "num_word_doc": 49.8256, "num_word_query": 23.2927, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4770.2503, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6904, "query_norm": 1.4782, "queue_k_norm": 1.5877, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3576, "sent_len_1": 66.7655, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.5625, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69600 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.5455, "doc_norm": 1.59, "encoder_q-embeddings": 2890.1487, "encoder_q-layer.0": 1966.8931, "encoder_q-layer.1": 2091.6807, "encoder_q-layer.10": 3203.0386, "encoder_q-layer.11": 7452.6025, "encoder_q-layer.2": 2373.4014, "encoder_q-layer.3": 2375.3005, "encoder_q-layer.4": 2492.3564, "encoder_q-layer.5": 2489.8425, "encoder_q-layer.6": 2773.8621, "encoder_q-layer.7": 3184.615, "encoder_q-layer.8": 3675.8188, "encoder_q-layer.9": 3163.4402, "epoch": 0.3, "inbatch_neg_score": 0.6923, "inbatch_pos_score": 1.2637, "learning_rate": 3.4289473684210525e-05, "loss": 3.5455, "norm_diff": 0.1172, "norm_loss": 0.0, "num_token_doc": 66.8643, "num_token_overlap": 11.6732, "num_token_query": 31.4069, "num_token_union": 65.1797, "num_word_context": 202.214, "num_word_doc": 49.8617, "num_word_query": 23.3344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5066.8489, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6924, "query_norm": 1.4728, "queue_k_norm": 1.5865, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4069, "sent_len_1": 66.8643, "sent_len_max_0": 127.9925, "sent_len_max_1": 211.8787, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69700 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5456, "doc_norm": 1.5875, "encoder_q-embeddings": 2545.8872, "encoder_q-layer.0": 1743.5985, "encoder_q-layer.1": 1848.4591, "encoder_q-layer.10": 3592.3513, "encoder_q-layer.11": 7926.4248, "encoder_q-layer.2": 2040.7937, "encoder_q-layer.3": 2108.8506, "encoder_q-layer.4": 2254.4636, "encoder_q-layer.5": 2294.4583, "encoder_q-layer.6": 2512.3992, "encoder_q-layer.7": 2953.8108, "encoder_q-layer.8": 3724.1309, "encoder_q-layer.9": 3472.1348, "epoch": 0.3, "inbatch_neg_score": 0.691, "inbatch_pos_score": 1.2725, "learning_rate": 3.426315789473684e-05, "loss": 3.5456, "norm_diff": 0.1013, "norm_loss": 0.0, "num_token_doc": 66.74, "num_token_overlap": 11.6458, "num_token_query": 31.3422, "num_token_union": 65.1068, "num_word_context": 202.2297, "num_word_doc": 49.8311, "num_word_query": 23.2836, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4997.6946, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.6909, "query_norm": 1.4863, "queue_k_norm": 1.5875, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3422, "sent_len_1": 66.74, "sent_len_max_0": 128.0, "sent_len_max_1": 207.3225, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69800 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.5343, "doc_norm": 1.5877, "encoder_q-embeddings": 2683.9946, "encoder_q-layer.0": 1860.5216, "encoder_q-layer.1": 1922.0089, "encoder_q-layer.10": 3149.2969, "encoder_q-layer.11": 7133.1172, "encoder_q-layer.2": 2151.4692, "encoder_q-layer.3": 2247.0505, "encoder_q-layer.4": 2258.6567, "encoder_q-layer.5": 2306.6533, "encoder_q-layer.6": 2651.8987, "encoder_q-layer.7": 3058.4092, "encoder_q-layer.8": 3418.8352, "encoder_q-layer.9": 3084.6626, "epoch": 0.3, "inbatch_neg_score": 0.7017, "inbatch_pos_score": 1.3018, "learning_rate": 3.423684210526316e-05, "loss": 3.5343, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.8514, "num_token_overlap": 11.7349, "num_token_query": 31.4871, "num_token_union": 65.2042, "num_word_context": 202.5519, "num_word_doc": 49.8659, "num_word_query": 23.3892, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4747.085, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7007, "query_norm": 1.5084, "queue_k_norm": 1.5879, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4871, "sent_len_1": 66.8514, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.9313, "stdk": 0.0489, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 69900 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.5176, "doc_norm": 1.589, "encoder_q-embeddings": 2627.4543, "encoder_q-layer.0": 1704.4766, "encoder_q-layer.1": 1775.533, "encoder_q-layer.10": 3224.5461, "encoder_q-layer.11": 7289.4443, "encoder_q-layer.2": 1939.2633, "encoder_q-layer.3": 2020.4424, "encoder_q-layer.4": 2157.1946, "encoder_q-layer.5": 2180.5195, "encoder_q-layer.6": 2459.1182, "encoder_q-layer.7": 2739.324, "encoder_q-layer.8": 3441.8931, "encoder_q-layer.9": 3060.3987, "epoch": 0.3, "inbatch_neg_score": 0.7069, "inbatch_pos_score": 1.3018, "learning_rate": 3.421052631578947e-05, "loss": 3.5176, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.8856, "num_token_overlap": 11.7076, "num_token_query": 31.4265, "num_token_union": 65.2221, "num_word_context": 202.6154, "num_word_doc": 49.9187, "num_word_query": 23.3358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4657.7088, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7065, "query_norm": 1.5173, "queue_k_norm": 1.5886, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4265, "sent_len_1": 66.8856, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.1113, "stdk": 0.0489, "stdq": 0.0433, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70000 }, { "dev_runtime": 27.7006, "dev_samples_per_second": 1.155, "dev_steps_per_second": 0.036, "epoch": 0.3, "step": 70000, "test_accuracy": 93.6279296875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3584204316139221, "test_doc_norm": 1.558947205543518, "test_inbatch_neg_score": 0.9807655811309814, "test_inbatch_pos_score": 1.97483491897583, "test_loss": 0.3584204316139221, "test_loss_align": 1.1543536186218262, "test_loss_unif": 3.040229320526123, "test_loss_unif_q@queue": 3.040229558944702, "test_norm_diff": 0.06857950985431671, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.7001984119415283, "test_query_norm": 1.6275267601013184, "test_queue_k_norm": 1.588363766670227, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043525390326976776, "test_stdq": 0.04370199516415596, "test_stdqueue_k": 0.04892833158373833, "test_stdqueue_q": 0.0 }, { "dev_runtime": 27.7006, "dev_samples_per_second": 1.155, "dev_steps_per_second": 0.036, "epoch": 0.3, "eval_beir-arguana_ndcg@10": 0.36938, "eval_beir-arguana_recall@10": 0.63158, "eval_beir-arguana_recall@100": 0.92248, "eval_beir-arguana_recall@20": 0.77169, "eval_beir-avg_ndcg@10": 0.3905355833333333, "eval_beir-avg_recall@10": 0.463302, "eval_beir-avg_recall@100": 0.6371299166666666, "eval_beir-avg_recall@20": 0.5221353333333334, "eval_beir-cqadupstack_ndcg@10": 0.28709583333333333, "eval_beir-cqadupstack_recall@10": 0.38733999999999996, "eval_beir-cqadupstack_recall@100": 0.6185391666666668, "eval_beir-cqadupstack_recall@20": 0.4539233333333333, "eval_beir-fiqa_ndcg@10": 0.25963, "eval_beir-fiqa_recall@10": 0.33291, "eval_beir-fiqa_recall@100": 0.58694, "eval_beir-fiqa_recall@20": 0.40243, "eval_beir-nfcorpus_ndcg@10": 0.31788, "eval_beir-nfcorpus_recall@10": 0.15606, "eval_beir-nfcorpus_recall@100": 0.29478, "eval_beir-nfcorpus_recall@20": 0.18548, "eval_beir-nq_ndcg@10": 0.30803, "eval_beir-nq_recall@10": 0.49556, "eval_beir-nq_recall@100": 0.82457, "eval_beir-nq_recall@20": 0.60839, "eval_beir-quora_ndcg@10": 0.75669, "eval_beir-quora_recall@10": 0.87085, "eval_beir-quora_recall@100": 0.97317, "eval_beir-quora_recall@20": 0.91654, "eval_beir-scidocs_ndcg@10": 0.1581, "eval_beir-scidocs_recall@10": 0.16752, "eval_beir-scidocs_recall@100": 0.37785, "eval_beir-scidocs_recall@20": 0.22805, "eval_beir-scifact_ndcg@10": 0.64113, "eval_beir-scifact_recall@10": 0.79567, "eval_beir-scifact_recall@100": 0.90489, "eval_beir-scifact_recall@20": 0.84022, "eval_beir-trec-covid_ndcg@10": 0.62343, "eval_beir-trec-covid_recall@10": 0.668, "eval_beir-trec-covid_recall@100": 0.443, "eval_beir-trec-covid_recall@20": 0.604, "eval_beir-webis-touche2020_ndcg@10": 0.18399, "eval_beir-webis-touche2020_recall@10": 0.12753, "eval_beir-webis-touche2020_recall@100": 0.42508, "eval_beir-webis-touche2020_recall@20": 0.21063, "eval_senteval-avg_sts": 0.7537021923793231, "eval_senteval-sickr_spearman": 0.7316208803208656, "eval_senteval-stsb_spearman": 0.7757835044377807, "step": 70000, "test_accuracy": 93.6279296875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3584204316139221, "test_doc_norm": 1.558947205543518, "test_inbatch_neg_score": 0.9807655811309814, "test_inbatch_pos_score": 1.97483491897583, "test_loss": 0.3584204316139221, "test_loss_align": 1.1543536186218262, "test_loss_unif": 3.040229320526123, "test_loss_unif_q@queue": 3.040229558944702, "test_norm_diff": 0.06857950985431671, "test_norm_loss": 0.0, "test_q@queue_neg_score": 0.7001984119415283, "test_query_norm": 1.6275267601013184, "test_queue_k_norm": 1.588363766670227, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043525390326976776, "test_stdq": 0.04370199516415596, "test_stdqueue_k": 0.04892833158373833, "test_stdqueue_q": 0.0 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.5347, "doc_norm": 1.589, "encoder_q-embeddings": 2807.4851, "encoder_q-layer.0": 1855.5056, "encoder_q-layer.1": 2014.3788, "encoder_q-layer.10": 3593.3379, "encoder_q-layer.11": 7818.689, "encoder_q-layer.2": 2217.0869, "encoder_q-layer.3": 2230.4998, "encoder_q-layer.4": 2326.6384, "encoder_q-layer.5": 2391.1726, "encoder_q-layer.6": 2680.2268, "encoder_q-layer.7": 3079.8667, "encoder_q-layer.8": 3769.0159, "encoder_q-layer.9": 3279.5247, "epoch": 0.3, "inbatch_neg_score": 0.7194, "inbatch_pos_score": 1.3213, "learning_rate": 3.418421052631579e-05, "loss": 3.5347, "norm_diff": 0.0503, "norm_loss": 0.0, "num_token_doc": 66.7323, "num_token_overlap": 11.7266, "num_token_query": 31.5282, "num_token_union": 65.1567, "num_word_context": 202.3902, "num_word_doc": 49.8096, "num_word_query": 23.4309, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5110.2562, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7202, "query_norm": 1.5387, "queue_k_norm": 1.5881, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5282, "sent_len_1": 66.7323, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.6312, "stdk": 0.0489, "stdq": 0.0438, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70100 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.5538, "doc_norm": 1.5949, "encoder_q-embeddings": 2674.7664, "encoder_q-layer.0": 1757.2596, "encoder_q-layer.1": 1834.2665, "encoder_q-layer.10": 3286.3628, "encoder_q-layer.11": 7504.2368, "encoder_q-layer.2": 2064.2983, "encoder_q-layer.3": 2135.166, "encoder_q-layer.4": 2309.9211, "encoder_q-layer.5": 2381.8821, "encoder_q-layer.6": 2703.5002, "encoder_q-layer.7": 2975.7625, "encoder_q-layer.8": 3489.0273, "encoder_q-layer.9": 3243.656, "epoch": 0.3, "inbatch_neg_score": 0.7255, "inbatch_pos_score": 1.3174, "learning_rate": 3.41578947368421e-05, "loss": 3.5538, "norm_diff": 0.0631, "norm_loss": 0.0, "num_token_doc": 66.747, "num_token_overlap": 11.6911, "num_token_query": 31.4122, "num_token_union": 65.1166, "num_word_context": 202.1505, "num_word_doc": 49.7954, "num_word_query": 23.3191, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4876.0816, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7266, "query_norm": 1.5318, "queue_k_norm": 1.5888, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4122, "sent_len_1": 66.747, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.8475, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70200 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.5361, "doc_norm": 1.5928, "encoder_q-embeddings": 3857.7317, "encoder_q-layer.0": 2603.0112, "encoder_q-layer.1": 2966.6719, "encoder_q-layer.10": 3345.1904, "encoder_q-layer.11": 7473.8169, "encoder_q-layer.2": 3382.0308, "encoder_q-layer.3": 3495.7183, "encoder_q-layer.4": 3595.2573, "encoder_q-layer.5": 3685.9873, "encoder_q-layer.6": 3692.7461, "encoder_q-layer.7": 3623.4089, "encoder_q-layer.8": 3570.5481, "encoder_q-layer.9": 3141.9209, "epoch": 0.31, "inbatch_neg_score": 0.7405, "inbatch_pos_score": 1.3428, "learning_rate": 3.413157894736842e-05, "loss": 3.5361, "norm_diff": 0.0525, "norm_loss": 0.0, "num_token_doc": 66.8327, "num_token_overlap": 11.6546, "num_token_query": 31.429, "num_token_union": 65.1809, "num_word_context": 202.7193, "num_word_doc": 49.8283, "num_word_query": 23.3254, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5810.0466, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.7412, "query_norm": 1.5403, "queue_k_norm": 1.5898, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.429, "sent_len_1": 66.8327, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.7587, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70300 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.5292, "doc_norm": 1.5905, "encoder_q-embeddings": 2566.0723, "encoder_q-layer.0": 1730.823, "encoder_q-layer.1": 1840.6278, "encoder_q-layer.10": 3220.2939, "encoder_q-layer.11": 7428.3945, "encoder_q-layer.2": 2001.7489, "encoder_q-layer.3": 2011.8761, "encoder_q-layer.4": 2093.7739, "encoder_q-layer.5": 2146.4246, "encoder_q-layer.6": 2419.5693, "encoder_q-layer.7": 2740.4084, "encoder_q-layer.8": 3295.1309, "encoder_q-layer.9": 3122.4543, "epoch": 0.31, "inbatch_neg_score": 0.7484, "inbatch_pos_score": 1.3379, "learning_rate": 3.410526315789474e-05, "loss": 3.5292, "norm_diff": 0.068, "norm_loss": 0.0, "num_token_doc": 66.7745, "num_token_overlap": 11.6873, "num_token_query": 31.3709, "num_token_union": 65.0914, "num_word_context": 201.9561, "num_word_doc": 49.8125, "num_word_query": 23.301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4671.1799, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.748, "query_norm": 1.5224, "queue_k_norm": 1.5915, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3709, "sent_len_1": 66.7745, "sent_len_max_0": 127.975, "sent_len_max_1": 209.305, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70400 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.5285, "doc_norm": 1.5932, "encoder_q-embeddings": 2651.8367, "encoder_q-layer.0": 1832.8485, "encoder_q-layer.1": 2003.2588, "encoder_q-layer.10": 3536.6572, "encoder_q-layer.11": 8077.8525, "encoder_q-layer.2": 2220.4255, "encoder_q-layer.3": 2237.1841, "encoder_q-layer.4": 2322.9458, "encoder_q-layer.5": 2329.4285, "encoder_q-layer.6": 2622.7319, "encoder_q-layer.7": 2970.3176, "encoder_q-layer.8": 3762.6562, "encoder_q-layer.9": 3362.4109, "epoch": 0.31, "inbatch_neg_score": 0.7541, "inbatch_pos_score": 1.3369, "learning_rate": 3.407894736842105e-05, "loss": 3.5285, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.9302, "num_token_overlap": 11.677, "num_token_query": 31.3967, "num_token_union": 65.2281, "num_word_context": 202.61, "num_word_doc": 49.9384, "num_word_query": 23.3177, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5074.5925, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7544, "query_norm": 1.5101, "queue_k_norm": 1.5923, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3967, "sent_len_1": 66.9302, "sent_len_max_0": 127.9575, "sent_len_max_1": 208.8088, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70500 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.5436, "doc_norm": 1.5964, "encoder_q-embeddings": 2704.1697, "encoder_q-layer.0": 1806.2783, "encoder_q-layer.1": 1863.5868, "encoder_q-layer.10": 3155.4497, "encoder_q-layer.11": 7452.521, "encoder_q-layer.2": 2003.5903, "encoder_q-layer.3": 2065.6082, "encoder_q-layer.4": 2188.2583, "encoder_q-layer.5": 2253.5312, "encoder_q-layer.6": 2493.9971, "encoder_q-layer.7": 2949.0928, "encoder_q-layer.8": 3271.9282, "encoder_q-layer.9": 3022.2849, "epoch": 0.31, "inbatch_neg_score": 0.755, "inbatch_pos_score": 1.3496, "learning_rate": 3.405263157894737e-05, "loss": 3.5436, "norm_diff": 0.0964, "norm_loss": 0.0, "num_token_doc": 66.7521, "num_token_overlap": 11.6967, "num_token_query": 31.4147, "num_token_union": 65.1098, "num_word_context": 202.2459, "num_word_doc": 49.8219, "num_word_query": 23.3199, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4736.1143, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7554, "query_norm": 1.5, "queue_k_norm": 1.5964, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4147, "sent_len_1": 66.7521, "sent_len_max_0": 127.97, "sent_len_max_1": 207.665, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 70600 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.5436, "doc_norm": 1.5928, "encoder_q-embeddings": 2677.7117, "encoder_q-layer.0": 1732.9567, "encoder_q-layer.1": 1838.3418, "encoder_q-layer.10": 3295.561, "encoder_q-layer.11": 8105.4473, "encoder_q-layer.2": 2063.459, "encoder_q-layer.3": 2151.98, "encoder_q-layer.4": 2307.3521, "encoder_q-layer.5": 2358.8699, "encoder_q-layer.6": 2627.9902, "encoder_q-layer.7": 3084.77, "encoder_q-layer.8": 3599.8262, "encoder_q-layer.9": 3226.8997, "epoch": 0.31, "inbatch_neg_score": 0.7595, "inbatch_pos_score": 1.3516, "learning_rate": 3.402631578947368e-05, "loss": 3.5436, "norm_diff": 0.0853, "norm_loss": 0.0, "num_token_doc": 66.5865, "num_token_overlap": 11.6291, "num_token_query": 31.2876, "num_token_union": 64.9945, "num_word_context": 202.2413, "num_word_doc": 49.6915, "num_word_query": 23.2324, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5069.195, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7598, "query_norm": 1.5075, "queue_k_norm": 1.5975, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2876, "sent_len_1": 66.5865, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.9462, "stdk": 0.0487, "stdq": 0.0434, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70700 }, { "accuracy": 47.1191, "active_queue_size": 16384.0, "cl_loss": 3.5355, "doc_norm": 1.5977, "encoder_q-embeddings": 3159.575, "encoder_q-layer.0": 2203.9504, "encoder_q-layer.1": 2361.6738, "encoder_q-layer.10": 3757.0334, "encoder_q-layer.11": 7946.6284, "encoder_q-layer.2": 2727.8689, "encoder_q-layer.3": 2710.7644, "encoder_q-layer.4": 2911.2698, "encoder_q-layer.5": 3051.3652, "encoder_q-layer.6": 3442.1121, "encoder_q-layer.7": 3631.4866, "encoder_q-layer.8": 4300.1987, "encoder_q-layer.9": 3605.9436, "epoch": 0.31, "inbatch_neg_score": 0.7538, "inbatch_pos_score": 1.3262, "learning_rate": 3.4000000000000007e-05, "loss": 3.5355, "norm_diff": 0.1079, "norm_loss": 0.0, "num_token_doc": 66.9053, "num_token_overlap": 11.6887, "num_token_query": 31.3591, "num_token_union": 65.1395, "num_word_context": 202.2747, "num_word_doc": 49.9032, "num_word_query": 23.3034, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5557.2905, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.7534, "query_norm": 1.4898, "queue_k_norm": 1.6, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3591, "sent_len_1": 66.9053, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.3525, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 70800 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.5403, "doc_norm": 1.5967, "encoder_q-embeddings": 2617.2253, "encoder_q-layer.0": 1723.437, "encoder_q-layer.1": 1802.9607, "encoder_q-layer.10": 3331.8491, "encoder_q-layer.11": 7577.3589, "encoder_q-layer.2": 2020.4702, "encoder_q-layer.3": 2105.9954, "encoder_q-layer.4": 2243.6777, "encoder_q-layer.5": 2235.6731, "encoder_q-layer.6": 2493.1357, "encoder_q-layer.7": 3009.9768, "encoder_q-layer.8": 3778.7319, "encoder_q-layer.9": 3372.8796, "epoch": 0.31, "inbatch_neg_score": 0.7566, "inbatch_pos_score": 1.3555, "learning_rate": 3.397368421052632e-05, "loss": 3.5403, "norm_diff": 0.1037, "norm_loss": 0.0, "num_token_doc": 66.6983, "num_token_overlap": 11.6785, "num_token_query": 31.3965, "num_token_union": 65.0954, "num_word_context": 202.3168, "num_word_doc": 49.7596, "num_word_query": 23.3289, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4926.9359, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7573, "query_norm": 1.4931, "queue_k_norm": 1.5991, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3965, "sent_len_1": 66.6983, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.2038, "stdk": 0.0488, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 70900 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.5438, "doc_norm": 1.6023, "encoder_q-embeddings": 2500.95, "encoder_q-layer.0": 1663.2186, "encoder_q-layer.1": 1785.9424, "encoder_q-layer.10": 3193.8364, "encoder_q-layer.11": 7809.0288, "encoder_q-layer.2": 2013.5652, "encoder_q-layer.3": 2090.6365, "encoder_q-layer.4": 2166.0125, "encoder_q-layer.5": 2179.1138, "encoder_q-layer.6": 2409.8347, "encoder_q-layer.7": 2789.3413, "encoder_q-layer.8": 3155.1099, "encoder_q-layer.9": 2947.6909, "epoch": 0.31, "inbatch_neg_score": 0.7571, "inbatch_pos_score": 1.3613, "learning_rate": 3.3947368421052636e-05, "loss": 3.5438, "norm_diff": 0.103, "norm_loss": 0.0, "num_token_doc": 66.8216, "num_token_overlap": 11.6714, "num_token_query": 31.3431, "num_token_union": 65.1441, "num_word_context": 202.3973, "num_word_doc": 49.8475, "num_word_query": 23.2718, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4751.4834, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7568, "query_norm": 1.4993, "queue_k_norm": 1.601, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3431, "sent_len_1": 66.8216, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.6987, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71000 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.547, "doc_norm": 1.6014, "encoder_q-embeddings": 5874.7451, "encoder_q-layer.0": 3814.2429, "encoder_q-layer.1": 4012.3818, "encoder_q-layer.10": 7517.5015, "encoder_q-layer.11": 17232.3105, "encoder_q-layer.2": 4374.375, "encoder_q-layer.3": 4707.6157, "encoder_q-layer.4": 5035.5371, "encoder_q-layer.5": 5155.7349, "encoder_q-layer.6": 5687.3735, "encoder_q-layer.7": 6422.3354, "encoder_q-layer.8": 7465.5415, "encoder_q-layer.9": 7113.8281, "epoch": 0.31, "inbatch_neg_score": 0.7596, "inbatch_pos_score": 1.3564, "learning_rate": 3.392105263157895e-05, "loss": 3.547, "norm_diff": 0.1062, "norm_loss": 0.0, "num_token_doc": 66.7797, "num_token_overlap": 11.6507, "num_token_query": 31.2966, "num_token_union": 65.1168, "num_word_context": 202.4329, "num_word_doc": 49.8565, "num_word_query": 23.2262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10751.3351, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.7593, "query_norm": 1.4953, "queue_k_norm": 1.6019, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2966, "sent_len_1": 66.7797, "sent_len_max_0": 127.985, "sent_len_max_1": 206.06, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71100 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.521, "doc_norm": 1.5991, "encoder_q-embeddings": 2494.7205, "encoder_q-layer.0": 1693.8649, "encoder_q-layer.1": 1809.6849, "encoder_q-layer.10": 3361.542, "encoder_q-layer.11": 8094.9482, "encoder_q-layer.2": 1999.3285, "encoder_q-layer.3": 2078.6665, "encoder_q-layer.4": 2168.2842, "encoder_q-layer.5": 2231.5657, "encoder_q-layer.6": 2588.3794, "encoder_q-layer.7": 2992.9258, "encoder_q-layer.8": 3506.8274, "encoder_q-layer.9": 3240.5493, "epoch": 0.31, "inbatch_neg_score": 0.76, "inbatch_pos_score": 1.3477, "learning_rate": 3.389473684210526e-05, "loss": 3.521, "norm_diff": 0.0919, "norm_loss": 0.0, "num_token_doc": 66.7526, "num_token_overlap": 11.6763, "num_token_query": 31.3812, "num_token_union": 65.1276, "num_word_context": 202.4244, "num_word_doc": 49.8219, "num_word_query": 23.3073, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4832.1917, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7603, "query_norm": 1.5072, "queue_k_norm": 1.6037, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3812, "sent_len_1": 66.7526, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.0025, "stdk": 0.0487, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71200 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.5428, "doc_norm": 1.6041, "encoder_q-embeddings": 2673.4641, "encoder_q-layer.0": 1733.9486, "encoder_q-layer.1": 1793.9717, "encoder_q-layer.10": 3422.083, "encoder_q-layer.11": 7356.2104, "encoder_q-layer.2": 2042.4282, "encoder_q-layer.3": 2078.6982, "encoder_q-layer.4": 2207.6809, "encoder_q-layer.5": 2289.1335, "encoder_q-layer.6": 2580.3789, "encoder_q-layer.7": 2870.4558, "encoder_q-layer.8": 3622.4048, "encoder_q-layer.9": 3177.656, "epoch": 0.31, "inbatch_neg_score": 0.7607, "inbatch_pos_score": 1.3613, "learning_rate": 3.3868421052631585e-05, "loss": 3.5428, "norm_diff": 0.1131, "norm_loss": 0.0, "num_token_doc": 66.6562, "num_token_overlap": 11.6716, "num_token_query": 31.4246, "num_token_union": 65.0855, "num_word_context": 202.1471, "num_word_doc": 49.7567, "num_word_query": 23.3302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4829.3303, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7607, "query_norm": 1.491, "queue_k_norm": 1.6039, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4246, "sent_len_1": 66.6562, "sent_len_max_0": 127.9625, "sent_len_max_1": 206.9675, "stdk": 0.0489, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71300 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5271, "doc_norm": 1.6022, "encoder_q-embeddings": 2412.6331, "encoder_q-layer.0": 1639.8982, "encoder_q-layer.1": 1730.8365, "encoder_q-layer.10": 3256.3198, "encoder_q-layer.11": 7404.1733, "encoder_q-layer.2": 1933.7455, "encoder_q-layer.3": 1962.5907, "encoder_q-layer.4": 2091.5657, "encoder_q-layer.5": 2153.4531, "encoder_q-layer.6": 2362.8813, "encoder_q-layer.7": 2704.625, "encoder_q-layer.8": 3366.4172, "encoder_q-layer.9": 3120.3911, "epoch": 0.31, "inbatch_neg_score": 0.7602, "inbatch_pos_score": 1.3516, "learning_rate": 3.3842105263157896e-05, "loss": 3.5271, "norm_diff": 0.1085, "norm_loss": 0.0, "num_token_doc": 66.8229, "num_token_overlap": 11.6807, "num_token_query": 31.3653, "num_token_union": 65.1375, "num_word_context": 202.3066, "num_word_doc": 49.8689, "num_word_query": 23.2952, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4670.3054, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7603, "query_norm": 1.4937, "queue_k_norm": 1.6039, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3653, "sent_len_1": 66.8229, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.005, "stdk": 0.0488, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 71400 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.5103, "doc_norm": 1.6064, "encoder_q-embeddings": 2747.2134, "encoder_q-layer.0": 1880.6329, "encoder_q-layer.1": 2005.8441, "encoder_q-layer.10": 3288.3135, "encoder_q-layer.11": 7391.4795, "encoder_q-layer.2": 2272.0654, "encoder_q-layer.3": 2496.6733, "encoder_q-layer.4": 2591.6116, "encoder_q-layer.5": 2732.228, "encoder_q-layer.6": 2866.6296, "encoder_q-layer.7": 3110.3433, "encoder_q-layer.8": 3455.8655, "encoder_q-layer.9": 3174.2327, "epoch": 0.31, "inbatch_neg_score": 0.765, "inbatch_pos_score": 1.3652, "learning_rate": 3.3815789473684215e-05, "loss": 3.5103, "norm_diff": 0.0954, "norm_loss": 0.0, "num_token_doc": 66.7854, "num_token_overlap": 11.6815, "num_token_query": 31.3338, "num_token_union": 65.0855, "num_word_context": 202.1137, "num_word_doc": 49.8455, "num_word_query": 23.2645, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4957.958, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7651, "query_norm": 1.511, "queue_k_norm": 1.6042, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3338, "sent_len_1": 66.7854, "sent_len_max_0": 127.975, "sent_len_max_1": 208.0488, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 71500 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.5223, "doc_norm": 1.61, "encoder_q-embeddings": 2443.449, "encoder_q-layer.0": 1655.2792, "encoder_q-layer.1": 1729.502, "encoder_q-layer.10": 3342.9014, "encoder_q-layer.11": 7316.0293, "encoder_q-layer.2": 1915.1223, "encoder_q-layer.3": 2000.4575, "encoder_q-layer.4": 2138.0281, "encoder_q-layer.5": 2138.668, "encoder_q-layer.6": 2455.6343, "encoder_q-layer.7": 2783.0569, "encoder_q-layer.8": 3369.5166, "encoder_q-layer.9": 3128.3765, "epoch": 0.31, "inbatch_neg_score": 0.7725, "inbatch_pos_score": 1.3779, "learning_rate": 3.3789473684210526e-05, "loss": 3.5223, "norm_diff": 0.09, "norm_loss": 0.0, "num_token_doc": 66.891, "num_token_overlap": 11.6949, "num_token_query": 31.3521, "num_token_union": 65.1766, "num_word_context": 202.0988, "num_word_doc": 49.9075, "num_word_query": 23.2945, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4646.395, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.772, "query_norm": 1.52, "queue_k_norm": 1.607, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3521, "sent_len_1": 66.891, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.7463, "stdk": 0.0491, "stdq": 0.0437, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71600 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.5184, "doc_norm": 1.6033, "encoder_q-embeddings": 2641.491, "encoder_q-layer.0": 1757.6155, "encoder_q-layer.1": 1840.3312, "encoder_q-layer.10": 3220.9585, "encoder_q-layer.11": 7146.353, "encoder_q-layer.2": 2042.1736, "encoder_q-layer.3": 2095.2009, "encoder_q-layer.4": 2197.2993, "encoder_q-layer.5": 2213.2993, "encoder_q-layer.6": 2512.2004, "encoder_q-layer.7": 2845.5938, "encoder_q-layer.8": 3399.0391, "encoder_q-layer.9": 3085.8547, "epoch": 0.31, "inbatch_neg_score": 0.7719, "inbatch_pos_score": 1.3564, "learning_rate": 3.376315789473684e-05, "loss": 3.5184, "norm_diff": 0.0923, "norm_loss": 0.0, "num_token_doc": 66.8695, "num_token_overlap": 11.6977, "num_token_query": 31.4447, "num_token_union": 65.2081, "num_word_context": 202.2433, "num_word_doc": 49.8921, "num_word_query": 23.3452, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4712.0458, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7715, "query_norm": 1.511, "queue_k_norm": 1.6077, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4447, "sent_len_1": 66.8695, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.6562, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71700 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.5252, "doc_norm": 1.611, "encoder_q-embeddings": 2576.4543, "encoder_q-layer.0": 1779.0717, "encoder_q-layer.1": 1811.9956, "encoder_q-layer.10": 3323.3679, "encoder_q-layer.11": 7602.499, "encoder_q-layer.2": 2028.8097, "encoder_q-layer.3": 2100.2922, "encoder_q-layer.4": 2208.613, "encoder_q-layer.5": 2212.7654, "encoder_q-layer.6": 2570.2019, "encoder_q-layer.7": 2964.9956, "encoder_q-layer.8": 3384.8076, "encoder_q-layer.9": 3116.7439, "epoch": 0.31, "inbatch_neg_score": 0.768, "inbatch_pos_score": 1.3691, "learning_rate": 3.373684210526316e-05, "loss": 3.5252, "norm_diff": 0.1015, "norm_loss": 0.0, "num_token_doc": 66.8827, "num_token_overlap": 11.6887, "num_token_query": 31.355, "num_token_union": 65.1523, "num_word_context": 202.256, "num_word_doc": 49.903, "num_word_query": 23.2909, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4764.6494, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7686, "query_norm": 1.5095, "queue_k_norm": 1.608, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.355, "sent_len_1": 66.8827, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.6813, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71800 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.5407, "doc_norm": 1.6096, "encoder_q-embeddings": 2425.8743, "encoder_q-layer.0": 1632.2058, "encoder_q-layer.1": 1675.8085, "encoder_q-layer.10": 2976.0679, "encoder_q-layer.11": 7113.854, "encoder_q-layer.2": 1861.6152, "encoder_q-layer.3": 1925.6095, "encoder_q-layer.4": 2016.864, "encoder_q-layer.5": 2116.7917, "encoder_q-layer.6": 2317.7024, "encoder_q-layer.7": 2713.7197, "encoder_q-layer.8": 3091.2034, "encoder_q-layer.9": 2912.9446, "epoch": 0.31, "inbatch_neg_score": 0.7705, "inbatch_pos_score": 1.3594, "learning_rate": 3.3710526315789475e-05, "loss": 3.5407, "norm_diff": 0.1029, "norm_loss": 0.0, "num_token_doc": 66.7616, "num_token_overlap": 11.6644, "num_token_query": 31.3384, "num_token_union": 65.1006, "num_word_context": 202.3578, "num_word_doc": 49.8037, "num_word_query": 23.2729, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4524.6125, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7695, "query_norm": 1.5067, "queue_k_norm": 1.6085, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3384, "sent_len_1": 66.7616, "sent_len_max_0": 127.9363, "sent_len_max_1": 210.5462, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 71900 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.5266, "doc_norm": 1.608, "encoder_q-embeddings": 2512.2065, "encoder_q-layer.0": 1679.6093, "encoder_q-layer.1": 1800.6038, "encoder_q-layer.10": 3422.8728, "encoder_q-layer.11": 7787.3896, "encoder_q-layer.2": 2012.8179, "encoder_q-layer.3": 2043.7622, "encoder_q-layer.4": 2107.7681, "encoder_q-layer.5": 2190.4375, "encoder_q-layer.6": 2432.4504, "encoder_q-layer.7": 2795.208, "encoder_q-layer.8": 3398.842, "encoder_q-layer.9": 3073.1221, "epoch": 0.31, "inbatch_neg_score": 0.7706, "inbatch_pos_score": 1.3711, "learning_rate": 3.368421052631579e-05, "loss": 3.5266, "norm_diff": 0.099, "norm_loss": 0.0, "num_token_doc": 66.7017, "num_token_overlap": 11.6759, "num_token_query": 31.318, "num_token_union": 65.0205, "num_word_context": 201.9405, "num_word_doc": 49.7439, "num_word_query": 23.2617, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4819.5163, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7715, "query_norm": 1.5089, "queue_k_norm": 1.6083, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.318, "sent_len_1": 66.7017, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.0913, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 72000 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.5226, "doc_norm": 1.6118, "encoder_q-embeddings": 2588.8833, "encoder_q-layer.0": 1736.019, "encoder_q-layer.1": 1809.5284, "encoder_q-layer.10": 3250.824, "encoder_q-layer.11": 7550.7881, "encoder_q-layer.2": 2087.7544, "encoder_q-layer.3": 2140.3257, "encoder_q-layer.4": 2242.3032, "encoder_q-layer.5": 2327.0005, "encoder_q-layer.6": 2586.6758, "encoder_q-layer.7": 2816.0984, "encoder_q-layer.8": 3345.2346, "encoder_q-layer.9": 3126.8945, "epoch": 0.31, "inbatch_neg_score": 0.7763, "inbatch_pos_score": 1.3574, "learning_rate": 3.3657894736842105e-05, "loss": 3.5226, "norm_diff": 0.1079, "norm_loss": 0.0, "num_token_doc": 66.7656, "num_token_overlap": 11.6744, "num_token_query": 31.3584, "num_token_union": 65.1068, "num_word_context": 202.3404, "num_word_doc": 49.8264, "num_word_query": 23.2861, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4833.3736, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7754, "query_norm": 1.5038, "queue_k_norm": 1.609, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3584, "sent_len_1": 66.7656, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.1887, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 72100 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.5229, "doc_norm": 1.6116, "encoder_q-embeddings": 2620.426, "encoder_q-layer.0": 1791.6637, "encoder_q-layer.1": 1911.1575, "encoder_q-layer.10": 3642.3706, "encoder_q-layer.11": 8019.1133, "encoder_q-layer.2": 2157.8943, "encoder_q-layer.3": 2186.5647, "encoder_q-layer.4": 2335.021, "encoder_q-layer.5": 2472.1362, "encoder_q-layer.6": 2754.0493, "encoder_q-layer.7": 3136.7986, "encoder_q-layer.8": 3746.9219, "encoder_q-layer.9": 3341.0203, "epoch": 0.31, "inbatch_neg_score": 0.7762, "inbatch_pos_score": 1.3799, "learning_rate": 3.363157894736842e-05, "loss": 3.5229, "norm_diff": 0.0845, "norm_loss": 0.0, "num_token_doc": 66.6923, "num_token_overlap": 11.7122, "num_token_query": 31.4146, "num_token_union": 65.0518, "num_word_context": 202.3146, "num_word_doc": 49.7684, "num_word_query": 23.3338, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5104.1244, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7778, "query_norm": 1.5271, "queue_k_norm": 1.6096, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4146, "sent_len_1": 66.6923, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.9087, "stdk": 0.049, "stdq": 0.0437, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 72200 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.5208, "doc_norm": 1.61, "encoder_q-embeddings": 2653.9409, "encoder_q-layer.0": 1797.0076, "encoder_q-layer.1": 1919.5984, "encoder_q-layer.10": 3410.0332, "encoder_q-layer.11": 7474.4736, "encoder_q-layer.2": 2177.1208, "encoder_q-layer.3": 2237.3611, "encoder_q-layer.4": 2455.3281, "encoder_q-layer.5": 2501.3735, "encoder_q-layer.6": 2608.3486, "encoder_q-layer.7": 2981.2463, "encoder_q-layer.8": 3603.2012, "encoder_q-layer.9": 3254.5283, "epoch": 0.31, "inbatch_neg_score": 0.7811, "inbatch_pos_score": 1.3691, "learning_rate": 3.360526315789474e-05, "loss": 3.5208, "norm_diff": 0.0962, "norm_loss": 0.0, "num_token_doc": 66.8439, "num_token_overlap": 11.6835, "num_token_query": 31.3513, "num_token_union": 65.1524, "num_word_context": 202.4179, "num_word_doc": 49.8988, "num_word_query": 23.2867, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4947.6257, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7798, "query_norm": 1.5138, "queue_k_norm": 1.6103, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3513, "sent_len_1": 66.8439, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.15, "stdk": 0.0488, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 72300 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5295, "doc_norm": 1.6171, "encoder_q-embeddings": 2616.0364, "encoder_q-layer.0": 1769.0988, "encoder_q-layer.1": 1839.3153, "encoder_q-layer.10": 3444.3325, "encoder_q-layer.11": 7698.7197, "encoder_q-layer.2": 2088.2109, "encoder_q-layer.3": 2162.0066, "encoder_q-layer.4": 2239.4868, "encoder_q-layer.5": 2211.2983, "encoder_q-layer.6": 2460.0713, "encoder_q-layer.7": 2949.5303, "encoder_q-layer.8": 3800.179, "encoder_q-layer.9": 3362.2354, "epoch": 0.31, "inbatch_neg_score": 0.7846, "inbatch_pos_score": 1.3848, "learning_rate": 3.357894736842105e-05, "loss": 3.5295, "norm_diff": 0.0913, "norm_loss": 0.0, "num_token_doc": 66.8706, "num_token_overlap": 11.653, "num_token_query": 31.329, "num_token_union": 65.1728, "num_word_context": 202.415, "num_word_doc": 49.8714, "num_word_query": 23.2577, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4911.0956, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7842, "query_norm": 1.5258, "queue_k_norm": 1.6141, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.329, "sent_len_1": 66.8706, "sent_len_max_0": 127.9838, "sent_len_max_1": 211.0137, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 72400 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5207, "doc_norm": 1.6127, "encoder_q-embeddings": 2524.1265, "encoder_q-layer.0": 1719.2982, "encoder_q-layer.1": 1850.224, "encoder_q-layer.10": 3345.8022, "encoder_q-layer.11": 7876.5942, "encoder_q-layer.2": 1990.0348, "encoder_q-layer.3": 2050.4001, "encoder_q-layer.4": 2171.9849, "encoder_q-layer.5": 2249.7998, "encoder_q-layer.6": 2567.2344, "encoder_q-layer.7": 2922.8879, "encoder_q-layer.8": 3596.1074, "encoder_q-layer.9": 3223.0347, "epoch": 0.31, "inbatch_neg_score": 0.7873, "inbatch_pos_score": 1.3857, "learning_rate": 3.355263157894737e-05, "loss": 3.5207, "norm_diff": 0.0883, "norm_loss": 0.0, "num_token_doc": 66.7653, "num_token_overlap": 11.7074, "num_token_query": 31.3768, "num_token_union": 65.1009, "num_word_context": 202.403, "num_word_doc": 49.808, "num_word_query": 23.3063, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4932.8956, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7871, "query_norm": 1.5244, "queue_k_norm": 1.6137, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3768, "sent_len_1": 66.7653, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.7925, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 72500 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.519, "doc_norm": 1.617, "encoder_q-embeddings": 2635.0564, "encoder_q-layer.0": 1796.6086, "encoder_q-layer.1": 1894.1273, "encoder_q-layer.10": 3274.6895, "encoder_q-layer.11": 7551.0298, "encoder_q-layer.2": 2137.9014, "encoder_q-layer.3": 2156.0779, "encoder_q-layer.4": 2228.1038, "encoder_q-layer.5": 2246.9919, "encoder_q-layer.6": 2483.7437, "encoder_q-layer.7": 3073.0415, "encoder_q-layer.8": 3461.187, "encoder_q-layer.9": 3050.9539, "epoch": 0.32, "inbatch_neg_score": 0.7914, "inbatch_pos_score": 1.3857, "learning_rate": 3.352631578947368e-05, "loss": 3.519, "norm_diff": 0.0844, "norm_loss": 0.0, "num_token_doc": 66.952, "num_token_overlap": 11.7207, "num_token_query": 31.3791, "num_token_union": 65.1659, "num_word_context": 202.4107, "num_word_doc": 49.9201, "num_word_query": 23.2974, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4802.9735, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.7915, "query_norm": 1.5326, "queue_k_norm": 1.6134, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3791, "sent_len_1": 66.952, "sent_len_max_0": 127.9862, "sent_len_max_1": 211.4275, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 72600 }, { "accuracy": 46.582, "active_queue_size": 16384.0, "cl_loss": 3.5338, "doc_norm": 1.6135, "encoder_q-embeddings": 2724.6099, "encoder_q-layer.0": 1886.0471, "encoder_q-layer.1": 1971.0118, "encoder_q-layer.10": 3467.6509, "encoder_q-layer.11": 7817.2168, "encoder_q-layer.2": 2155.8315, "encoder_q-layer.3": 2208.1941, "encoder_q-layer.4": 2360.5129, "encoder_q-layer.5": 2361.4836, "encoder_q-layer.6": 2631.8027, "encoder_q-layer.7": 2951.8403, "encoder_q-layer.8": 3423.9011, "encoder_q-layer.9": 3230.9978, "epoch": 0.32, "inbatch_neg_score": 0.8042, "inbatch_pos_score": 1.3779, "learning_rate": 3.35e-05, "loss": 3.5338, "norm_diff": 0.0685, "norm_loss": 0.0, "num_token_doc": 66.9898, "num_token_overlap": 11.6549, "num_token_query": 31.3474, "num_token_union": 65.2394, "num_word_context": 202.6262, "num_word_doc": 49.9639, "num_word_query": 23.2732, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5003.3384, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.8037, "query_norm": 1.545, "queue_k_norm": 1.6159, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3474, "sent_len_1": 66.9898, "sent_len_max_0": 127.975, "sent_len_max_1": 209.5637, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 72700 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.5393, "doc_norm": 1.6171, "encoder_q-embeddings": 3603.2585, "encoder_q-layer.0": 2491.2458, "encoder_q-layer.1": 2872.4324, "encoder_q-layer.10": 3439.2197, "encoder_q-layer.11": 7361.6753, "encoder_q-layer.2": 3162.7119, "encoder_q-layer.3": 3351.8037, "encoder_q-layer.4": 3688.4404, "encoder_q-layer.5": 3589.2209, "encoder_q-layer.6": 3459.4885, "encoder_q-layer.7": 3325.2866, "encoder_q-layer.8": 3610.4565, "encoder_q-layer.9": 3302.7439, "epoch": 0.32, "inbatch_neg_score": 0.8147, "inbatch_pos_score": 1.4033, "learning_rate": 3.347368421052632e-05, "loss": 3.5393, "norm_diff": 0.0567, "norm_loss": 0.0, "num_token_doc": 66.8482, "num_token_overlap": 11.6482, "num_token_query": 31.2669, "num_token_union": 65.153, "num_word_context": 202.4164, "num_word_doc": 49.906, "num_word_query": 23.2127, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5595.2591, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.814, "query_norm": 1.5604, "queue_k_norm": 1.6162, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2669, "sent_len_1": 66.8482, "sent_len_max_0": 127.9825, "sent_len_max_1": 206.71, "stdk": 0.049, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 72800 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5422, "doc_norm": 1.6203, "encoder_q-embeddings": 2480.0918, "encoder_q-layer.0": 1703.9302, "encoder_q-layer.1": 1784.4369, "encoder_q-layer.10": 3243.7041, "encoder_q-layer.11": 7401.311, "encoder_q-layer.2": 2005.7311, "encoder_q-layer.3": 2103.0154, "encoder_q-layer.4": 2157.6345, "encoder_q-layer.5": 2195.335, "encoder_q-layer.6": 2407.1833, "encoder_q-layer.7": 2999.4854, "encoder_q-layer.8": 3363.147, "encoder_q-layer.9": 3159.9915, "epoch": 0.32, "inbatch_neg_score": 0.8301, "inbatch_pos_score": 1.4268, "learning_rate": 3.344736842105263e-05, "loss": 3.5422, "norm_diff": 0.0442, "norm_loss": 0.0, "num_token_doc": 66.6054, "num_token_overlap": 11.6101, "num_token_query": 31.2428, "num_token_union": 65.0235, "num_word_context": 202.277, "num_word_doc": 49.7233, "num_word_query": 23.2108, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4758.5431, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.8296, "query_norm": 1.5761, "queue_k_norm": 1.6174, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2428, "sent_len_1": 66.6054, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.1438, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 72900 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.516, "doc_norm": 1.6228, "encoder_q-embeddings": 6580.1284, "encoder_q-layer.0": 4733.1875, "encoder_q-layer.1": 5530.0283, "encoder_q-layer.10": 3622.2063, "encoder_q-layer.11": 7901.5391, "encoder_q-layer.2": 6396.1987, "encoder_q-layer.3": 6823.2612, "encoder_q-layer.4": 8071.5483, "encoder_q-layer.5": 8226.1689, "encoder_q-layer.6": 7869.6831, "encoder_q-layer.7": 7142.5034, "encoder_q-layer.8": 6388.0591, "encoder_q-layer.9": 4032.863, "epoch": 0.32, "inbatch_neg_score": 0.8448, "inbatch_pos_score": 1.4424, "learning_rate": 3.342105263157895e-05, "loss": 3.516, "norm_diff": 0.051, "norm_loss": 0.0, "num_token_doc": 66.6202, "num_token_overlap": 11.7136, "num_token_query": 31.5251, "num_token_union": 65.0864, "num_word_context": 202.209, "num_word_doc": 49.7273, "num_word_query": 23.4172, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9605.8649, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8438, "query_norm": 1.5718, "queue_k_norm": 1.6175, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5251, "sent_len_1": 66.6202, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.4062, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 73000 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.5176, "doc_norm": 1.6228, "encoder_q-embeddings": 2593.6501, "encoder_q-layer.0": 1672.719, "encoder_q-layer.1": 1792.8156, "encoder_q-layer.10": 3278.5007, "encoder_q-layer.11": 7623.9307, "encoder_q-layer.2": 1965.797, "encoder_q-layer.3": 2067.5605, "encoder_q-layer.4": 2174.1921, "encoder_q-layer.5": 2299.2363, "encoder_q-layer.6": 2420.98, "encoder_q-layer.7": 3027.1489, "encoder_q-layer.8": 3442.8625, "encoder_q-layer.9": 3222.2559, "epoch": 0.32, "inbatch_neg_score": 0.8553, "inbatch_pos_score": 1.4551, "learning_rate": 3.339473684210526e-05, "loss": 3.5176, "norm_diff": 0.0477, "norm_loss": 0.0, "num_token_doc": 66.7569, "num_token_overlap": 11.6985, "num_token_query": 31.4526, "num_token_union": 65.1936, "num_word_context": 202.1274, "num_word_doc": 49.8476, "num_word_query": 23.3693, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4798.4135, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 0.8555, "query_norm": 1.5751, "queue_k_norm": 1.6214, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4526, "sent_len_1": 66.7569, "sent_len_max_0": 127.975, "sent_len_max_1": 205.735, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73100 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.5083, "doc_norm": 1.6232, "encoder_q-embeddings": 4916.5898, "encoder_q-layer.0": 3363.2092, "encoder_q-layer.1": 3544.8174, "encoder_q-layer.10": 7165.5259, "encoder_q-layer.11": 15643.25, "encoder_q-layer.2": 3939.7859, "encoder_q-layer.3": 3949.925, "encoder_q-layer.4": 4173.6001, "encoder_q-layer.5": 4211.9487, "encoder_q-layer.6": 4767.5718, "encoder_q-layer.7": 5790.1436, "encoder_q-layer.8": 6777.6094, "encoder_q-layer.9": 6438.5239, "epoch": 0.32, "inbatch_neg_score": 0.8584, "inbatch_pos_score": 1.4375, "learning_rate": 3.336842105263158e-05, "loss": 3.5083, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 67.042, "num_token_overlap": 11.7106, "num_token_query": 31.3659, "num_token_union": 65.2224, "num_word_context": 202.481, "num_word_doc": 50.0057, "num_word_query": 23.2786, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9413.0582, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8584, "query_norm": 1.5462, "queue_k_norm": 1.6229, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3659, "sent_len_1": 67.042, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.7512, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73200 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.5328, "doc_norm": 1.6239, "encoder_q-embeddings": 5257.1943, "encoder_q-layer.0": 3549.1672, "encoder_q-layer.1": 3829.147, "encoder_q-layer.10": 7165.4326, "encoder_q-layer.11": 16187.7754, "encoder_q-layer.2": 4300.1357, "encoder_q-layer.3": 4399.5156, "encoder_q-layer.4": 4517.0669, "encoder_q-layer.5": 4552.0757, "encoder_q-layer.6": 5074.1274, "encoder_q-layer.7": 5838.4126, "encoder_q-layer.8": 7322.6455, "encoder_q-layer.9": 6977.9375, "epoch": 0.32, "inbatch_neg_score": 0.8647, "inbatch_pos_score": 1.4443, "learning_rate": 3.33421052631579e-05, "loss": 3.5328, "norm_diff": 0.0856, "norm_loss": 0.0, "num_token_doc": 66.6855, "num_token_overlap": 11.6732, "num_token_query": 31.4075, "num_token_union": 65.0913, "num_word_context": 202.4218, "num_word_doc": 49.7519, "num_word_query": 23.3363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10127.6183, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8643, "query_norm": 1.5383, "queue_k_norm": 1.6251, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4075, "sent_len_1": 66.6855, "sent_len_max_0": 128.0, "sent_len_max_1": 209.425, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73300 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.5157, "doc_norm": 1.6273, "encoder_q-embeddings": 9962.9473, "encoder_q-layer.0": 6999.499, "encoder_q-layer.1": 8266.0732, "encoder_q-layer.10": 6473.249, "encoder_q-layer.11": 15305.9805, "encoder_q-layer.2": 8629.6162, "encoder_q-layer.3": 8230.6846, "encoder_q-layer.4": 6963.5674, "encoder_q-layer.5": 6845.9038, "encoder_q-layer.6": 7026.1587, "encoder_q-layer.7": 7772.6992, "encoder_q-layer.8": 8464.0225, "encoder_q-layer.9": 6734.4014, "epoch": 0.32, "inbatch_neg_score": 0.864, "inbatch_pos_score": 1.458, "learning_rate": 3.331578947368421e-05, "loss": 3.5157, "norm_diff": 0.0899, "norm_loss": 0.0, "num_token_doc": 66.8589, "num_token_overlap": 11.6717, "num_token_query": 31.2959, "num_token_union": 65.1289, "num_word_context": 202.3565, "num_word_doc": 49.8803, "num_word_query": 23.2285, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13142.7888, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8628, "query_norm": 1.5375, "queue_k_norm": 1.6278, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2959, "sent_len_1": 66.8589, "sent_len_max_0": 127.975, "sent_len_max_1": 210.2237, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73400 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.5351, "doc_norm": 1.6257, "encoder_q-embeddings": 4885.4507, "encoder_q-layer.0": 3308.8713, "encoder_q-layer.1": 3608.146, "encoder_q-layer.10": 6218.6406, "encoder_q-layer.11": 13956.8691, "encoder_q-layer.2": 4166.8223, "encoder_q-layer.3": 4174.0928, "encoder_q-layer.4": 4579.3472, "encoder_q-layer.5": 4657.5659, "encoder_q-layer.6": 5075.9419, "encoder_q-layer.7": 5691.5723, "encoder_q-layer.8": 6637.644, "encoder_q-layer.9": 6223.1909, "epoch": 0.32, "inbatch_neg_score": 0.8667, "inbatch_pos_score": 1.4521, "learning_rate": 3.328947368421053e-05, "loss": 3.5351, "norm_diff": 0.0863, "norm_loss": 0.0, "num_token_doc": 66.9013, "num_token_overlap": 11.6928, "num_token_query": 31.4437, "num_token_union": 65.2552, "num_word_context": 202.5913, "num_word_doc": 49.939, "num_word_query": 23.3422, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9269.4703, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8662, "query_norm": 1.5394, "queue_k_norm": 1.6299, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4437, "sent_len_1": 66.9013, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.265, "stdk": 0.0488, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73500 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.5328, "doc_norm": 1.6301, "encoder_q-embeddings": 5941.5059, "encoder_q-layer.0": 4072.4363, "encoder_q-layer.1": 4578.1382, "encoder_q-layer.10": 6657.7231, "encoder_q-layer.11": 16187.9082, "encoder_q-layer.2": 4916.938, "encoder_q-layer.3": 4998.7993, "encoder_q-layer.4": 5459.9985, "encoder_q-layer.5": 5581.4302, "encoder_q-layer.6": 5885.4175, "encoder_q-layer.7": 6556.748, "encoder_q-layer.8": 7359.9194, "encoder_q-layer.9": 6425.1406, "epoch": 0.32, "inbatch_neg_score": 0.8617, "inbatch_pos_score": 1.4531, "learning_rate": 3.3263157894736846e-05, "loss": 3.5328, "norm_diff": 0.0961, "norm_loss": 0.0, "num_token_doc": 66.5111, "num_token_overlap": 11.673, "num_token_query": 31.355, "num_token_union": 64.9453, "num_word_context": 201.8849, "num_word_doc": 49.6384, "num_word_query": 23.284, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10573.8185, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8623, "query_norm": 1.534, "queue_k_norm": 1.6297, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.355, "sent_len_1": 66.5111, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.12, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73600 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5246, "doc_norm": 1.6307, "encoder_q-embeddings": 5333.9829, "encoder_q-layer.0": 3544.3816, "encoder_q-layer.1": 3761.3496, "encoder_q-layer.10": 6897.0942, "encoder_q-layer.11": 15638.2646, "encoder_q-layer.2": 4146.0542, "encoder_q-layer.3": 4330.5581, "encoder_q-layer.4": 4497.0073, "encoder_q-layer.5": 4743.71, "encoder_q-layer.6": 5132.7373, "encoder_q-layer.7": 5623.8579, "encoder_q-layer.8": 6936.2969, "encoder_q-layer.9": 6194.9077, "epoch": 0.32, "inbatch_neg_score": 0.8664, "inbatch_pos_score": 1.4658, "learning_rate": 3.323684210526316e-05, "loss": 3.5246, "norm_diff": 0.0914, "norm_loss": 0.0, "num_token_doc": 66.8748, "num_token_overlap": 11.6961, "num_token_query": 31.3991, "num_token_union": 65.1663, "num_word_context": 202.5575, "num_word_doc": 49.8802, "num_word_query": 23.3293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10073.5366, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8672, "query_norm": 1.5393, "queue_k_norm": 1.6326, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3991, "sent_len_1": 66.8748, "sent_len_max_0": 127.995, "sent_len_max_1": 207.7438, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73700 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.5384, "doc_norm": 1.6335, "encoder_q-embeddings": 6011.0801, "encoder_q-layer.0": 4126.6768, "encoder_q-layer.1": 4285.5898, "encoder_q-layer.10": 7464.3828, "encoder_q-layer.11": 15202.6973, "encoder_q-layer.2": 4660.2856, "encoder_q-layer.3": 4745.7329, "encoder_q-layer.4": 5065.9414, "encoder_q-layer.5": 5210.1699, "encoder_q-layer.6": 5508.2163, "encoder_q-layer.7": 6307.6401, "encoder_q-layer.8": 7753.3545, "encoder_q-layer.9": 7260.8125, "epoch": 0.32, "inbatch_neg_score": 0.8689, "inbatch_pos_score": 1.46, "learning_rate": 3.3210526315789476e-05, "loss": 3.5384, "norm_diff": 0.0844, "norm_loss": 0.0, "num_token_doc": 66.6332, "num_token_overlap": 11.623, "num_token_query": 31.288, "num_token_union": 65.0545, "num_word_context": 202.3585, "num_word_doc": 49.7376, "num_word_query": 23.2475, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10353.6843, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8687, "query_norm": 1.549, "queue_k_norm": 1.6332, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.288, "sent_len_1": 66.6332, "sent_len_max_0": 127.98, "sent_len_max_1": 207.1362, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73800 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.5241, "doc_norm": 1.6358, "encoder_q-embeddings": 5473.8003, "encoder_q-layer.0": 3767.6646, "encoder_q-layer.1": 4070.4546, "encoder_q-layer.10": 6738.7915, "encoder_q-layer.11": 15241.751, "encoder_q-layer.2": 4490.6211, "encoder_q-layer.3": 4420.5044, "encoder_q-layer.4": 4687.8901, "encoder_q-layer.5": 4802.8804, "encoder_q-layer.6": 5482.2354, "encoder_q-layer.7": 5950.5474, "encoder_q-layer.8": 6949.1997, "encoder_q-layer.9": 6296.9976, "epoch": 0.32, "inbatch_neg_score": 0.874, "inbatch_pos_score": 1.4727, "learning_rate": 3.318421052631579e-05, "loss": 3.5241, "norm_diff": 0.0878, "norm_loss": 0.0, "num_token_doc": 66.7557, "num_token_overlap": 11.6657, "num_token_query": 31.3583, "num_token_union": 65.0966, "num_word_context": 202.2967, "num_word_doc": 49.7887, "num_word_query": 23.2806, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9960.5113, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8745, "query_norm": 1.548, "queue_k_norm": 1.6347, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3583, "sent_len_1": 66.7557, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.8575, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 73900 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.5198, "doc_norm": 1.6348, "encoder_q-embeddings": 5864.6694, "encoder_q-layer.0": 4160.4629, "encoder_q-layer.1": 4573.0518, "encoder_q-layer.10": 6177.793, "encoder_q-layer.11": 14732.3359, "encoder_q-layer.2": 5052.769, "encoder_q-layer.3": 5330.7832, "encoder_q-layer.4": 5850.5674, "encoder_q-layer.5": 6281.1421, "encoder_q-layer.6": 6244.4907, "encoder_q-layer.7": 6758.4785, "encoder_q-layer.8": 7152.1533, "encoder_q-layer.9": 6291.4688, "epoch": 0.32, "inbatch_neg_score": 0.8755, "inbatch_pos_score": 1.4688, "learning_rate": 3.3157894736842106e-05, "loss": 3.5198, "norm_diff": 0.0933, "norm_loss": 0.0, "num_token_doc": 66.7077, "num_token_overlap": 11.6701, "num_token_query": 31.3558, "num_token_union": 65.0536, "num_word_context": 201.9532, "num_word_doc": 49.7534, "num_word_query": 23.2999, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10460.0057, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.875, "query_norm": 1.5416, "queue_k_norm": 1.6351, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3558, "sent_len_1": 66.7077, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.6012, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74000 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.5207, "doc_norm": 1.6359, "encoder_q-embeddings": 5196.3838, "encoder_q-layer.0": 3473.0374, "encoder_q-layer.1": 3817.0925, "encoder_q-layer.10": 6379.3618, "encoder_q-layer.11": 14808.5322, "encoder_q-layer.2": 4171.3911, "encoder_q-layer.3": 4357.835, "encoder_q-layer.4": 4669.0898, "encoder_q-layer.5": 4855.001, "encoder_q-layer.6": 5360.0312, "encoder_q-layer.7": 6066.4932, "encoder_q-layer.8": 6635.0825, "encoder_q-layer.9": 5881.3789, "epoch": 0.32, "inbatch_neg_score": 0.8754, "inbatch_pos_score": 1.4766, "learning_rate": 3.3131578947368424e-05, "loss": 3.5207, "norm_diff": 0.0967, "norm_loss": 0.0, "num_token_doc": 66.6892, "num_token_overlap": 11.6914, "num_token_query": 31.372, "num_token_union": 65.065, "num_word_context": 201.8935, "num_word_doc": 49.7437, "num_word_query": 23.3104, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9666.7172, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.876, "query_norm": 1.5392, "queue_k_norm": 1.6378, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.372, "sent_len_1": 66.6892, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.2438, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74100 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.5289, "doc_norm": 1.6436, "encoder_q-embeddings": 5754.4204, "encoder_q-layer.0": 4144.8062, "encoder_q-layer.1": 4298.8032, "encoder_q-layer.10": 7179.2495, "encoder_q-layer.11": 15229.7373, "encoder_q-layer.2": 4860.8833, "encoder_q-layer.3": 5153.6104, "encoder_q-layer.4": 5698.5581, "encoder_q-layer.5": 5132.5386, "encoder_q-layer.6": 5631.9692, "encoder_q-layer.7": 6222.4248, "encoder_q-layer.8": 6877.4248, "encoder_q-layer.9": 6682.4019, "epoch": 0.32, "inbatch_neg_score": 0.8773, "inbatch_pos_score": 1.4668, "learning_rate": 3.3105263157894736e-05, "loss": 3.5289, "norm_diff": 0.1219, "norm_loss": 0.0, "num_token_doc": 66.8022, "num_token_overlap": 11.6227, "num_token_query": 31.2839, "num_token_union": 65.1432, "num_word_context": 202.4074, "num_word_doc": 49.8696, "num_word_query": 23.2291, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10157.399, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.877, "query_norm": 1.5217, "queue_k_norm": 1.6372, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2839, "sent_len_1": 66.8022, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.4487, "stdk": 0.0492, "stdq": 0.0422, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74200 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.5241, "doc_norm": 1.6433, "encoder_q-embeddings": 5161.1011, "encoder_q-layer.0": 3496.6331, "encoder_q-layer.1": 3702.9636, "encoder_q-layer.10": 6307.3887, "encoder_q-layer.11": 14724.834, "encoder_q-layer.2": 4105.0342, "encoder_q-layer.3": 4244.6201, "encoder_q-layer.4": 4546.5859, "encoder_q-layer.5": 4700.0659, "encoder_q-layer.6": 5230.605, "encoder_q-layer.7": 5849.5771, "encoder_q-layer.8": 7086.5366, "encoder_q-layer.9": 6179.8354, "epoch": 0.32, "inbatch_neg_score": 0.8782, "inbatch_pos_score": 1.4766, "learning_rate": 3.3078947368421054e-05, "loss": 3.5241, "norm_diff": 0.0977, "norm_loss": 0.0, "num_token_doc": 66.9528, "num_token_overlap": 11.689, "num_token_query": 31.3166, "num_token_union": 65.1899, "num_word_context": 202.3436, "num_word_doc": 49.9698, "num_word_query": 23.2574, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9678.7087, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8779, "query_norm": 1.5456, "queue_k_norm": 1.6398, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3166, "sent_len_1": 66.9528, "sent_len_max_0": 127.9737, "sent_len_max_1": 207.645, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74300 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.519, "doc_norm": 1.6383, "encoder_q-embeddings": 5023.3115, "encoder_q-layer.0": 3430.7957, "encoder_q-layer.1": 3647.8496, "encoder_q-layer.10": 6611.2163, "encoder_q-layer.11": 14740.2627, "encoder_q-layer.2": 4043.3813, "encoder_q-layer.3": 4154.1655, "encoder_q-layer.4": 4462.2783, "encoder_q-layer.5": 4613.6807, "encoder_q-layer.6": 5153.0259, "encoder_q-layer.7": 5958.1504, "encoder_q-layer.8": 6788.6919, "encoder_q-layer.9": 6178.6729, "epoch": 0.32, "inbatch_neg_score": 0.8828, "inbatch_pos_score": 1.4746, "learning_rate": 3.3052631578947366e-05, "loss": 3.519, "norm_diff": 0.0955, "norm_loss": 0.0, "num_token_doc": 66.7543, "num_token_overlap": 11.6592, "num_token_query": 31.3308, "num_token_union": 65.09, "num_word_context": 202.2429, "num_word_doc": 49.8134, "num_word_query": 23.2722, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9494.1768, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8823, "query_norm": 1.5428, "queue_k_norm": 1.6395, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3308, "sent_len_1": 66.7543, "sent_len_max_0": 127.985, "sent_len_max_1": 207.7325, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74400 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.5112, "doc_norm": 1.6429, "encoder_q-embeddings": 16210.9463, "encoder_q-layer.0": 12435.4346, "encoder_q-layer.1": 14671.9883, "encoder_q-layer.10": 6680.7056, "encoder_q-layer.11": 16036.5742, "encoder_q-layer.2": 16315.3232, "encoder_q-layer.3": 13895.4775, "encoder_q-layer.4": 14894.5381, "encoder_q-layer.5": 15643.1953, "encoder_q-layer.6": 13672.3945, "encoder_q-layer.7": 11445.502, "encoder_q-layer.8": 8804.3477, "encoder_q-layer.9": 6709.2676, "epoch": 0.32, "inbatch_neg_score": 0.8898, "inbatch_pos_score": 1.4873, "learning_rate": 3.302631578947369e-05, "loss": 3.5112, "norm_diff": 0.0901, "norm_loss": 0.0, "num_token_doc": 66.6521, "num_token_overlap": 11.6858, "num_token_query": 31.3857, "num_token_union": 65.0474, "num_word_context": 202.1752, "num_word_doc": 49.7441, "num_word_query": 23.3204, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20307.8666, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.8892, "query_norm": 1.5528, "queue_k_norm": 1.6399, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3857, "sent_len_1": 66.6521, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.6, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74500 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.5264, "doc_norm": 1.6436, "encoder_q-embeddings": 7433.0845, "encoder_q-layer.0": 5200.1299, "encoder_q-layer.1": 5927.3145, "encoder_q-layer.10": 6608.311, "encoder_q-layer.11": 15042.7607, "encoder_q-layer.2": 7046.7549, "encoder_q-layer.3": 7641.9683, "encoder_q-layer.4": 8050.6743, "encoder_q-layer.5": 7936.1841, "encoder_q-layer.6": 7543.5889, "encoder_q-layer.7": 7438.1987, "encoder_q-layer.8": 7400.5049, "encoder_q-layer.9": 6234.7217, "epoch": 0.32, "inbatch_neg_score": 0.885, "inbatch_pos_score": 1.4756, "learning_rate": 3.3e-05, "loss": 3.5264, "norm_diff": 0.1052, "norm_loss": 0.0, "num_token_doc": 66.7688, "num_token_overlap": 11.6677, "num_token_query": 31.3828, "num_token_union": 65.152, "num_word_context": 202.4986, "num_word_doc": 49.847, "num_word_query": 23.3132, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11845.8723, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8853, "query_norm": 1.5384, "queue_k_norm": 1.641, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3828, "sent_len_1": 66.7688, "sent_len_max_0": 127.975, "sent_len_max_1": 207.77, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74600 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.5145, "doc_norm": 1.6429, "encoder_q-embeddings": 5681.498, "encoder_q-layer.0": 4073.7659, "encoder_q-layer.1": 4715.4111, "encoder_q-layer.10": 6504.5781, "encoder_q-layer.11": 14981.8086, "encoder_q-layer.2": 5389.6094, "encoder_q-layer.3": 5294.3911, "encoder_q-layer.4": 5819.7349, "encoder_q-layer.5": 5722.3027, "encoder_q-layer.6": 5678.4707, "encoder_q-layer.7": 6124.1826, "encoder_q-layer.8": 6866.1255, "encoder_q-layer.9": 5958.6787, "epoch": 0.32, "inbatch_neg_score": 0.8844, "inbatch_pos_score": 1.4902, "learning_rate": 3.297368421052632e-05, "loss": 3.5145, "norm_diff": 0.1077, "norm_loss": 0.0, "num_token_doc": 66.8244, "num_token_overlap": 11.6742, "num_token_query": 31.4098, "num_token_union": 65.1715, "num_word_context": 202.2319, "num_word_doc": 49.8329, "num_word_query": 23.3196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10247.269, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8848, "query_norm": 1.5352, "queue_k_norm": 1.6416, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4098, "sent_len_1": 66.8244, "sent_len_max_0": 127.97, "sent_len_max_1": 210.2025, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 74700 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.5134, "doc_norm": 1.6484, "encoder_q-embeddings": 4928.7251, "encoder_q-layer.0": 3302.7373, "encoder_q-layer.1": 3553.231, "encoder_q-layer.10": 6082.0645, "encoder_q-layer.11": 14539.0889, "encoder_q-layer.2": 3899.2454, "encoder_q-layer.3": 3929.8784, "encoder_q-layer.4": 4159.4805, "encoder_q-layer.5": 4354.4097, "encoder_q-layer.6": 4921.8442, "encoder_q-layer.7": 5363.8848, "encoder_q-layer.8": 6436.5659, "encoder_q-layer.9": 5889.8862, "epoch": 0.32, "inbatch_neg_score": 0.8873, "inbatch_pos_score": 1.4922, "learning_rate": 3.294736842105263e-05, "loss": 3.5134, "norm_diff": 0.1006, "norm_loss": 0.0, "num_token_doc": 66.8753, "num_token_overlap": 11.6518, "num_token_query": 31.325, "num_token_union": 65.2023, "num_word_context": 202.419, "num_word_doc": 49.8929, "num_word_query": 23.2611, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9292.4398, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8867, "query_norm": 1.5478, "queue_k_norm": 1.6445, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.325, "sent_len_1": 66.8753, "sent_len_max_0": 127.975, "sent_len_max_1": 207.735, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74800 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.5162, "doc_norm": 1.6473, "encoder_q-embeddings": 5455.9248, "encoder_q-layer.0": 3627.6201, "encoder_q-layer.1": 3915.8633, "encoder_q-layer.10": 6096.3384, "encoder_q-layer.11": 14415.1855, "encoder_q-layer.2": 4300.7222, "encoder_q-layer.3": 4491.4395, "encoder_q-layer.4": 4661.5205, "encoder_q-layer.5": 4645.2881, "encoder_q-layer.6": 5194.7954, "encoder_q-layer.7": 5968.9927, "encoder_q-layer.8": 6771.2002, "encoder_q-layer.9": 6014.3682, "epoch": 0.33, "inbatch_neg_score": 0.888, "inbatch_pos_score": 1.4873, "learning_rate": 3.2921052631578944e-05, "loss": 3.5162, "norm_diff": 0.1059, "norm_loss": 0.0, "num_token_doc": 66.7976, "num_token_overlap": 11.6831, "num_token_query": 31.3521, "num_token_union": 65.1045, "num_word_context": 202.3141, "num_word_doc": 49.8209, "num_word_query": 23.2777, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9520.4393, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8887, "query_norm": 1.5413, "queue_k_norm": 1.6454, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3521, "sent_len_1": 66.7976, "sent_len_max_0": 127.995, "sent_len_max_1": 208.295, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 74900 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.5305, "doc_norm": 1.6462, "encoder_q-embeddings": 5037.9644, "encoder_q-layer.0": 3492.8013, "encoder_q-layer.1": 3689.5117, "encoder_q-layer.10": 6169.2065, "encoder_q-layer.11": 15208.9697, "encoder_q-layer.2": 4083.4578, "encoder_q-layer.3": 4192.4419, "encoder_q-layer.4": 4466.0273, "encoder_q-layer.5": 4454.1626, "encoder_q-layer.6": 4816.7417, "encoder_q-layer.7": 5471.9585, "encoder_q-layer.8": 6399.6479, "encoder_q-layer.9": 6052.4053, "epoch": 0.33, "inbatch_neg_score": 0.8902, "inbatch_pos_score": 1.4727, "learning_rate": 3.289473684210527e-05, "loss": 3.5305, "norm_diff": 0.1202, "norm_loss": 0.0, "num_token_doc": 66.9089, "num_token_overlap": 11.6606, "num_token_query": 31.3375, "num_token_union": 65.1879, "num_word_context": 202.6517, "num_word_doc": 49.924, "num_word_query": 23.255, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9629.0901, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8896, "query_norm": 1.526, "queue_k_norm": 1.648, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3375, "sent_len_1": 66.9089, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.9313, "stdk": 0.049, "stdq": 0.0425, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 75000 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.5056, "doc_norm": 1.6446, "encoder_q-embeddings": 5634.6523, "encoder_q-layer.0": 3795.5164, "encoder_q-layer.1": 4177.375, "encoder_q-layer.10": 7391.7974, "encoder_q-layer.11": 15761.9258, "encoder_q-layer.2": 4566.5889, "encoder_q-layer.3": 4813.6021, "encoder_q-layer.4": 5145.6836, "encoder_q-layer.5": 5194.1392, "encoder_q-layer.6": 5486.4541, "encoder_q-layer.7": 6196.0288, "encoder_q-layer.8": 7268.2173, "encoder_q-layer.9": 6738.959, "epoch": 0.33, "inbatch_neg_score": 0.8889, "inbatch_pos_score": 1.4756, "learning_rate": 3.286842105263158e-05, "loss": 3.5056, "norm_diff": 0.1227, "norm_loss": 0.0, "num_token_doc": 66.7803, "num_token_overlap": 11.6923, "num_token_query": 31.3741, "num_token_union": 65.1267, "num_word_context": 202.2791, "num_word_doc": 49.8261, "num_word_query": 23.3079, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10313.6723, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8887, "query_norm": 1.522, "queue_k_norm": 1.645, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3741, "sent_len_1": 66.7803, "sent_len_max_0": 127.97, "sent_len_max_1": 209.4837, "stdk": 0.0489, "stdq": 0.0423, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75100 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.5271, "doc_norm": 1.6457, "encoder_q-embeddings": 22667.7324, "encoder_q-layer.0": 16617.9512, "encoder_q-layer.1": 19230.1055, "encoder_q-layer.10": 12758.832, "encoder_q-layer.11": 30175.6406, "encoder_q-layer.2": 23355.5117, "encoder_q-layer.3": 24483.0762, "encoder_q-layer.4": 26600.0371, "encoder_q-layer.5": 25887.0879, "encoder_q-layer.6": 26275.3027, "encoder_q-layer.7": 21434.2324, "encoder_q-layer.8": 17102.1426, "encoder_q-layer.9": 12946.2051, "epoch": 0.33, "inbatch_neg_score": 0.8899, "inbatch_pos_score": 1.4805, "learning_rate": 3.28421052631579e-05, "loss": 3.5271, "norm_diff": 0.1108, "norm_loss": 0.0, "num_token_doc": 66.6798, "num_token_overlap": 11.6779, "num_token_query": 31.3592, "num_token_union": 65.0644, "num_word_context": 202.4652, "num_word_doc": 49.7662, "num_word_query": 23.3031, "postclip_grad_norm": 1.0, "preclip_grad_norm": 32910.3803, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 0.8896, "query_norm": 1.5349, "queue_k_norm": 1.648, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3592, "sent_len_1": 66.6798, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.2988, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 75200 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.5152, "doc_norm": 1.6487, "encoder_q-embeddings": 13401.5215, "encoder_q-layer.0": 9406.0293, "encoder_q-layer.1": 10582.6865, "encoder_q-layer.10": 13550.1865, "encoder_q-layer.11": 30701.1094, "encoder_q-layer.2": 11007.4541, "encoder_q-layer.3": 11754.6152, "encoder_q-layer.4": 12181.4443, "encoder_q-layer.5": 12350.2598, "encoder_q-layer.6": 11855.5059, "encoder_q-layer.7": 12023.21, "encoder_q-layer.8": 14026.9492, "encoder_q-layer.9": 13044.6855, "epoch": 0.33, "inbatch_neg_score": 0.8873, "inbatch_pos_score": 1.5039, "learning_rate": 3.281578947368421e-05, "loss": 3.5152, "norm_diff": 0.1004, "norm_loss": 0.0, "num_token_doc": 66.8844, "num_token_overlap": 11.7096, "num_token_query": 31.5033, "num_token_union": 65.2238, "num_word_context": 202.3624, "num_word_doc": 49.8924, "num_word_query": 23.397, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21591.8762, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.8877, "query_norm": 1.5483, "queue_k_norm": 1.6469, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5033, "sent_len_1": 66.8844, "sent_len_max_0": 127.9638, "sent_len_max_1": 212.2663, "stdk": 0.049, "stdq": 0.0436, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75300 }, { "accuracy": 47.6562, "active_queue_size": 16384.0, "cl_loss": 3.5335, "doc_norm": 1.6515, "encoder_q-embeddings": 10349.9209, "encoder_q-layer.0": 6935.8828, "encoder_q-layer.1": 7152.0396, "encoder_q-layer.10": 13339.459, "encoder_q-layer.11": 30823.8125, "encoder_q-layer.2": 8120.5889, "encoder_q-layer.3": 8367.6318, "encoder_q-layer.4": 8989.9297, "encoder_q-layer.5": 9267.2754, "encoder_q-layer.6": 10192.5879, "encoder_q-layer.7": 12403.1738, "encoder_q-layer.8": 14471.3633, "encoder_q-layer.9": 12601.4453, "epoch": 0.33, "inbatch_neg_score": 0.8895, "inbatch_pos_score": 1.4697, "learning_rate": 3.278947368421052e-05, "loss": 3.5335, "norm_diff": 0.1228, "norm_loss": 0.0, "num_token_doc": 66.7194, "num_token_overlap": 11.6413, "num_token_query": 31.3423, "num_token_union": 65.1251, "num_word_context": 202.2483, "num_word_doc": 49.7976, "num_word_query": 23.2859, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19912.989, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.8896, "query_norm": 1.5287, "queue_k_norm": 1.6475, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3423, "sent_len_1": 66.7194, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.6075, "stdk": 0.0491, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75400 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.5209, "doc_norm": 1.6508, "encoder_q-embeddings": 10846.7344, "encoder_q-layer.0": 7183.4941, "encoder_q-layer.1": 7692.1914, "encoder_q-layer.10": 12939.7236, "encoder_q-layer.11": 30300.7324, "encoder_q-layer.2": 8483.8242, "encoder_q-layer.3": 8537.3174, "encoder_q-layer.4": 8938.6914, "encoder_q-layer.5": 9388.4902, "encoder_q-layer.6": 10466.4912, "encoder_q-layer.7": 11526.8701, "encoder_q-layer.8": 13955.5146, "encoder_q-layer.9": 12638.5977, "epoch": 0.33, "inbatch_neg_score": 0.8897, "inbatch_pos_score": 1.498, "learning_rate": 3.276315789473685e-05, "loss": 3.5209, "norm_diff": 0.1069, "norm_loss": 0.0, "num_token_doc": 66.8523, "num_token_overlap": 11.6721, "num_token_query": 31.3151, "num_token_union": 65.1792, "num_word_context": 202.411, "num_word_doc": 49.8944, "num_word_query": 23.2594, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19496.7988, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.8896, "query_norm": 1.544, "queue_k_norm": 1.6504, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3151, "sent_len_1": 66.8523, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.2925, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 75500 }, { "accuracy": 47.4121, "active_queue_size": 16384.0, "cl_loss": 3.5185, "doc_norm": 1.6471, "encoder_q-embeddings": 9835.0303, "encoder_q-layer.0": 6806.4897, "encoder_q-layer.1": 7380.5591, "encoder_q-layer.10": 15436.9941, "encoder_q-layer.11": 31697.9531, "encoder_q-layer.2": 8258.2344, "encoder_q-layer.3": 8600.5527, "encoder_q-layer.4": 9323.7344, "encoder_q-layer.5": 9468.6982, "encoder_q-layer.6": 10227.5723, "encoder_q-layer.7": 11595.8096, "encoder_q-layer.8": 14174.3389, "encoder_q-layer.9": 13806.4014, "epoch": 0.33, "inbatch_neg_score": 0.8941, "inbatch_pos_score": 1.4805, "learning_rate": 3.273684210526316e-05, "loss": 3.5185, "norm_diff": 0.0967, "norm_loss": 0.0, "num_token_doc": 66.6507, "num_token_overlap": 11.6861, "num_token_query": 31.4979, "num_token_union": 65.1453, "num_word_context": 202.451, "num_word_doc": 49.7463, "num_word_query": 23.3964, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19871.1866, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.894, "query_norm": 1.5504, "queue_k_norm": 1.6481, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4979, "sent_len_1": 66.6507, "sent_len_max_0": 127.965, "sent_len_max_1": 206.675, "stdk": 0.0489, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75600 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.5316, "doc_norm": 1.647, "encoder_q-embeddings": 4963.8691, "encoder_q-layer.0": 3409.3462, "encoder_q-layer.1": 3602.5803, "encoder_q-layer.10": 6367.3784, "encoder_q-layer.11": 15065.2842, "encoder_q-layer.2": 3951.2236, "encoder_q-layer.3": 4129.4917, "encoder_q-layer.4": 4278.2612, "encoder_q-layer.5": 4483.2441, "encoder_q-layer.6": 5212.8467, "encoder_q-layer.7": 6055.6875, "encoder_q-layer.8": 7129.7627, "encoder_q-layer.9": 6342.8174, "epoch": 0.33, "inbatch_neg_score": 0.8882, "inbatch_pos_score": 1.4805, "learning_rate": 3.271052631578948e-05, "loss": 3.5316, "norm_diff": 0.1112, "norm_loss": 0.0, "num_token_doc": 66.7813, "num_token_overlap": 11.6886, "num_token_query": 31.3904, "num_token_union": 65.1182, "num_word_context": 202.3153, "num_word_doc": 49.825, "num_word_query": 23.3076, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9786.3797, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8887, "query_norm": 1.5358, "queue_k_norm": 1.648, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3904, "sent_len_1": 66.7813, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.2825, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0489, "stdqueue_q": 0.0, "step": 75700 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.503, "doc_norm": 1.6491, "encoder_q-embeddings": 7404.7959, "encoder_q-layer.0": 5032.8281, "encoder_q-layer.1": 5264.0303, "encoder_q-layer.10": 6724.2441, "encoder_q-layer.11": 15142.834, "encoder_q-layer.2": 5889.207, "encoder_q-layer.3": 5741.0908, "encoder_q-layer.4": 6068.5088, "encoder_q-layer.5": 6289.104, "encoder_q-layer.6": 6856.5449, "encoder_q-layer.7": 6846.6357, "encoder_q-layer.8": 7192.8838, "encoder_q-layer.9": 6164.1841, "epoch": 0.33, "inbatch_neg_score": 0.893, "inbatch_pos_score": 1.4795, "learning_rate": 3.268421052631579e-05, "loss": 3.503, "norm_diff": 0.1132, "norm_loss": 0.0, "num_token_doc": 66.7594, "num_token_overlap": 11.6501, "num_token_query": 31.2919, "num_token_union": 65.0706, "num_word_context": 202.1637, "num_word_doc": 49.8056, "num_word_query": 23.2358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11028.6889, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8931, "query_norm": 1.5359, "queue_k_norm": 1.6507, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2919, "sent_len_1": 66.7594, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.33, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75800 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.5274, "doc_norm": 1.6514, "encoder_q-embeddings": 18203.9512, "encoder_q-layer.0": 13016.0156, "encoder_q-layer.1": 16461.8828, "encoder_q-layer.10": 6568.9111, "encoder_q-layer.11": 14946.3271, "encoder_q-layer.2": 18418.5859, "encoder_q-layer.3": 20456.2754, "encoder_q-layer.4": 22154.8945, "encoder_q-layer.5": 23723.5117, "encoder_q-layer.6": 22258.5645, "encoder_q-layer.7": 20383.2852, "encoder_q-layer.8": 14825.1875, "encoder_q-layer.9": 8183.2847, "epoch": 0.33, "inbatch_neg_score": 0.8925, "inbatch_pos_score": 1.4854, "learning_rate": 3.265789473684211e-05, "loss": 3.5274, "norm_diff": 0.1197, "norm_loss": 0.0, "num_token_doc": 66.7212, "num_token_overlap": 11.6851, "num_token_query": 31.3557, "num_token_union": 65.1038, "num_word_context": 202.366, "num_word_doc": 49.8148, "num_word_query": 23.2941, "postclip_grad_norm": 1.0, "preclip_grad_norm": 25807.949, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.8916, "query_norm": 1.5317, "queue_k_norm": 1.6495, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3557, "sent_len_1": 66.7212, "sent_len_max_0": 127.98, "sent_len_max_1": 208.47, "stdk": 0.049, "stdq": 0.0426, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 75900 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.5146, "doc_norm": 1.6523, "encoder_q-embeddings": 7349.1436, "encoder_q-layer.0": 4950.1465, "encoder_q-layer.1": 5873.5244, "encoder_q-layer.10": 6493.481, "encoder_q-layer.11": 15240.9248, "encoder_q-layer.2": 6583.2944, "encoder_q-layer.3": 6512.9678, "encoder_q-layer.4": 6719.6074, "encoder_q-layer.5": 7618.3579, "encoder_q-layer.6": 6794.0889, "encoder_q-layer.7": 8500.6396, "encoder_q-layer.8": 8707.7363, "encoder_q-layer.9": 6239.4482, "epoch": 0.33, "inbatch_neg_score": 0.8937, "inbatch_pos_score": 1.4805, "learning_rate": 3.2631578947368426e-05, "loss": 3.5146, "norm_diff": 0.1104, "norm_loss": 0.0, "num_token_doc": 66.7575, "num_token_overlap": 11.6648, "num_token_query": 31.3882, "num_token_union": 65.1161, "num_word_context": 202.4305, "num_word_doc": 49.8192, "num_word_query": 23.3064, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11831.7512, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8936, "query_norm": 1.5419, "queue_k_norm": 1.6507, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3882, "sent_len_1": 66.7575, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.875, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76000 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.5309, "doc_norm": 1.6495, "encoder_q-embeddings": 5513.3926, "encoder_q-layer.0": 3796.4539, "encoder_q-layer.1": 4042.9976, "encoder_q-layer.10": 8461.6973, "encoder_q-layer.11": 16442.2988, "encoder_q-layer.2": 4688.9604, "encoder_q-layer.3": 4890.5889, "encoder_q-layer.4": 5108.4146, "encoder_q-layer.5": 5210.3828, "encoder_q-layer.6": 5800.5205, "encoder_q-layer.7": 6808.3691, "encoder_q-layer.8": 8717.5225, "encoder_q-layer.9": 8035.4551, "epoch": 0.33, "inbatch_neg_score": 0.8935, "inbatch_pos_score": 1.4814, "learning_rate": 3.260526315789474e-05, "loss": 3.5309, "norm_diff": 0.0887, "norm_loss": 0.0, "num_token_doc": 66.7482, "num_token_overlap": 11.6537, "num_token_query": 31.3239, "num_token_union": 65.069, "num_word_context": 202.2387, "num_word_doc": 49.7779, "num_word_query": 23.237, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10773.3183, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8931, "query_norm": 1.5609, "queue_k_norm": 1.6513, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3239, "sent_len_1": 66.7482, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.4375, "stdk": 0.0489, "stdq": 0.0437, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76100 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.5179, "doc_norm": 1.6542, "encoder_q-embeddings": 5152.8774, "encoder_q-layer.0": 3431.4976, "encoder_q-layer.1": 3617.4265, "encoder_q-layer.10": 6508.1064, "encoder_q-layer.11": 15000.0137, "encoder_q-layer.2": 4020.7271, "encoder_q-layer.3": 4210.2505, "encoder_q-layer.4": 4524.1655, "encoder_q-layer.5": 4734.9561, "encoder_q-layer.6": 4983.6821, "encoder_q-layer.7": 6078.1392, "encoder_q-layer.8": 7030.9683, "encoder_q-layer.9": 6211.5645, "epoch": 0.33, "inbatch_neg_score": 0.8947, "inbatch_pos_score": 1.4922, "learning_rate": 3.2578947368421056e-05, "loss": 3.5179, "norm_diff": 0.1092, "norm_loss": 0.0, "num_token_doc": 67.1102, "num_token_overlap": 11.6881, "num_token_query": 31.3624, "num_token_union": 65.2969, "num_word_context": 202.6936, "num_word_doc": 50.0526, "num_word_query": 23.288, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9721.7364, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8945, "query_norm": 1.545, "queue_k_norm": 1.6504, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3624, "sent_len_1": 67.1102, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.5725, "stdk": 0.0491, "stdq": 0.0426, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76200 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.5091, "doc_norm": 1.6522, "encoder_q-embeddings": 5570.9062, "encoder_q-layer.0": 3818.8008, "encoder_q-layer.1": 4256.1899, "encoder_q-layer.10": 6938.5107, "encoder_q-layer.11": 15739.1816, "encoder_q-layer.2": 4557.332, "encoder_q-layer.3": 4432.6372, "encoder_q-layer.4": 4751.7998, "encoder_q-layer.5": 4686.3862, "encoder_q-layer.6": 5009.9893, "encoder_q-layer.7": 5784.8047, "encoder_q-layer.8": 6940.6016, "encoder_q-layer.9": 6304.4858, "epoch": 0.33, "inbatch_neg_score": 0.8997, "inbatch_pos_score": 1.4766, "learning_rate": 3.255263157894737e-05, "loss": 3.5091, "norm_diff": 0.0933, "norm_loss": 0.0, "num_token_doc": 66.8301, "num_token_overlap": 11.7099, "num_token_query": 31.4347, "num_token_union": 65.1703, "num_word_context": 202.1042, "num_word_doc": 49.8675, "num_word_query": 23.3481, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10017.9004, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.8999, "query_norm": 1.5589, "queue_k_norm": 1.6524, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4347, "sent_len_1": 66.8301, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.975, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76300 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.5154, "doc_norm": 1.6497, "encoder_q-embeddings": 5381.5547, "encoder_q-layer.0": 3793.9866, "encoder_q-layer.1": 4177.21, "encoder_q-layer.10": 6602.5576, "encoder_q-layer.11": 14707.0078, "encoder_q-layer.2": 4713.3306, "encoder_q-layer.3": 4819.1152, "encoder_q-layer.4": 5221.4697, "encoder_q-layer.5": 5646.7993, "encoder_q-layer.6": 5781.9468, "encoder_q-layer.7": 6534.416, "encoder_q-layer.8": 7162.4482, "encoder_q-layer.9": 6098.9053, "epoch": 0.33, "inbatch_neg_score": 0.9079, "inbatch_pos_score": 1.5049, "learning_rate": 3.2526315789473686e-05, "loss": 3.5154, "norm_diff": 0.0714, "norm_loss": 0.0, "num_token_doc": 66.7981, "num_token_overlap": 11.6794, "num_token_query": 31.3655, "num_token_union": 65.0774, "num_word_context": 202.3987, "num_word_doc": 49.851, "num_word_query": 23.3078, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10042.8788, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9077, "query_norm": 1.5783, "queue_k_norm": 1.652, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3655, "sent_len_1": 66.7981, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.6287, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76400 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.5181, "doc_norm": 1.6542, "encoder_q-embeddings": 5150.457, "encoder_q-layer.0": 3402.1765, "encoder_q-layer.1": 3673.4353, "encoder_q-layer.10": 6412.708, "encoder_q-layer.11": 14815.7627, "encoder_q-layer.2": 4163.521, "encoder_q-layer.3": 4205.769, "encoder_q-layer.4": 4495.7676, "encoder_q-layer.5": 4555.4521, "encoder_q-layer.6": 5084.0693, "encoder_q-layer.7": 6059.1323, "encoder_q-layer.8": 7079.6045, "encoder_q-layer.9": 6440.6162, "epoch": 0.33, "inbatch_neg_score": 0.9179, "inbatch_pos_score": 1.5225, "learning_rate": 3.2500000000000004e-05, "loss": 3.5181, "norm_diff": 0.0548, "norm_loss": 0.0, "num_token_doc": 66.6795, "num_token_overlap": 11.67, "num_token_query": 31.3537, "num_token_union": 65.076, "num_word_context": 202.1209, "num_word_doc": 49.7513, "num_word_query": 23.2924, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9680.548, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9185, "query_norm": 1.5993, "queue_k_norm": 1.6523, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3537, "sent_len_1": 66.6795, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.8988, "stdk": 0.049, "stdq": 0.0436, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76500 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.5098, "doc_norm": 1.6513, "encoder_q-embeddings": 5515.2695, "encoder_q-layer.0": 3815.8059, "encoder_q-layer.1": 4136.3872, "encoder_q-layer.10": 6390.7446, "encoder_q-layer.11": 14605.1797, "encoder_q-layer.2": 4478.7207, "encoder_q-layer.3": 4414.1582, "encoder_q-layer.4": 4647.3467, "encoder_q-layer.5": 4615.2388, "encoder_q-layer.6": 5258.1738, "encoder_q-layer.7": 5925.1689, "encoder_q-layer.8": 6568.7104, "encoder_q-layer.9": 6139.2583, "epoch": 0.33, "inbatch_neg_score": 0.9327, "inbatch_pos_score": 1.5234, "learning_rate": 3.2473684210526316e-05, "loss": 3.5098, "norm_diff": 0.0574, "norm_loss": 0.0, "num_token_doc": 66.682, "num_token_overlap": 11.6381, "num_token_query": 31.2697, "num_token_union": 65.0255, "num_word_context": 202.1532, "num_word_doc": 49.7567, "num_word_query": 23.2098, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9782.3094, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9312, "query_norm": 1.5939, "queue_k_norm": 1.6549, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2697, "sent_len_1": 66.682, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.4925, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 76600 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.486, "doc_norm": 1.6555, "encoder_q-embeddings": 6331.1572, "encoder_q-layer.0": 4625.8726, "encoder_q-layer.1": 4983.1812, "encoder_q-layer.10": 6719.6899, "encoder_q-layer.11": 14680.0127, "encoder_q-layer.2": 5177.2236, "encoder_q-layer.3": 5260.6265, "encoder_q-layer.4": 5498.874, "encoder_q-layer.5": 5298.1704, "encoder_q-layer.6": 6437.9834, "encoder_q-layer.7": 6782.6934, "encoder_q-layer.8": 7182.8213, "encoder_q-layer.9": 6512.4272, "epoch": 0.33, "inbatch_neg_score": 0.9483, "inbatch_pos_score": 1.5547, "learning_rate": 3.2447368421052634e-05, "loss": 3.486, "norm_diff": 0.0527, "norm_loss": 0.0, "num_token_doc": 66.828, "num_token_overlap": 11.7302, "num_token_query": 31.4832, "num_token_union": 65.168, "num_word_context": 202.3938, "num_word_doc": 49.8675, "num_word_query": 23.3953, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10418.3819, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9482, "query_norm": 1.6028, "queue_k_norm": 1.6568, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4832, "sent_len_1": 66.828, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.4588, "stdk": 0.049, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 76700 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.52, "doc_norm": 1.6593, "encoder_q-embeddings": 4906.4272, "encoder_q-layer.0": 3379.5898, "encoder_q-layer.1": 3529.2065, "encoder_q-layer.10": 6212.8374, "encoder_q-layer.11": 14666.7979, "encoder_q-layer.2": 3873.1233, "encoder_q-layer.3": 3921.6299, "encoder_q-layer.4": 4341.4199, "encoder_q-layer.5": 4351.5474, "encoder_q-layer.6": 4808.9277, "encoder_q-layer.7": 5405.8071, "encoder_q-layer.8": 6383.9429, "encoder_q-layer.9": 6027.9546, "epoch": 0.33, "inbatch_neg_score": 0.9581, "inbatch_pos_score": 1.5527, "learning_rate": 3.2421052631578945e-05, "loss": 3.52, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.7778, "num_token_overlap": 11.6773, "num_token_query": 31.4006, "num_token_union": 65.1345, "num_word_context": 202.3039, "num_word_doc": 49.8213, "num_word_query": 23.3184, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9367.9204, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9575, "query_norm": 1.5922, "queue_k_norm": 1.6564, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4006, "sent_len_1": 66.7778, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.0087, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 76800 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.5102, "doc_norm": 1.6597, "encoder_q-embeddings": 5711.5181, "encoder_q-layer.0": 3816.2759, "encoder_q-layer.1": 4219.5894, "encoder_q-layer.10": 6655.1353, "encoder_q-layer.11": 14744.4941, "encoder_q-layer.2": 4654.792, "encoder_q-layer.3": 4545.228, "encoder_q-layer.4": 4876.2007, "encoder_q-layer.5": 4851.9868, "encoder_q-layer.6": 5563.2417, "encoder_q-layer.7": 5975.96, "encoder_q-layer.8": 7411.3335, "encoder_q-layer.9": 6454.2759, "epoch": 0.33, "inbatch_neg_score": 0.9654, "inbatch_pos_score": 1.5762, "learning_rate": 3.2394736842105264e-05, "loss": 3.5102, "norm_diff": 0.062, "norm_loss": 0.0, "num_token_doc": 66.8289, "num_token_overlap": 11.6689, "num_token_query": 31.3025, "num_token_union": 65.1153, "num_word_context": 202.262, "num_word_doc": 49.8652, "num_word_query": 23.2548, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9878.7754, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9658, "query_norm": 1.5976, "queue_k_norm": 1.6601, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3025, "sent_len_1": 66.8289, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.1637, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 76900 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.5163, "doc_norm": 1.6634, "encoder_q-embeddings": 5326.6943, "encoder_q-layer.0": 3838.0862, "encoder_q-layer.1": 3998.9485, "encoder_q-layer.10": 6086.3848, "encoder_q-layer.11": 15283.4424, "encoder_q-layer.2": 4419.5918, "encoder_q-layer.3": 4443.6982, "encoder_q-layer.4": 4680.6934, "encoder_q-layer.5": 4594.7173, "encoder_q-layer.6": 4929.5142, "encoder_q-layer.7": 5502.6382, "encoder_q-layer.8": 6365.8691, "encoder_q-layer.9": 5896.6123, "epoch": 0.33, "inbatch_neg_score": 0.9739, "inbatch_pos_score": 1.5742, "learning_rate": 3.236842105263158e-05, "loss": 3.5163, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.6763, "num_token_overlap": 11.6733, "num_token_query": 31.3195, "num_token_union": 65.0199, "num_word_context": 202.2105, "num_word_doc": 49.7516, "num_word_query": 23.279, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9696.8463, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9731, "query_norm": 1.5931, "queue_k_norm": 1.6591, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3195, "sent_len_1": 66.6763, "sent_len_max_0": 127.9775, "sent_len_max_1": 211.8187, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77000 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.528, "doc_norm": 1.6623, "encoder_q-embeddings": 4871.3477, "encoder_q-layer.0": 3329.6235, "encoder_q-layer.1": 3554.0476, "encoder_q-layer.10": 7117.2646, "encoder_q-layer.11": 15361.8496, "encoder_q-layer.2": 4117.1748, "encoder_q-layer.3": 4372.5933, "encoder_q-layer.4": 4648.5049, "encoder_q-layer.5": 4924.0601, "encoder_q-layer.6": 5250.9956, "encoder_q-layer.7": 6073.5859, "encoder_q-layer.8": 7511.7788, "encoder_q-layer.9": 6761.2593, "epoch": 0.33, "inbatch_neg_score": 0.9778, "inbatch_pos_score": 1.584, "learning_rate": 3.2342105263157894e-05, "loss": 3.528, "norm_diff": 0.0765, "norm_loss": 0.0, "num_token_doc": 66.6905, "num_token_overlap": 11.6584, "num_token_query": 31.3071, "num_token_union": 64.9915, "num_word_context": 202.1929, "num_word_doc": 49.7346, "num_word_query": 23.2515, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10050.5009, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.979, "query_norm": 1.5858, "queue_k_norm": 1.6615, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3071, "sent_len_1": 66.6905, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.8288, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77100 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.5112, "doc_norm": 1.6683, "encoder_q-embeddings": 5079.2397, "encoder_q-layer.0": 3309.4182, "encoder_q-layer.1": 3697.2988, "encoder_q-layer.10": 6828.4009, "encoder_q-layer.11": 14853.9551, "encoder_q-layer.2": 4067.6008, "encoder_q-layer.3": 4194.0942, "encoder_q-layer.4": 4455.5713, "encoder_q-layer.5": 4453.2217, "encoder_q-layer.6": 4798.0264, "encoder_q-layer.7": 5489.5947, "encoder_q-layer.8": 6665.9287, "encoder_q-layer.9": 6163.6377, "epoch": 0.33, "inbatch_neg_score": 0.9767, "inbatch_pos_score": 1.5811, "learning_rate": 3.231578947368421e-05, "loss": 3.5112, "norm_diff": 0.0985, "norm_loss": 0.0, "num_token_doc": 66.8037, "num_token_overlap": 11.6795, "num_token_query": 31.3841, "num_token_union": 65.1247, "num_word_context": 202.2118, "num_word_doc": 49.8633, "num_word_query": 23.3043, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9384.4424, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9771, "query_norm": 1.5699, "queue_k_norm": 1.6623, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3841, "sent_len_1": 66.8037, "sent_len_max_0": 127.9862, "sent_len_max_1": 206.8787, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77200 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.5143, "doc_norm": 1.6679, "encoder_q-embeddings": 6779.9521, "encoder_q-layer.0": 4980.6431, "encoder_q-layer.1": 5442.2144, "encoder_q-layer.10": 6783.2988, "encoder_q-layer.11": 15342.9004, "encoder_q-layer.2": 6051.5146, "encoder_q-layer.3": 6329.2002, "encoder_q-layer.4": 6935.2598, "encoder_q-layer.5": 7314.8994, "encoder_q-layer.6": 7446.4004, "encoder_q-layer.7": 7249.1733, "encoder_q-layer.8": 7845.748, "encoder_q-layer.9": 6591.3477, "epoch": 0.34, "inbatch_neg_score": 0.9839, "inbatch_pos_score": 1.582, "learning_rate": 3.228947368421053e-05, "loss": 3.5143, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.8082, "num_token_overlap": 11.7023, "num_token_query": 31.3822, "num_token_union": 65.1322, "num_word_context": 202.3965, "num_word_doc": 49.8447, "num_word_query": 23.3128, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11501.5729, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9834, "query_norm": 1.585, "queue_k_norm": 1.6642, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3822, "sent_len_1": 66.8082, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.9187, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77300 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.5069, "doc_norm": 1.6621, "encoder_q-embeddings": 5183.1025, "encoder_q-layer.0": 3450.1897, "encoder_q-layer.1": 3680.2834, "encoder_q-layer.10": 6712.1152, "encoder_q-layer.11": 16026.582, "encoder_q-layer.2": 4108.1948, "encoder_q-layer.3": 4276.417, "encoder_q-layer.4": 4403.3862, "encoder_q-layer.5": 4415.2222, "encoder_q-layer.6": 4944.3457, "encoder_q-layer.7": 5474.7095, "encoder_q-layer.8": 6855.6895, "encoder_q-layer.9": 6197.855, "epoch": 0.34, "inbatch_neg_score": 0.9882, "inbatch_pos_score": 1.5752, "learning_rate": 3.226315789473684e-05, "loss": 3.5069, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.8375, "num_token_overlap": 11.6525, "num_token_query": 31.4074, "num_token_union": 65.1814, "num_word_context": 202.3543, "num_word_doc": 49.8597, "num_word_query": 23.3275, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9835.0683, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9878, "query_norm": 1.5825, "queue_k_norm": 1.6661, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4074, "sent_len_1": 66.8375, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.6612, "stdk": 0.0488, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77400 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.5149, "doc_norm": 1.6673, "encoder_q-embeddings": 5539.9746, "encoder_q-layer.0": 3766.0068, "encoder_q-layer.1": 3839.3611, "encoder_q-layer.10": 6569.8101, "encoder_q-layer.11": 15307.7109, "encoder_q-layer.2": 4322.1226, "encoder_q-layer.3": 4434.5571, "encoder_q-layer.4": 4655.7446, "encoder_q-layer.5": 4751.5713, "encoder_q-layer.6": 5196.167, "encoder_q-layer.7": 5803.7759, "encoder_q-layer.8": 6911.936, "encoder_q-layer.9": 6307.1372, "epoch": 0.34, "inbatch_neg_score": 0.991, "inbatch_pos_score": 1.5918, "learning_rate": 3.223684210526316e-05, "loss": 3.5149, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 67.0661, "num_token_overlap": 11.6783, "num_token_query": 31.307, "num_token_union": 65.223, "num_word_context": 202.3613, "num_word_doc": 50.0213, "num_word_query": 23.256, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9952.9831, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9907, "query_norm": 1.5847, "queue_k_norm": 1.6692, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.307, "sent_len_1": 67.0661, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.4412, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 77500 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.5079, "doc_norm": 1.6668, "encoder_q-embeddings": 4997.3594, "encoder_q-layer.0": 3435.731, "encoder_q-layer.1": 3694.7671, "encoder_q-layer.10": 6042.3086, "encoder_q-layer.11": 14085.7383, "encoder_q-layer.2": 4151.9326, "encoder_q-layer.3": 4279.4717, "encoder_q-layer.4": 4604.6177, "encoder_q-layer.5": 4489.9282, "encoder_q-layer.6": 5146.4443, "encoder_q-layer.7": 5788.5762, "encoder_q-layer.8": 6408.665, "encoder_q-layer.9": 5865.5317, "epoch": 0.34, "inbatch_neg_score": 0.9945, "inbatch_pos_score": 1.5801, "learning_rate": 3.221052631578947e-05, "loss": 3.5079, "norm_diff": 0.0895, "norm_loss": 0.0, "num_token_doc": 66.7876, "num_token_overlap": 11.6673, "num_token_query": 31.3582, "num_token_union": 65.1286, "num_word_context": 202.2905, "num_word_doc": 49.809, "num_word_query": 23.2949, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9196.7271, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 0.9941, "query_norm": 1.5773, "queue_k_norm": 1.6699, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3582, "sent_len_1": 66.7876, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.1788, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77600 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.516, "doc_norm": 1.6727, "encoder_q-embeddings": 10021.5146, "encoder_q-layer.0": 6620.6924, "encoder_q-layer.1": 7213.3208, "encoder_q-layer.10": 13271.4932, "encoder_q-layer.11": 29694.5371, "encoder_q-layer.2": 8132.7036, "encoder_q-layer.3": 8455.3936, "encoder_q-layer.4": 9030.3643, "encoder_q-layer.5": 8908.666, "encoder_q-layer.6": 10195.3779, "encoder_q-layer.7": 11072.6387, "encoder_q-layer.8": 13024.9619, "encoder_q-layer.9": 12057.1094, "epoch": 0.34, "inbatch_neg_score": 0.9961, "inbatch_pos_score": 1.6045, "learning_rate": 3.218421052631579e-05, "loss": 3.516, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.8441, "num_token_overlap": 11.6673, "num_token_query": 31.366, "num_token_union": 65.1648, "num_word_context": 202.3528, "num_word_doc": 49.9035, "num_word_query": 23.3088, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18838.941, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.9956, "query_norm": 1.5913, "queue_k_norm": 1.6706, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.366, "sent_len_1": 66.8441, "sent_len_max_0": 127.98, "sent_len_max_1": 209.16, "stdk": 0.049, "stdq": 0.0437, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77700 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.5049, "doc_norm": 1.6751, "encoder_q-embeddings": 11230.0371, "encoder_q-layer.0": 7914.144, "encoder_q-layer.1": 8592.3896, "encoder_q-layer.10": 12854.1807, "encoder_q-layer.11": 29879.6152, "encoder_q-layer.2": 9514.1172, "encoder_q-layer.3": 10166.8525, "encoder_q-layer.4": 10347.5645, "encoder_q-layer.5": 10353.4561, "encoder_q-layer.6": 10935.7402, "encoder_q-layer.7": 11621.125, "encoder_q-layer.8": 14118.8711, "encoder_q-layer.9": 12799.0908, "epoch": 0.34, "inbatch_neg_score": 0.9997, "inbatch_pos_score": 1.6016, "learning_rate": 3.215789473684211e-05, "loss": 3.5049, "norm_diff": 0.0893, "norm_loss": 0.0, "num_token_doc": 66.9059, "num_token_overlap": 11.6848, "num_token_query": 31.4331, "num_token_union": 65.2296, "num_word_context": 202.6186, "num_word_doc": 49.9419, "num_word_query": 23.3495, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20092.8115, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.9995, "query_norm": 1.5859, "queue_k_norm": 1.6723, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4331, "sent_len_1": 66.9059, "sent_len_max_0": 127.9525, "sent_len_max_1": 208.245, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77800 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.5016, "doc_norm": 1.6745, "encoder_q-embeddings": 10209.0088, "encoder_q-layer.0": 6957.0044, "encoder_q-layer.1": 7292.4927, "encoder_q-layer.10": 13168.8369, "encoder_q-layer.11": 30838.1094, "encoder_q-layer.2": 8114.5747, "encoder_q-layer.3": 8593.501, "encoder_q-layer.4": 9110.4902, "encoder_q-layer.5": 8991.9053, "encoder_q-layer.6": 10150.457, "encoder_q-layer.7": 11453.2334, "encoder_q-layer.8": 13976.6904, "encoder_q-layer.9": 12381.5625, "epoch": 0.34, "inbatch_neg_score": 1.0009, "inbatch_pos_score": 1.5947, "learning_rate": 3.213157894736842e-05, "loss": 3.5016, "norm_diff": 0.0976, "norm_loss": 0.0, "num_token_doc": 66.6398, "num_token_overlap": 11.6603, "num_token_query": 31.3967, "num_token_union": 65.0665, "num_word_context": 201.9886, "num_word_doc": 49.7169, "num_word_query": 23.3202, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19496.5317, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.001, "query_norm": 1.5769, "queue_k_norm": 1.6732, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3967, "sent_len_1": 66.6398, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.8175, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 77900 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5158, "doc_norm": 1.6763, "encoder_q-embeddings": 10024.0977, "encoder_q-layer.0": 6703.3057, "encoder_q-layer.1": 7226.2314, "encoder_q-layer.10": 12486.4707, "encoder_q-layer.11": 30829.7266, "encoder_q-layer.2": 8031.3311, "encoder_q-layer.3": 8384.2773, "encoder_q-layer.4": 8823.2637, "encoder_q-layer.5": 8869.0059, "encoder_q-layer.6": 9549.5117, "encoder_q-layer.7": 10920.0918, "encoder_q-layer.8": 12771.9766, "encoder_q-layer.9": 11873.5186, "epoch": 0.34, "inbatch_neg_score": 0.9995, "inbatch_pos_score": 1.5947, "learning_rate": 3.210526315789474e-05, "loss": 3.5158, "norm_diff": 0.1047, "norm_loss": 0.0, "num_token_doc": 66.7421, "num_token_overlap": 11.6757, "num_token_query": 31.3922, "num_token_union": 65.1043, "num_word_context": 202.0799, "num_word_doc": 49.7784, "num_word_query": 23.3038, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18846.9861, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.9995, "query_norm": 1.5716, "queue_k_norm": 1.6751, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3922, "sent_len_1": 66.7421, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.4112, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 78000 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5196, "doc_norm": 1.6753, "encoder_q-embeddings": 10075.0703, "encoder_q-layer.0": 6855.6899, "encoder_q-layer.1": 7217.4141, "encoder_q-layer.10": 15211.9492, "encoder_q-layer.11": 33514.957, "encoder_q-layer.2": 8309.9375, "encoder_q-layer.3": 8548.4893, "encoder_q-layer.4": 9101.0889, "encoder_q-layer.5": 8962.3467, "encoder_q-layer.6": 10225.584, "encoder_q-layer.7": 11902.8828, "encoder_q-layer.8": 15156.7832, "encoder_q-layer.9": 13739.8691, "epoch": 0.34, "inbatch_neg_score": 1.0009, "inbatch_pos_score": 1.6074, "learning_rate": 3.207894736842105e-05, "loss": 3.5196, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.7093, "num_token_overlap": 11.7216, "num_token_query": 31.4549, "num_token_union": 65.0888, "num_word_context": 202.2737, "num_word_doc": 49.7681, "num_word_query": 23.3566, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20265.9125, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.001, "query_norm": 1.5924, "queue_k_norm": 1.677, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4549, "sent_len_1": 66.7093, "sent_len_max_0": 127.975, "sent_len_max_1": 208.5575, "stdk": 0.0489, "stdq": 0.0439, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 78100 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.5199, "doc_norm": 1.6788, "encoder_q-embeddings": 10078.0195, "encoder_q-layer.0": 6917.8057, "encoder_q-layer.1": 7130.6689, "encoder_q-layer.10": 14622.3418, "encoder_q-layer.11": 30321.7461, "encoder_q-layer.2": 7970.644, "encoder_q-layer.3": 8153.4829, "encoder_q-layer.4": 8655.1221, "encoder_q-layer.5": 9011.6104, "encoder_q-layer.6": 9701.707, "encoder_q-layer.7": 10985.6172, "encoder_q-layer.8": 13560.373, "encoder_q-layer.9": 13373.8896, "epoch": 0.34, "inbatch_neg_score": 0.996, "inbatch_pos_score": 1.5869, "learning_rate": 3.205263157894737e-05, "loss": 3.5199, "norm_diff": 0.1021, "norm_loss": 0.0, "num_token_doc": 66.6216, "num_token_overlap": 11.6195, "num_token_query": 31.2541, "num_token_union": 64.9961, "num_word_context": 202.1448, "num_word_doc": 49.7162, "num_word_query": 23.2045, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19277.954, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 0.9966, "query_norm": 1.5767, "queue_k_norm": 1.6794, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2541, "sent_len_1": 66.6216, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.4137, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 78200 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.5253, "doc_norm": 1.6776, "encoder_q-embeddings": 9925.7656, "encoder_q-layer.0": 6807.1689, "encoder_q-layer.1": 7194.9819, "encoder_q-layer.10": 13803.9639, "encoder_q-layer.11": 29784.3105, "encoder_q-layer.2": 7837.7041, "encoder_q-layer.3": 8178.1499, "encoder_q-layer.4": 8880.6826, "encoder_q-layer.5": 8704.7139, "encoder_q-layer.6": 9667.4971, "encoder_q-layer.7": 12203.9131, "encoder_q-layer.8": 14098.0898, "encoder_q-layer.9": 13125.25, "epoch": 0.34, "inbatch_neg_score": 1.0004, "inbatch_pos_score": 1.5898, "learning_rate": 3.202631578947369e-05, "loss": 3.5253, "norm_diff": 0.099, "norm_loss": 0.0, "num_token_doc": 66.7781, "num_token_overlap": 11.625, "num_token_query": 31.2605, "num_token_union": 65.0408, "num_word_context": 201.9245, "num_word_doc": 49.7909, "num_word_query": 23.2164, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19353.8569, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0, "query_norm": 1.5786, "queue_k_norm": 1.6766, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2605, "sent_len_1": 66.7781, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.6113, "stdk": 0.0489, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 78300 }, { "accuracy": 46.8262, "active_queue_size": 16384.0, "cl_loss": 3.4885, "doc_norm": 1.6774, "encoder_q-embeddings": 10406.6602, "encoder_q-layer.0": 7062.2598, "encoder_q-layer.1": 7670.4858, "encoder_q-layer.10": 14674.9258, "encoder_q-layer.11": 32086.459, "encoder_q-layer.2": 8835.9727, "encoder_q-layer.3": 9043.1855, "encoder_q-layer.4": 9779.6709, "encoder_q-layer.5": 10075.166, "encoder_q-layer.6": 11136.6543, "encoder_q-layer.7": 12819.7461, "encoder_q-layer.8": 14725.8887, "encoder_q-layer.9": 13759.0938, "epoch": 0.34, "inbatch_neg_score": 1.0013, "inbatch_pos_score": 1.584, "learning_rate": 3.2000000000000005e-05, "loss": 3.4885, "norm_diff": 0.0973, "norm_loss": 0.0, "num_token_doc": 66.8296, "num_token_overlap": 11.6939, "num_token_query": 31.4512, "num_token_union": 65.1769, "num_word_context": 202.2324, "num_word_doc": 49.8445, "num_word_query": 23.3545, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20292.54, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.002, "query_norm": 1.5802, "queue_k_norm": 1.6791, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4512, "sent_len_1": 66.8296, "sent_len_max_0": 127.9675, "sent_len_max_1": 211.4375, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 78400 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.5216, "doc_norm": 1.6799, "encoder_q-embeddings": 13211.5938, "encoder_q-layer.0": 9270.167, "encoder_q-layer.1": 10154.8164, "encoder_q-layer.10": 12281.8945, "encoder_q-layer.11": 29261.9727, "encoder_q-layer.2": 11029.7148, "encoder_q-layer.3": 10558.2373, "encoder_q-layer.4": 11136.5918, "encoder_q-layer.5": 10474.1523, "encoder_q-layer.6": 10962.8545, "encoder_q-layer.7": 11854.5078, "encoder_q-layer.8": 13764.6719, "encoder_q-layer.9": 12085.8213, "epoch": 0.34, "inbatch_neg_score": 1.0046, "inbatch_pos_score": 1.583, "learning_rate": 3.197368421052632e-05, "loss": 3.5216, "norm_diff": 0.1092, "norm_loss": 0.0, "num_token_doc": 66.9, "num_token_overlap": 11.6613, "num_token_query": 31.3169, "num_token_union": 65.1556, "num_word_context": 202.5207, "num_word_doc": 49.8885, "num_word_query": 23.2573, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20665.1907, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0039, "query_norm": 1.5707, "queue_k_norm": 1.6805, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3169, "sent_len_1": 66.9, "sent_len_max_0": 127.99, "sent_len_max_1": 210.6475, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 78500 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.5059, "doc_norm": 1.682, "encoder_q-embeddings": 9509.5205, "encoder_q-layer.0": 6742.9634, "encoder_q-layer.1": 7181.5762, "encoder_q-layer.10": 13148.0283, "encoder_q-layer.11": 28314.3613, "encoder_q-layer.2": 7858.918, "encoder_q-layer.3": 8086.9961, "encoder_q-layer.4": 8773.6875, "encoder_q-layer.5": 8891.8994, "encoder_q-layer.6": 9707.1025, "encoder_q-layer.7": 10993.9932, "encoder_q-layer.8": 13550.8926, "encoder_q-layer.9": 12433.0195, "epoch": 0.34, "inbatch_neg_score": 1.005, "inbatch_pos_score": 1.5967, "learning_rate": 3.194736842105263e-05, "loss": 3.5059, "norm_diff": 0.1093, "norm_loss": 0.0, "num_token_doc": 66.7027, "num_token_overlap": 11.6887, "num_token_query": 31.3504, "num_token_union": 65.0694, "num_word_context": 202.3039, "num_word_doc": 49.765, "num_word_query": 23.2866, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18507.5127, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0049, "query_norm": 1.5727, "queue_k_norm": 1.6819, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3504, "sent_len_1": 66.7027, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.9638, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 78600 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.5006, "doc_norm": 1.6845, "encoder_q-embeddings": 10128.9062, "encoder_q-layer.0": 6738.5649, "encoder_q-layer.1": 7129.5532, "encoder_q-layer.10": 12391.3008, "encoder_q-layer.11": 27796.8145, "encoder_q-layer.2": 7814.9487, "encoder_q-layer.3": 8243.6729, "encoder_q-layer.4": 8646.0859, "encoder_q-layer.5": 8678.4609, "encoder_q-layer.6": 10238.1846, "encoder_q-layer.7": 10510.6387, "encoder_q-layer.8": 12719.6328, "encoder_q-layer.9": 11911.3867, "epoch": 0.34, "inbatch_neg_score": 1.005, "inbatch_pos_score": 1.6211, "learning_rate": 3.1921052631578954e-05, "loss": 3.5006, "norm_diff": 0.1176, "norm_loss": 0.0, "num_token_doc": 66.8254, "num_token_overlap": 11.6618, "num_token_query": 31.3712, "num_token_union": 65.1934, "num_word_context": 202.5145, "num_word_doc": 49.8897, "num_word_query": 23.2875, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18058.7904, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0049, "query_norm": 1.5669, "queue_k_norm": 1.681, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3712, "sent_len_1": 66.8254, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.7475, "stdk": 0.0491, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 78700 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.5246, "doc_norm": 1.6809, "encoder_q-embeddings": 17024.3965, "encoder_q-layer.0": 11475.542, "encoder_q-layer.1": 13101.5107, "encoder_q-layer.10": 12333.8037, "encoder_q-layer.11": 30986.5762, "encoder_q-layer.2": 14437.3652, "encoder_q-layer.3": 15394.9854, "encoder_q-layer.4": 16262.415, "encoder_q-layer.5": 16657.457, "encoder_q-layer.6": 14859.8545, "encoder_q-layer.7": 12871.6426, "encoder_q-layer.8": 14172.3809, "encoder_q-layer.9": 12024.5547, "epoch": 0.34, "inbatch_neg_score": 1.0076, "inbatch_pos_score": 1.6113, "learning_rate": 3.1894736842105265e-05, "loss": 3.5246, "norm_diff": 0.1025, "norm_loss": 0.0, "num_token_doc": 66.5506, "num_token_overlap": 11.6528, "num_token_query": 31.3638, "num_token_union": 64.9793, "num_word_context": 202.1901, "num_word_doc": 49.6209, "num_word_query": 23.2909, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24349.1275, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0078, "query_norm": 1.5784, "queue_k_norm": 1.6839, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3638, "sent_len_1": 66.5506, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.4888, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 78800 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.5163, "doc_norm": 1.6867, "encoder_q-embeddings": 9848.1553, "encoder_q-layer.0": 6827.772, "encoder_q-layer.1": 7367.1748, "encoder_q-layer.10": 12658.5479, "encoder_q-layer.11": 29858.6484, "encoder_q-layer.2": 8198.4688, "encoder_q-layer.3": 8363.6768, "encoder_q-layer.4": 8964.334, "encoder_q-layer.5": 9332.916, "encoder_q-layer.6": 10298.4912, "encoder_q-layer.7": 11254.9189, "encoder_q-layer.8": 13437.9834, "encoder_q-layer.9": 12422.2021, "epoch": 0.34, "inbatch_neg_score": 1.0056, "inbatch_pos_score": 1.6074, "learning_rate": 3.1868421052631584e-05, "loss": 3.5163, "norm_diff": 0.1036, "norm_loss": 0.0, "num_token_doc": 66.6514, "num_token_overlap": 11.658, "num_token_query": 31.3841, "num_token_union": 65.0651, "num_word_context": 202.3327, "num_word_doc": 49.7264, "num_word_query": 23.3157, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19087.445, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0059, "query_norm": 1.5831, "queue_k_norm": 1.6823, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3841, "sent_len_1": 66.6514, "sent_len_max_0": 127.975, "sent_len_max_1": 206.8663, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 78900 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.5089, "doc_norm": 1.6887, "encoder_q-embeddings": 11027.2969, "encoder_q-layer.0": 7406.6479, "encoder_q-layer.1": 7944.2446, "encoder_q-layer.10": 14784.4082, "encoder_q-layer.11": 29602.9727, "encoder_q-layer.2": 8900.7129, "encoder_q-layer.3": 9157.5283, "encoder_q-layer.4": 10055.1562, "encoder_q-layer.5": 10239.3574, "encoder_q-layer.6": 11816.2959, "encoder_q-layer.7": 12301.7314, "encoder_q-layer.8": 15241.2881, "encoder_q-layer.9": 12910.835, "epoch": 0.34, "inbatch_neg_score": 1.009, "inbatch_pos_score": 1.6113, "learning_rate": 3.1842105263157895e-05, "loss": 3.5089, "norm_diff": 0.1039, "norm_loss": 0.0, "num_token_doc": 66.5947, "num_token_overlap": 11.6421, "num_token_query": 31.3498, "num_token_union": 65.0227, "num_word_context": 202.2081, "num_word_doc": 49.691, "num_word_query": 23.2852, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20087.0637, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0098, "query_norm": 1.5848, "queue_k_norm": 1.6848, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3498, "sent_len_1": 66.5947, "sent_len_max_0": 127.995, "sent_len_max_1": 207.0863, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79000 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.5145, "doc_norm": 1.6811, "encoder_q-embeddings": 9897.3027, "encoder_q-layer.0": 6748.1924, "encoder_q-layer.1": 7121.8789, "encoder_q-layer.10": 12761.4502, "encoder_q-layer.11": 28644.0664, "encoder_q-layer.2": 7896.2109, "encoder_q-layer.3": 8046.98, "encoder_q-layer.4": 9001.8877, "encoder_q-layer.5": 9072.4023, "encoder_q-layer.6": 10447.3203, "encoder_q-layer.7": 11482.8545, "encoder_q-layer.8": 13907.3135, "encoder_q-layer.9": 11961.2402, "epoch": 0.34, "inbatch_neg_score": 1.013, "inbatch_pos_score": 1.6055, "learning_rate": 3.181578947368421e-05, "loss": 3.5145, "norm_diff": 0.1094, "norm_loss": 0.0, "num_token_doc": 66.6883, "num_token_overlap": 11.6884, "num_token_query": 31.3827, "num_token_union": 65.0858, "num_word_context": 202.3644, "num_word_doc": 49.7609, "num_word_query": 23.3161, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19082.3811, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0127, "query_norm": 1.5717, "queue_k_norm": 1.6849, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3827, "sent_len_1": 66.6883, "sent_len_max_0": 127.9712, "sent_len_max_1": 207.7275, "stdk": 0.0488, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79100 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.5108, "doc_norm": 1.685, "encoder_q-embeddings": 9889.7646, "encoder_q-layer.0": 6720.1602, "encoder_q-layer.1": 7158.0967, "encoder_q-layer.10": 13274.5137, "encoder_q-layer.11": 29414.8555, "encoder_q-layer.2": 7837.8931, "encoder_q-layer.3": 8116.98, "encoder_q-layer.4": 8675.4258, "encoder_q-layer.5": 8578.1953, "encoder_q-layer.6": 9758.8096, "encoder_q-layer.7": 11295.7764, "encoder_q-layer.8": 13228.9258, "encoder_q-layer.9": 12304.9854, "epoch": 0.34, "inbatch_neg_score": 1.0144, "inbatch_pos_score": 1.6055, "learning_rate": 3.178947368421053e-05, "loss": 3.5108, "norm_diff": 0.1003, "norm_loss": 0.0, "num_token_doc": 66.8748, "num_token_overlap": 11.7197, "num_token_query": 31.5595, "num_token_union": 65.27, "num_word_context": 202.552, "num_word_doc": 49.8825, "num_word_query": 23.4498, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18922.9279, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0146, "query_norm": 1.5847, "queue_k_norm": 1.6866, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5595, "sent_len_1": 66.8748, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.8862, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79200 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.513, "doc_norm": 1.694, "encoder_q-embeddings": 10069.0752, "encoder_q-layer.0": 7109.4897, "encoder_q-layer.1": 7409.7896, "encoder_q-layer.10": 14653.4961, "encoder_q-layer.11": 30786.6113, "encoder_q-layer.2": 8346.6875, "encoder_q-layer.3": 8277.7432, "encoder_q-layer.4": 8734.3691, "encoder_q-layer.5": 8913.9668, "encoder_q-layer.6": 9910.75, "encoder_q-layer.7": 11155.5791, "encoder_q-layer.8": 14514.6367, "encoder_q-layer.9": 13821.1621, "epoch": 0.34, "inbatch_neg_score": 1.0149, "inbatch_pos_score": 1.6182, "learning_rate": 3.1763157894736843e-05, "loss": 3.513, "norm_diff": 0.1135, "norm_loss": 0.0, "num_token_doc": 66.6897, "num_token_overlap": 11.6766, "num_token_query": 31.4289, "num_token_union": 65.1287, "num_word_context": 202.3019, "num_word_doc": 49.7786, "num_word_query": 23.3503, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19288.9542, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0156, "query_norm": 1.5806, "queue_k_norm": 1.6859, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4289, "sent_len_1": 66.6897, "sent_len_max_0": 127.995, "sent_len_max_1": 207.7837, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 79300 }, { "accuracy": 48.1445, "active_queue_size": 16384.0, "cl_loss": 3.5142, "doc_norm": 1.6884, "encoder_q-embeddings": 14384.6221, "encoder_q-layer.0": 10031.167, "encoder_q-layer.1": 11118.5518, "encoder_q-layer.10": 12649.6768, "encoder_q-layer.11": 29533.4395, "encoder_q-layer.2": 12207.501, "encoder_q-layer.3": 12474.6709, "encoder_q-layer.4": 12460.6455, "encoder_q-layer.5": 11764.2139, "encoder_q-layer.6": 11958.0674, "encoder_q-layer.7": 12558.3418, "encoder_q-layer.8": 13639.9531, "encoder_q-layer.9": 11970.1162, "epoch": 0.34, "inbatch_neg_score": 1.0142, "inbatch_pos_score": 1.6055, "learning_rate": 3.173684210526316e-05, "loss": 3.5142, "norm_diff": 0.1208, "norm_loss": 0.0, "num_token_doc": 66.7939, "num_token_overlap": 11.6755, "num_token_query": 31.3831, "num_token_union": 65.141, "num_word_context": 202.3391, "num_word_doc": 49.8081, "num_word_query": 23.2974, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21656.0995, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0146, "query_norm": 1.5676, "queue_k_norm": 1.6866, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3831, "sent_len_1": 66.7939, "sent_len_max_0": 127.9587, "sent_len_max_1": 209.3738, "stdk": 0.0491, "stdq": 0.0425, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 79400 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.5044, "doc_norm": 1.6864, "encoder_q-embeddings": 11326.373, "encoder_q-layer.0": 7786.2617, "encoder_q-layer.1": 8393.4951, "encoder_q-layer.10": 11756.9971, "encoder_q-layer.11": 28170.3633, "encoder_q-layer.2": 9364.9609, "encoder_q-layer.3": 9286.3877, "encoder_q-layer.4": 9784.6152, "encoder_q-layer.5": 10238.8037, "encoder_q-layer.6": 11105.2295, "encoder_q-layer.7": 12222.5684, "encoder_q-layer.8": 13219.2822, "encoder_q-layer.9": 11721.9551, "epoch": 0.34, "inbatch_neg_score": 1.0179, "inbatch_pos_score": 1.6416, "learning_rate": 3.1710526315789473e-05, "loss": 3.5044, "norm_diff": 0.0967, "norm_loss": 0.0, "num_token_doc": 66.7968, "num_token_overlap": 11.6726, "num_token_query": 31.3693, "num_token_union": 65.1556, "num_word_context": 202.2277, "num_word_doc": 49.8713, "num_word_query": 23.3079, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19094.7024, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0176, "query_norm": 1.5897, "queue_k_norm": 1.6873, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3693, "sent_len_1": 66.7968, "sent_len_max_0": 127.98, "sent_len_max_1": 208.8887, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79500 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.4946, "doc_norm": 1.692, "encoder_q-embeddings": 10440.8672, "encoder_q-layer.0": 7104.4863, "encoder_q-layer.1": 7539.6353, "encoder_q-layer.10": 12499.5625, "encoder_q-layer.11": 28493.3848, "encoder_q-layer.2": 8295.1514, "encoder_q-layer.3": 8535.6982, "encoder_q-layer.4": 9430.0908, "encoder_q-layer.5": 9260.125, "encoder_q-layer.6": 10007.4307, "encoder_q-layer.7": 10920.1689, "encoder_q-layer.8": 12460.1357, "encoder_q-layer.9": 11944.8945, "epoch": 0.35, "inbatch_neg_score": 1.0192, "inbatch_pos_score": 1.6299, "learning_rate": 3.168421052631579e-05, "loss": 3.4946, "norm_diff": 0.1, "norm_loss": 0.0, "num_token_doc": 66.7555, "num_token_overlap": 11.688, "num_token_query": 31.3852, "num_token_union": 65.1202, "num_word_context": 202.2966, "num_word_doc": 49.809, "num_word_query": 23.3199, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18895.6778, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0195, "query_norm": 1.592, "queue_k_norm": 1.6883, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3852, "sent_len_1": 66.7555, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.8413, "stdk": 0.0492, "stdq": 0.0436, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79600 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.5119, "doc_norm": 1.6934, "encoder_q-embeddings": 19689.4219, "encoder_q-layer.0": 13085.7734, "encoder_q-layer.1": 13837.3584, "encoder_q-layer.10": 28092.9336, "encoder_q-layer.11": 62170.3711, "encoder_q-layer.2": 15582.1094, "encoder_q-layer.3": 15899.5625, "encoder_q-layer.4": 17574.6758, "encoder_q-layer.5": 17678.2266, "encoder_q-layer.6": 20168.6895, "encoder_q-layer.7": 23035.6191, "encoder_q-layer.8": 27850.2383, "encoder_q-layer.9": 28320.8926, "epoch": 0.35, "inbatch_neg_score": 1.0191, "inbatch_pos_score": 1.6182, "learning_rate": 3.165789473684211e-05, "loss": 3.5119, "norm_diff": 0.1072, "norm_loss": 0.0, "num_token_doc": 66.8195, "num_token_overlap": 11.671, "num_token_query": 31.3617, "num_token_union": 65.1533, "num_word_context": 202.5159, "num_word_doc": 49.8393, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39137.9008, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.0205, "query_norm": 1.5861, "queue_k_norm": 1.6887, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3617, "sent_len_1": 66.8195, "sent_len_max_0": 127.9562, "sent_len_max_1": 210.8725, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79700 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.4969, "doc_norm": 1.6837, "encoder_q-embeddings": 29574.2598, "encoder_q-layer.0": 21441.4453, "encoder_q-layer.1": 22349.6543, "encoder_q-layer.10": 26563.9785, "encoder_q-layer.11": 58512.5352, "encoder_q-layer.2": 26097.0215, "encoder_q-layer.3": 25355.2715, "encoder_q-layer.4": 28225.9863, "encoder_q-layer.5": 28708.3125, "encoder_q-layer.6": 30691.3496, "encoder_q-layer.7": 30557.2188, "encoder_q-layer.8": 31124.0352, "encoder_q-layer.9": 25648.8984, "epoch": 0.35, "inbatch_neg_score": 1.0254, "inbatch_pos_score": 1.627, "learning_rate": 3.163157894736842e-05, "loss": 3.4969, "norm_diff": 0.0903, "norm_loss": 0.0, "num_token_doc": 66.8626, "num_token_overlap": 11.7271, "num_token_query": 31.4317, "num_token_union": 65.1827, "num_word_context": 202.3819, "num_word_doc": 49.8951, "num_word_query": 23.3488, "postclip_grad_norm": 1.0, "preclip_grad_norm": 46144.8215, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.0254, "query_norm": 1.5935, "queue_k_norm": 1.6879, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4317, "sent_len_1": 66.8626, "sent_len_max_0": 127.97, "sent_len_max_1": 209.8525, "stdk": 0.0488, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 79800 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.5069, "doc_norm": 1.6914, "encoder_q-embeddings": 21174.4023, "encoder_q-layer.0": 14882.6064, "encoder_q-layer.1": 16252.4619, "encoder_q-layer.10": 26812.8789, "encoder_q-layer.11": 60271.8203, "encoder_q-layer.2": 17793.7871, "encoder_q-layer.3": 18146.3301, "encoder_q-layer.4": 18831.1367, "encoder_q-layer.5": 19696.2246, "encoder_q-layer.6": 21045.0, "encoder_q-layer.7": 23647.5586, "encoder_q-layer.8": 27261.2324, "encoder_q-layer.9": 25126.9629, "epoch": 0.35, "inbatch_neg_score": 1.0277, "inbatch_pos_score": 1.6348, "learning_rate": 3.160526315789474e-05, "loss": 3.5069, "norm_diff": 0.0989, "norm_loss": 0.0, "num_token_doc": 66.7192, "num_token_overlap": 11.6568, "num_token_query": 31.3317, "num_token_union": 65.0816, "num_word_context": 202.3026, "num_word_doc": 49.7879, "num_word_query": 23.2701, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39751.9696, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.0283, "query_norm": 1.5925, "queue_k_norm": 1.6908, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3317, "sent_len_1": 66.7192, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.1912, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 79900 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.5037, "doc_norm": 1.6908, "encoder_q-embeddings": 19809.1367, "encoder_q-layer.0": 13520.3477, "encoder_q-layer.1": 14428.5723, "encoder_q-layer.10": 26771.334, "encoder_q-layer.11": 58914.0508, "encoder_q-layer.2": 16148.5928, "encoder_q-layer.3": 17007.0879, "encoder_q-layer.4": 18138.0723, "encoder_q-layer.5": 17904.5859, "encoder_q-layer.6": 20240.1035, "encoder_q-layer.7": 22839.3145, "encoder_q-layer.8": 27701.9531, "encoder_q-layer.9": 24949.293, "epoch": 0.35, "inbatch_neg_score": 1.0288, "inbatch_pos_score": 1.626, "learning_rate": 3.157894736842105e-05, "loss": 3.5037, "norm_diff": 0.1019, "norm_loss": 0.0, "num_token_doc": 66.6678, "num_token_overlap": 11.6451, "num_token_query": 31.3483, "num_token_union": 65.1054, "num_word_context": 202.3935, "num_word_doc": 49.7421, "num_word_query": 23.2675, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38280.9456, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.0293, "query_norm": 1.5889, "queue_k_norm": 1.6894, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3483, "sent_len_1": 66.6678, "sent_len_max_0": 127.9725, "sent_len_max_1": 206.8825, "stdk": 0.049, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 80000 }, { "dev_runtime": 29.0973, "dev_samples_per_second": 1.1, "dev_steps_per_second": 0.034, "epoch": 0.35, "step": 80000, "test_accuracy": 93.701171875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3554224371910095, "test_doc_norm": 1.664008378982544, "test_inbatch_neg_score": 1.3156574964523315, "test_inbatch_pos_score": 2.3161299228668213, "test_loss": 0.3554224371910095, "test_loss_align": 1.0588059425354004, "test_loss_unif": 1.9411154985427856, "test_loss_unif_q@queue": 1.9411154985427856, "test_norm_diff": 0.03842584788799286, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.031087875366211, "test_query_norm": 1.7024343013763428, "test_queue_k_norm": 1.6896429061889648, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043644979596138, "test_stdq": 0.04390425235033035, "test_stdqueue_k": 0.04903624206781387, "test_stdqueue_q": 0.0 }, { "dev_runtime": 29.0973, "dev_samples_per_second": 1.1, "dev_steps_per_second": 0.034, "epoch": 0.35, "eval_beir-arguana_ndcg@10": 0.3791, "eval_beir-arguana_recall@10": 0.65434, "eval_beir-arguana_recall@100": 0.93101, "eval_beir-arguana_recall@20": 0.77098, "eval_beir-avg_ndcg@10": 0.3937554166666667, "eval_beir-avg_recall@10": 0.46750791666666663, "eval_beir-avg_recall@100": 0.6437425000000001, "eval_beir-avg_recall@20": 0.5253990833333334, "eval_beir-cqadupstack_ndcg@10": 0.2750641666666667, "eval_beir-cqadupstack_recall@10": 0.3732491666666667, "eval_beir-cqadupstack_recall@100": 0.6167349999999999, "eval_beir-cqadupstack_recall@20": 0.44556083333333335, "eval_beir-fiqa_ndcg@10": 0.26378, "eval_beir-fiqa_recall@10": 0.33465, "eval_beir-fiqa_recall@100": 0.60601, "eval_beir-fiqa_recall@20": 0.41273, "eval_beir-nfcorpus_ndcg@10": 0.31755, "eval_beir-nfcorpus_recall@10": 0.152, "eval_beir-nfcorpus_recall@100": 0.29569, "eval_beir-nfcorpus_recall@20": 0.1886, "eval_beir-nq_ndcg@10": 0.31822, "eval_beir-nq_recall@10": 0.50686, "eval_beir-nq_recall@100": 0.83075, "eval_beir-nq_recall@20": 0.62877, "eval_beir-quora_ndcg@10": 0.78781, "eval_beir-quora_recall@10": 0.89495, "eval_beir-quora_recall@100": 0.98001, "eval_beir-quora_recall@20": 0.93649, "eval_beir-scidocs_ndcg@10": 0.15601, "eval_beir-scidocs_recall@10": 0.16687, "eval_beir-scidocs_recall@100": 0.37718, "eval_beir-scidocs_recall@20": 0.22427, "eval_beir-scifact_ndcg@10": 0.64876, "eval_beir-scifact_recall@10": 0.79833, "eval_beir-scifact_recall@100": 0.90822, "eval_beir-scifact_recall@20": 0.83744, "eval_beir-trec-covid_ndcg@10": 0.59694, "eval_beir-trec-covid_recall@10": 0.656, "eval_beir-trec-covid_recall@100": 0.4578, "eval_beir-trec-covid_recall@20": 0.6, "eval_beir-webis-touche2020_ndcg@10": 0.19432, "eval_beir-webis-touche2020_recall@10": 0.13783, "eval_beir-webis-touche2020_recall@100": 0.43402, "eval_beir-webis-touche2020_recall@20": 0.20915, "eval_senteval-avg_sts": 0.7444577700787828, "eval_senteval-sickr_spearman": 0.7256974447360104, "eval_senteval-stsb_spearman": 0.7632180954215552, "step": 80000, "test_accuracy": 93.701171875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3554224371910095, "test_doc_norm": 1.664008378982544, "test_inbatch_neg_score": 1.3156574964523315, "test_inbatch_pos_score": 2.3161299228668213, "test_loss": 0.3554224371910095, "test_loss_align": 1.0588059425354004, "test_loss_unif": 1.9411154985427856, "test_loss_unif_q@queue": 1.9411154985427856, "test_norm_diff": 0.03842584788799286, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.031087875366211, "test_query_norm": 1.7024343013763428, "test_queue_k_norm": 1.6896429061889648, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043644979596138, "test_stdq": 0.04390425235033035, "test_stdqueue_k": 0.04903624206781387, "test_stdqueue_q": 0.0 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.4876, "doc_norm": 1.6911, "encoder_q-embeddings": 19974.3867, "encoder_q-layer.0": 13730.2402, "encoder_q-layer.1": 15095.0986, "encoder_q-layer.10": 24964.8438, "encoder_q-layer.11": 58625.0938, "encoder_q-layer.2": 17140.4531, "encoder_q-layer.3": 16675.0918, "encoder_q-layer.4": 18460.2422, "encoder_q-layer.5": 19677.6582, "encoder_q-layer.6": 21131.6895, "encoder_q-layer.7": 22802.168, "encoder_q-layer.8": 25900.9824, "encoder_q-layer.9": 23960.7148, "epoch": 0.35, "inbatch_neg_score": 1.0336, "inbatch_pos_score": 1.6309, "learning_rate": 3.155263157894737e-05, "loss": 3.4876, "norm_diff": 0.0969, "norm_loss": 0.0, "num_token_doc": 66.872, "num_token_overlap": 11.6794, "num_token_query": 31.3672, "num_token_union": 65.168, "num_word_context": 202.4646, "num_word_doc": 49.9127, "num_word_query": 23.2998, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37960.492, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 1.0332, "query_norm": 1.5942, "queue_k_norm": 1.6919, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3672, "sent_len_1": 66.872, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.4125, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80100 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.5068, "doc_norm": 1.6897, "encoder_q-embeddings": 20249.6738, "encoder_q-layer.0": 14150.5938, "encoder_q-layer.1": 15115.8408, "encoder_q-layer.10": 24838.1074, "encoder_q-layer.11": 57044.3164, "encoder_q-layer.2": 17458.7422, "encoder_q-layer.3": 17408.8945, "encoder_q-layer.4": 18587.6406, "encoder_q-layer.5": 17971.2422, "encoder_q-layer.6": 19608.4941, "encoder_q-layer.7": 21404.4688, "encoder_q-layer.8": 25571.498, "encoder_q-layer.9": 23830.0312, "epoch": 0.35, "inbatch_neg_score": 1.0371, "inbatch_pos_score": 1.6318, "learning_rate": 3.152631578947369e-05, "loss": 3.5068, "norm_diff": 0.1021, "norm_loss": 0.0, "num_token_doc": 66.768, "num_token_overlap": 11.6575, "num_token_query": 31.3408, "num_token_union": 65.1429, "num_word_context": 202.3819, "num_word_doc": 49.8512, "num_word_query": 23.2859, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37406.57, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 1.0381, "query_norm": 1.5877, "queue_k_norm": 1.6915, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3408, "sent_len_1": 66.768, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.0838, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80200 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.5101, "doc_norm": 1.692, "encoder_q-embeddings": 9774.0615, "encoder_q-layer.0": 6610.4365, "encoder_q-layer.1": 6992.9844, "encoder_q-layer.10": 13426.5664, "encoder_q-layer.11": 29107.6758, "encoder_q-layer.2": 7565.2305, "encoder_q-layer.3": 7889.0, "encoder_q-layer.4": 8161.3071, "encoder_q-layer.5": 8544.3799, "encoder_q-layer.6": 9669.2852, "encoder_q-layer.7": 10470.2646, "encoder_q-layer.8": 13022.3086, "encoder_q-layer.9": 12076.1758, "epoch": 0.35, "inbatch_neg_score": 1.0398, "inbatch_pos_score": 1.6504, "learning_rate": 3.15e-05, "loss": 3.5101, "norm_diff": 0.0892, "norm_loss": 0.0, "num_token_doc": 66.7394, "num_token_overlap": 11.6638, "num_token_query": 31.3865, "num_token_union": 65.0899, "num_word_context": 202.3297, "num_word_doc": 49.7438, "num_word_query": 23.3106, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18445.4803, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.04, "query_norm": 1.6028, "queue_k_norm": 1.6927, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3865, "sent_len_1": 66.7394, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.305, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80300 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.5216, "doc_norm": 1.69, "encoder_q-embeddings": 10355.3555, "encoder_q-layer.0": 6923.5562, "encoder_q-layer.1": 7110.2378, "encoder_q-layer.10": 13434.3613, "encoder_q-layer.11": 30561.9648, "encoder_q-layer.2": 8128.5986, "encoder_q-layer.3": 8214.0986, "encoder_q-layer.4": 8676.9463, "encoder_q-layer.5": 8642.3203, "encoder_q-layer.6": 9597.5742, "encoder_q-layer.7": 11527.6748, "encoder_q-layer.8": 14515.1357, "encoder_q-layer.9": 13202.2832, "epoch": 0.35, "inbatch_neg_score": 1.0447, "inbatch_pos_score": 1.6426, "learning_rate": 3.147368421052632e-05, "loss": 3.5216, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.7736, "num_token_overlap": 11.6747, "num_token_query": 31.314, "num_token_union": 65.0723, "num_word_context": 202.2786, "num_word_doc": 49.8291, "num_word_query": 23.253, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19307.0931, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.0449, "query_norm": 1.6054, "queue_k_norm": 1.6937, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.314, "sent_len_1": 66.7736, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.3713, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80400 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.5252, "doc_norm": 1.6943, "encoder_q-embeddings": 4988.8159, "encoder_q-layer.0": 3415.8474, "encoder_q-layer.1": 3669.9382, "encoder_q-layer.10": 6627.437, "encoder_q-layer.11": 15349.8291, "encoder_q-layer.2": 3973.3623, "encoder_q-layer.3": 4040.0908, "encoder_q-layer.4": 4300.9678, "encoder_q-layer.5": 4355.2793, "encoder_q-layer.6": 5325.5312, "encoder_q-layer.7": 5818.0947, "encoder_q-layer.8": 7529.2412, "encoder_q-layer.9": 6621.4287, "epoch": 0.35, "inbatch_neg_score": 1.0444, "inbatch_pos_score": 1.6523, "learning_rate": 3.144736842105263e-05, "loss": 3.5252, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.8175, "num_token_overlap": 11.6439, "num_token_query": 31.2436, "num_token_union": 65.1103, "num_word_context": 202.5075, "num_word_doc": 49.8749, "num_word_query": 23.2007, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9894.531, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.0449, "query_norm": 1.6186, "queue_k_norm": 1.6947, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2436, "sent_len_1": 66.8175, "sent_len_max_0": 127.965, "sent_len_max_1": 208.0425, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80500 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4995, "doc_norm": 1.6923, "encoder_q-embeddings": 5072.0957, "encoder_q-layer.0": 3376.9849, "encoder_q-layer.1": 3605.0713, "encoder_q-layer.10": 6709.188, "encoder_q-layer.11": 14866.9834, "encoder_q-layer.2": 4071.5063, "encoder_q-layer.3": 4048.9722, "encoder_q-layer.4": 4250.1025, "encoder_q-layer.5": 4293.2412, "encoder_q-layer.6": 4619.3672, "encoder_q-layer.7": 5160.4375, "encoder_q-layer.8": 6686.1831, "encoder_q-layer.9": 6236.2896, "epoch": 0.35, "inbatch_neg_score": 1.0551, "inbatch_pos_score": 1.6514, "learning_rate": 3.142105263157895e-05, "loss": 3.4995, "norm_diff": 0.0701, "norm_loss": 0.0, "num_token_doc": 66.7682, "num_token_overlap": 11.7095, "num_token_query": 31.4292, "num_token_union": 65.1343, "num_word_context": 202.416, "num_word_doc": 49.8324, "num_word_query": 23.3362, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9437.8776, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.0547, "query_norm": 1.6221, "queue_k_norm": 1.6955, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4292, "sent_len_1": 66.7682, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.1587, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80600 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4996, "doc_norm": 1.6988, "encoder_q-embeddings": 5833.4121, "encoder_q-layer.0": 3934.2822, "encoder_q-layer.1": 4444.8486, "encoder_q-layer.10": 7187.4429, "encoder_q-layer.11": 15333.666, "encoder_q-layer.2": 5020.666, "encoder_q-layer.3": 5202.5601, "encoder_q-layer.4": 5579.1328, "encoder_q-layer.5": 5900.1792, "encoder_q-layer.6": 5989.5049, "encoder_q-layer.7": 6546.5059, "encoder_q-layer.8": 7577.9795, "encoder_q-layer.9": 6871.0088, "epoch": 0.35, "inbatch_neg_score": 1.0617, "inbatch_pos_score": 1.6748, "learning_rate": 3.139473684210527e-05, "loss": 3.4996, "norm_diff": 0.0617, "norm_loss": 0.0, "num_token_doc": 66.7189, "num_token_overlap": 11.6602, "num_token_query": 31.323, "num_token_union": 65.0975, "num_word_context": 202.4377, "num_word_doc": 49.7957, "num_word_query": 23.2537, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10508.105, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.0615, "query_norm": 1.6371, "queue_k_norm": 1.6971, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.323, "sent_len_1": 66.7189, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.78, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80700 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.5076, "doc_norm": 1.6978, "encoder_q-embeddings": 5247.5654, "encoder_q-layer.0": 3606.3108, "encoder_q-layer.1": 3740.3855, "encoder_q-layer.10": 6140.1025, "encoder_q-layer.11": 14560.8955, "encoder_q-layer.2": 4267.5415, "encoder_q-layer.3": 4421.0083, "encoder_q-layer.4": 4749.439, "encoder_q-layer.5": 4851.9751, "encoder_q-layer.6": 5079.686, "encoder_q-layer.7": 5822.6509, "encoder_q-layer.8": 6601.4766, "encoder_q-layer.9": 5858.063, "epoch": 0.35, "inbatch_neg_score": 1.0723, "inbatch_pos_score": 1.6807, "learning_rate": 3.136842105263158e-05, "loss": 3.5076, "norm_diff": 0.0602, "norm_loss": 0.0, "num_token_doc": 66.697, "num_token_overlap": 11.6131, "num_token_query": 31.2051, "num_token_union": 65.0263, "num_word_context": 202.0458, "num_word_doc": 49.7397, "num_word_query": 23.1512, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9482.4692, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.0723, "query_norm": 1.6376, "queue_k_norm": 1.6966, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2051, "sent_len_1": 66.697, "sent_len_max_0": 127.99, "sent_len_max_1": 210.6375, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 80800 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.5041, "doc_norm": 1.6979, "encoder_q-embeddings": 5987.1152, "encoder_q-layer.0": 4314.5132, "encoder_q-layer.1": 4422.0156, "encoder_q-layer.10": 6532.459, "encoder_q-layer.11": 15128.7803, "encoder_q-layer.2": 4918.4824, "encoder_q-layer.3": 4980.4141, "encoder_q-layer.4": 5321.1128, "encoder_q-layer.5": 5207.3999, "encoder_q-layer.6": 5752.9229, "encoder_q-layer.7": 6322.1636, "encoder_q-layer.8": 7088.0293, "encoder_q-layer.9": 6381.8203, "epoch": 0.35, "inbatch_neg_score": 1.0854, "inbatch_pos_score": 1.6748, "learning_rate": 3.1342105263157897e-05, "loss": 3.5041, "norm_diff": 0.0552, "norm_loss": 0.0, "num_token_doc": 66.7377, "num_token_overlap": 11.6836, "num_token_query": 31.3983, "num_token_union": 65.1035, "num_word_context": 202.0284, "num_word_doc": 49.7762, "num_word_query": 23.3104, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10292.2986, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.0859, "query_norm": 1.6427, "queue_k_norm": 1.6971, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3983, "sent_len_1": 66.7377, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.7612, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 80900 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.5091, "doc_norm": 1.701, "encoder_q-embeddings": 5123.1978, "encoder_q-layer.0": 3552.9021, "encoder_q-layer.1": 3773.3721, "encoder_q-layer.10": 6451.0244, "encoder_q-layer.11": 14798.5332, "encoder_q-layer.2": 4451.3975, "encoder_q-layer.3": 4636.4238, "encoder_q-layer.4": 5178.188, "encoder_q-layer.5": 5209.521, "encoder_q-layer.6": 5643.0889, "encoder_q-layer.7": 6456.2778, "encoder_q-layer.8": 8030.2236, "encoder_q-layer.9": 6575.3687, "epoch": 0.35, "inbatch_neg_score": 1.1008, "inbatch_pos_score": 1.7002, "learning_rate": 3.131578947368421e-05, "loss": 3.5091, "norm_diff": 0.0496, "norm_loss": 0.0, "num_token_doc": 66.7237, "num_token_overlap": 11.6923, "num_token_query": 31.4289, "num_token_union": 65.098, "num_word_context": 202.3975, "num_word_doc": 49.8035, "num_word_query": 23.3379, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9966.018, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1006, "query_norm": 1.6514, "queue_k_norm": 1.6993, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4289, "sent_len_1": 66.7237, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.3237, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 81000 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.5131, "doc_norm": 1.6986, "encoder_q-embeddings": 5063.2549, "encoder_q-layer.0": 3569.4373, "encoder_q-layer.1": 3966.7766, "encoder_q-layer.10": 6706.7344, "encoder_q-layer.11": 15082.5039, "encoder_q-layer.2": 4263.3418, "encoder_q-layer.3": 4392.2715, "encoder_q-layer.4": 4654.1318, "encoder_q-layer.5": 4636.9971, "encoder_q-layer.6": 5324.9189, "encoder_q-layer.7": 5857.3618, "encoder_q-layer.8": 6786.25, "encoder_q-layer.9": 6079.0273, "epoch": 0.35, "inbatch_neg_score": 1.1114, "inbatch_pos_score": 1.6895, "learning_rate": 3.1289473684210526e-05, "loss": 3.5131, "norm_diff": 0.0533, "norm_loss": 0.0, "num_token_doc": 66.6084, "num_token_overlap": 11.6474, "num_token_query": 31.4336, "num_token_union": 65.0656, "num_word_context": 202.2744, "num_word_doc": 49.6742, "num_word_query": 23.3344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9664.5559, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1104, "query_norm": 1.6453, "queue_k_norm": 1.7019, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4336, "sent_len_1": 66.6084, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.1488, "stdk": 0.0489, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 81100 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.4924, "doc_norm": 1.7026, "encoder_q-embeddings": 5470.6768, "encoder_q-layer.0": 3986.6675, "encoder_q-layer.1": 4264.7725, "encoder_q-layer.10": 6348.7891, "encoder_q-layer.11": 15162.2012, "encoder_q-layer.2": 4766.2363, "encoder_q-layer.3": 4794.5557, "encoder_q-layer.4": 4819.0366, "encoder_q-layer.5": 4822.5156, "encoder_q-layer.6": 5447.9434, "encoder_q-layer.7": 5996.6514, "encoder_q-layer.8": 6733.2451, "encoder_q-layer.9": 6108.8057, "epoch": 0.35, "inbatch_neg_score": 1.1178, "inbatch_pos_score": 1.6943, "learning_rate": 3.1263157894736845e-05, "loss": 3.4924, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.7027, "num_token_overlap": 11.6734, "num_token_query": 31.328, "num_token_union": 65.0557, "num_word_context": 201.9833, "num_word_doc": 49.7894, "num_word_query": 23.2821, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9930.3993, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1172, "query_norm": 1.6267, "queue_k_norm": 1.7027, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.328, "sent_len_1": 66.7027, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.2463, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 81200 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.5149, "doc_norm": 1.7016, "encoder_q-embeddings": 5052.9736, "encoder_q-layer.0": 3451.7456, "encoder_q-layer.1": 3655.4807, "encoder_q-layer.10": 6294.5723, "encoder_q-layer.11": 14544.9414, "encoder_q-layer.2": 4002.0347, "encoder_q-layer.3": 4039.1604, "encoder_q-layer.4": 4184.3135, "encoder_q-layer.5": 4333.3179, "encoder_q-layer.6": 4877.7671, "encoder_q-layer.7": 5504.5151, "encoder_q-layer.8": 6698.9658, "encoder_q-layer.9": 6218.6621, "epoch": 0.35, "inbatch_neg_score": 1.1215, "inbatch_pos_score": 1.707, "learning_rate": 3.1236842105263156e-05, "loss": 3.5149, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.6769, "num_token_overlap": 11.653, "num_token_query": 31.343, "num_token_union": 65.056, "num_word_context": 202.22, "num_word_doc": 49.7156, "num_word_query": 23.2841, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9358.6532, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1211, "query_norm": 1.6166, "queue_k_norm": 1.7053, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.343, "sent_len_1": 66.6769, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.1875, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 81300 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.503, "doc_norm": 1.7098, "encoder_q-embeddings": 5092.6436, "encoder_q-layer.0": 3496.9685, "encoder_q-layer.1": 3738.5347, "encoder_q-layer.10": 6955.561, "encoder_q-layer.11": 16154.6816, "encoder_q-layer.2": 4254.166, "encoder_q-layer.3": 4434.1748, "encoder_q-layer.4": 4692.6768, "encoder_q-layer.5": 4792.8867, "encoder_q-layer.6": 5274.6768, "encoder_q-layer.7": 5849.9404, "encoder_q-layer.8": 6952.8291, "encoder_q-layer.9": 6440.8184, "epoch": 0.35, "inbatch_neg_score": 1.1209, "inbatch_pos_score": 1.71, "learning_rate": 3.1210526315789475e-05, "loss": 3.503, "norm_diff": 0.0877, "norm_loss": 0.0, "num_token_doc": 66.8228, "num_token_overlap": 11.6479, "num_token_query": 31.3657, "num_token_union": 65.1473, "num_word_context": 202.2576, "num_word_doc": 49.8392, "num_word_query": 23.299, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9957.1217, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1211, "query_norm": 1.6221, "queue_k_norm": 1.7057, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3657, "sent_len_1": 66.8228, "sent_len_max_0": 127.9838, "sent_len_max_1": 211.12, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 81400 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4954, "doc_norm": 1.7122, "encoder_q-embeddings": 5838.6157, "encoder_q-layer.0": 4097.0317, "encoder_q-layer.1": 4688.9277, "encoder_q-layer.10": 6578.7832, "encoder_q-layer.11": 15219.1406, "encoder_q-layer.2": 5470.5581, "encoder_q-layer.3": 6013.7075, "encoder_q-layer.4": 6875.1196, "encoder_q-layer.5": 8188.2368, "encoder_q-layer.6": 7942.7007, "encoder_q-layer.7": 7928.5767, "encoder_q-layer.8": 7858.0645, "encoder_q-layer.9": 6586.9976, "epoch": 0.35, "inbatch_neg_score": 1.1193, "inbatch_pos_score": 1.707, "learning_rate": 3.118421052631579e-05, "loss": 3.4954, "norm_diff": 0.1001, "norm_loss": 0.0, "num_token_doc": 66.8013, "num_token_overlap": 11.6836, "num_token_query": 31.3159, "num_token_union": 65.1002, "num_word_context": 202.1267, "num_word_doc": 49.8468, "num_word_query": 23.2408, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11213.8623, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1191, "query_norm": 1.6121, "queue_k_norm": 1.7069, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3159, "sent_len_1": 66.8013, "sent_len_max_0": 127.98, "sent_len_max_1": 206.3887, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 81500 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.504, "doc_norm": 1.7069, "encoder_q-embeddings": 5141.792, "encoder_q-layer.0": 3409.0056, "encoder_q-layer.1": 3600.6799, "encoder_q-layer.10": 6465.7769, "encoder_q-layer.11": 16001.3838, "encoder_q-layer.2": 3989.2944, "encoder_q-layer.3": 4120.8008, "encoder_q-layer.4": 4412.874, "encoder_q-layer.5": 4268.3296, "encoder_q-layer.6": 4853.2749, "encoder_q-layer.7": 5612.3232, "encoder_q-layer.8": 6730.688, "encoder_q-layer.9": 6366.4326, "epoch": 0.35, "inbatch_neg_score": 1.1228, "inbatch_pos_score": 1.71, "learning_rate": 3.115789473684211e-05, "loss": 3.504, "norm_diff": 0.0936, "norm_loss": 0.0, "num_token_doc": 66.8343, "num_token_overlap": 11.6728, "num_token_query": 31.4026, "num_token_union": 65.2122, "num_word_context": 202.4471, "num_word_doc": 49.9011, "num_word_query": 23.3186, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9836.1221, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.123, "query_norm": 1.6133, "queue_k_norm": 1.7093, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4026, "sent_len_1": 66.8343, "sent_len_max_0": 127.985, "sent_len_max_1": 206.3537, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 81600 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.5051, "doc_norm": 1.7095, "encoder_q-embeddings": 5023.6113, "encoder_q-layer.0": 3421.845, "encoder_q-layer.1": 3622.3711, "encoder_q-layer.10": 7137.7266, "encoder_q-layer.11": 14903.835, "encoder_q-layer.2": 4084.9973, "encoder_q-layer.3": 4211.4683, "encoder_q-layer.4": 4392.9404, "encoder_q-layer.5": 4498.5537, "encoder_q-layer.6": 5122.8823, "encoder_q-layer.7": 5628.9951, "encoder_q-layer.8": 6743.541, "encoder_q-layer.9": 6309.5503, "epoch": 0.35, "inbatch_neg_score": 1.1281, "inbatch_pos_score": 1.7148, "learning_rate": 3.113157894736842e-05, "loss": 3.5051, "norm_diff": 0.0906, "norm_loss": 0.0, "num_token_doc": 66.7631, "num_token_overlap": 11.6236, "num_token_query": 31.2626, "num_token_union": 65.1141, "num_word_context": 202.1629, "num_word_doc": 49.8386, "num_word_query": 23.2196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9491.4443, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.127, "query_norm": 1.6189, "queue_k_norm": 1.7118, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2626, "sent_len_1": 66.7631, "sent_len_max_0": 127.9425, "sent_len_max_1": 207.6438, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 81700 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.5029, "doc_norm": 1.7116, "encoder_q-embeddings": 5304.0552, "encoder_q-layer.0": 3535.2485, "encoder_q-layer.1": 3740.6589, "encoder_q-layer.10": 6712.3784, "encoder_q-layer.11": 15142.9512, "encoder_q-layer.2": 4190.8135, "encoder_q-layer.3": 4426.2422, "encoder_q-layer.4": 4910.5596, "encoder_q-layer.5": 5075.3965, "encoder_q-layer.6": 5465.7334, "encoder_q-layer.7": 6029.5815, "encoder_q-layer.8": 6851.3428, "encoder_q-layer.9": 6601.5273, "epoch": 0.35, "inbatch_neg_score": 1.1242, "inbatch_pos_score": 1.7295, "learning_rate": 3.1105263157894735e-05, "loss": 3.5029, "norm_diff": 0.0894, "norm_loss": 0.0, "num_token_doc": 66.665, "num_token_overlap": 11.6667, "num_token_query": 31.3779, "num_token_union": 65.0752, "num_word_context": 202.2544, "num_word_doc": 49.7647, "num_word_query": 23.3096, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9883.0481, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.125, "query_norm": 1.6222, "queue_k_norm": 1.7104, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3779, "sent_len_1": 66.665, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.0662, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 81800 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.516, "doc_norm": 1.7098, "encoder_q-embeddings": 5147.8203, "encoder_q-layer.0": 3568.0669, "encoder_q-layer.1": 3757.4802, "encoder_q-layer.10": 6683.2607, "encoder_q-layer.11": 15154.2207, "encoder_q-layer.2": 4310.0166, "encoder_q-layer.3": 4514.7144, "encoder_q-layer.4": 4592.7251, "encoder_q-layer.5": 4750.2217, "encoder_q-layer.6": 5266.2046, "encoder_q-layer.7": 5788.4868, "encoder_q-layer.8": 6770.2817, "encoder_q-layer.9": 6210.543, "epoch": 0.36, "inbatch_neg_score": 1.1186, "inbatch_pos_score": 1.7021, "learning_rate": 3.107894736842105e-05, "loss": 3.516, "norm_diff": 0.0953, "norm_loss": 0.0, "num_token_doc": 66.7838, "num_token_overlap": 11.6336, "num_token_query": 31.2752, "num_token_union": 65.0968, "num_word_context": 202.2718, "num_word_doc": 49.8246, "num_word_query": 23.2439, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9802.889, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1182, "query_norm": 1.6145, "queue_k_norm": 1.7132, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2752, "sent_len_1": 66.7838, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.0387, "stdk": 0.0489, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 81900 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.5018, "doc_norm": 1.7164, "encoder_q-embeddings": 16976.1973, "encoder_q-layer.0": 12149.7822, "encoder_q-layer.1": 13157.6768, "encoder_q-layer.10": 6231.5962, "encoder_q-layer.11": 14597.749, "encoder_q-layer.2": 14081.6748, "encoder_q-layer.3": 11615.876, "encoder_q-layer.4": 9352.5869, "encoder_q-layer.5": 7664.0728, "encoder_q-layer.6": 6845.5874, "encoder_q-layer.7": 6461.7212, "encoder_q-layer.8": 7185.7803, "encoder_q-layer.9": 6346.3652, "epoch": 0.36, "inbatch_neg_score": 1.1183, "inbatch_pos_score": 1.7188, "learning_rate": 3.105263157894737e-05, "loss": 3.5018, "norm_diff": 0.0918, "norm_loss": 0.0, "num_token_doc": 66.8237, "num_token_overlap": 11.6535, "num_token_query": 31.306, "num_token_union": 65.1465, "num_word_context": 202.4125, "num_word_doc": 49.8811, "num_word_query": 23.2475, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16960.8466, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1182, "query_norm": 1.6246, "queue_k_norm": 1.7121, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.306, "sent_len_1": 66.8237, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.4938, "stdk": 0.0491, "stdq": 0.0436, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 82000 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.5041, "doc_norm": 1.7158, "encoder_q-embeddings": 5334.1289, "encoder_q-layer.0": 3680.4878, "encoder_q-layer.1": 3902.6123, "encoder_q-layer.10": 6377.4385, "encoder_q-layer.11": 15755.2109, "encoder_q-layer.2": 4451.27, "encoder_q-layer.3": 4586.3081, "encoder_q-layer.4": 4911.4741, "encoder_q-layer.5": 5008.1016, "encoder_q-layer.6": 5300.4482, "encoder_q-layer.7": 6075.9341, "encoder_q-layer.8": 7123.0244, "encoder_q-layer.9": 6327.1953, "epoch": 0.36, "inbatch_neg_score": 1.1223, "inbatch_pos_score": 1.7188, "learning_rate": 3.102631578947369e-05, "loss": 3.5041, "norm_diff": 0.0976, "norm_loss": 0.0, "num_token_doc": 66.7872, "num_token_overlap": 11.6313, "num_token_query": 31.227, "num_token_union": 65.0687, "num_word_context": 202.5282, "num_word_doc": 49.8064, "num_word_query": 23.1843, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9991.5772, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1221, "query_norm": 1.6182, "queue_k_norm": 1.7153, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.227, "sent_len_1": 66.7872, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.5213, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 82100 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4994, "doc_norm": 1.7196, "encoder_q-embeddings": 5002.252, "encoder_q-layer.0": 3486.9011, "encoder_q-layer.1": 3640.363, "encoder_q-layer.10": 6562.9614, "encoder_q-layer.11": 14487.0029, "encoder_q-layer.2": 4038.4487, "encoder_q-layer.3": 4146.7607, "encoder_q-layer.4": 4282.6738, "encoder_q-layer.5": 4382.4458, "encoder_q-layer.6": 4785.8657, "encoder_q-layer.7": 5688.9995, "encoder_q-layer.8": 6510.5527, "encoder_q-layer.9": 6193.6172, "epoch": 0.36, "inbatch_neg_score": 1.121, "inbatch_pos_score": 1.7158, "learning_rate": 3.1e-05, "loss": 3.4994, "norm_diff": 0.1012, "norm_loss": 0.0, "num_token_doc": 66.758, "num_token_overlap": 11.6622, "num_token_query": 31.358, "num_token_union": 65.1009, "num_word_context": 202.1345, "num_word_doc": 49.8256, "num_word_query": 23.2929, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9276.0784, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.1211, "query_norm": 1.6184, "queue_k_norm": 1.7148, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.358, "sent_len_1": 66.758, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.0437, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 82200 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.5024, "doc_norm": 1.7163, "encoder_q-embeddings": 6335.7622, "encoder_q-layer.0": 4479.7002, "encoder_q-layer.1": 4680.6396, "encoder_q-layer.10": 6446.311, "encoder_q-layer.11": 15114.5146, "encoder_q-layer.2": 5511.2529, "encoder_q-layer.3": 5560.5684, "encoder_q-layer.4": 6174.2212, "encoder_q-layer.5": 6121.4019, "encoder_q-layer.6": 6746.4951, "encoder_q-layer.7": 6877.2822, "encoder_q-layer.8": 6680.4556, "encoder_q-layer.9": 5966.5708, "epoch": 0.36, "inbatch_neg_score": 1.1259, "inbatch_pos_score": 1.7344, "learning_rate": 3.097368421052631e-05, "loss": 3.5024, "norm_diff": 0.0892, "norm_loss": 0.0, "num_token_doc": 66.8918, "num_token_overlap": 11.6847, "num_token_query": 31.3591, "num_token_union": 65.1953, "num_word_context": 202.4667, "num_word_doc": 49.9093, "num_word_query": 23.2879, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10640.2889, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.126, "query_norm": 1.6271, "queue_k_norm": 1.7161, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3591, "sent_len_1": 66.8918, "sent_len_max_0": 127.98, "sent_len_max_1": 210.7488, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 82300 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.5136, "doc_norm": 1.7155, "encoder_q-embeddings": 5313.3833, "encoder_q-layer.0": 3690.199, "encoder_q-layer.1": 3841.0496, "encoder_q-layer.10": 6675.1846, "encoder_q-layer.11": 15513.0986, "encoder_q-layer.2": 4367.792, "encoder_q-layer.3": 4550.3447, "encoder_q-layer.4": 4806.5229, "encoder_q-layer.5": 5066.7065, "encoder_q-layer.6": 5597.9326, "encoder_q-layer.7": 6304.6289, "encoder_q-layer.8": 7298.3193, "encoder_q-layer.9": 6442.8677, "epoch": 0.36, "inbatch_neg_score": 1.1261, "inbatch_pos_score": 1.709, "learning_rate": 3.094736842105263e-05, "loss": 3.5136, "norm_diff": 0.1007, "norm_loss": 0.0, "num_token_doc": 66.6786, "num_token_overlap": 11.6544, "num_token_query": 31.2995, "num_token_union": 65.0367, "num_word_context": 202.3831, "num_word_doc": 49.759, "num_word_query": 23.2572, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10130.2136, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.126, "query_norm": 1.6148, "queue_k_norm": 1.7167, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2995, "sent_len_1": 66.6786, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.2088, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 82400 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.4997, "doc_norm": 1.7185, "encoder_q-embeddings": 10316.6885, "encoder_q-layer.0": 7112.6792, "encoder_q-layer.1": 7669.6152, "encoder_q-layer.10": 12381.9863, "encoder_q-layer.11": 30427.625, "encoder_q-layer.2": 8488.0371, "encoder_q-layer.3": 9025.4424, "encoder_q-layer.4": 9739.1289, "encoder_q-layer.5": 9784.7275, "encoder_q-layer.6": 10346.6162, "encoder_q-layer.7": 11073.0889, "encoder_q-layer.8": 13696.958, "encoder_q-layer.9": 12361.2725, "epoch": 0.36, "inbatch_neg_score": 1.1262, "inbatch_pos_score": 1.7139, "learning_rate": 3.092105263157895e-05, "loss": 3.4997, "norm_diff": 0.105, "norm_loss": 0.0, "num_token_doc": 66.6785, "num_token_overlap": 11.6196, "num_token_query": 31.2562, "num_token_union": 65.0092, "num_word_context": 202.4285, "num_word_doc": 49.7389, "num_word_query": 23.1993, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19500.6621, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.125, "query_norm": 1.6135, "queue_k_norm": 1.7178, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2562, "sent_len_1": 66.6785, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.2063, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 82500 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.4963, "doc_norm": 1.7211, "encoder_q-embeddings": 10864.2539, "encoder_q-layer.0": 7341.0918, "encoder_q-layer.1": 7803.8569, "encoder_q-layer.10": 12434.8428, "encoder_q-layer.11": 30455.8301, "encoder_q-layer.2": 8808.4697, "encoder_q-layer.3": 8868.3232, "encoder_q-layer.4": 9721.1953, "encoder_q-layer.5": 9402.0732, "encoder_q-layer.6": 11098.7539, "encoder_q-layer.7": 12724.4619, "encoder_q-layer.8": 13636.9092, "encoder_q-layer.9": 12479.584, "epoch": 0.36, "inbatch_neg_score": 1.1291, "inbatch_pos_score": 1.7178, "learning_rate": 3.089473684210527e-05, "loss": 3.4963, "norm_diff": 0.1101, "norm_loss": 0.0, "num_token_doc": 66.8594, "num_token_overlap": 11.6417, "num_token_query": 31.1441, "num_token_union": 65.0358, "num_word_context": 202.3982, "num_word_doc": 49.8979, "num_word_query": 23.119, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19600.542, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1289, "query_norm": 1.611, "queue_k_norm": 1.7192, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.1441, "sent_len_1": 66.8594, "sent_len_max_0": 127.9525, "sent_len_max_1": 208.3825, "stdk": 0.0491, "stdq": 0.0426, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 82600 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.4915, "doc_norm": 1.7231, "encoder_q-embeddings": 9309.3242, "encoder_q-layer.0": 6466.7959, "encoder_q-layer.1": 6897.9937, "encoder_q-layer.10": 13867.373, "encoder_q-layer.11": 30129.625, "encoder_q-layer.2": 7491.5762, "encoder_q-layer.3": 7672.7031, "encoder_q-layer.4": 8088.9346, "encoder_q-layer.5": 8340.3604, "encoder_q-layer.6": 9991.4668, "encoder_q-layer.7": 10504.4141, "encoder_q-layer.8": 13170.9961, "encoder_q-layer.9": 13016.6689, "epoch": 0.36, "inbatch_neg_score": 1.1298, "inbatch_pos_score": 1.7334, "learning_rate": 3.086842105263158e-05, "loss": 3.4915, "norm_diff": 0.1061, "norm_loss": 0.0, "num_token_doc": 66.7828, "num_token_overlap": 11.6841, "num_token_query": 31.3961, "num_token_union": 65.1315, "num_word_context": 202.3254, "num_word_doc": 49.8439, "num_word_query": 23.3331, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18803.1003, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1299, "query_norm": 1.617, "queue_k_norm": 1.72, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3961, "sent_len_1": 66.7828, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.0412, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 82700 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4873, "doc_norm": 1.7214, "encoder_q-embeddings": 10535.0469, "encoder_q-layer.0": 7068.7466, "encoder_q-layer.1": 7331.563, "encoder_q-layer.10": 13309.9961, "encoder_q-layer.11": 30329.1152, "encoder_q-layer.2": 8363.375, "encoder_q-layer.3": 8725.9678, "encoder_q-layer.4": 9044.6602, "encoder_q-layer.5": 9120.9287, "encoder_q-layer.6": 10420.8252, "encoder_q-layer.7": 11041.707, "encoder_q-layer.8": 13365.5791, "encoder_q-layer.9": 13275.8379, "epoch": 0.36, "inbatch_neg_score": 1.1334, "inbatch_pos_score": 1.7246, "learning_rate": 3.084210526315789e-05, "loss": 3.4873, "norm_diff": 0.1045, "norm_loss": 0.0, "num_token_doc": 66.6552, "num_token_overlap": 11.6716, "num_token_query": 31.3639, "num_token_union": 65.0421, "num_word_context": 201.9782, "num_word_doc": 49.7605, "num_word_query": 23.2847, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19298.8187, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1318, "query_norm": 1.6169, "queue_k_norm": 1.7196, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3639, "sent_len_1": 66.6552, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.5488, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 82800 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.5091, "doc_norm": 1.7182, "encoder_q-embeddings": 10651.376, "encoder_q-layer.0": 7439.3472, "encoder_q-layer.1": 7970.1357, "encoder_q-layer.10": 12629.998, "encoder_q-layer.11": 29456.0684, "encoder_q-layer.2": 8879.7412, "encoder_q-layer.3": 9121.3252, "encoder_q-layer.4": 9806.9629, "encoder_q-layer.5": 10379.5322, "encoder_q-layer.6": 11190.5654, "encoder_q-layer.7": 11848.9473, "encoder_q-layer.8": 14226.5098, "encoder_q-layer.9": 12196.085, "epoch": 0.36, "inbatch_neg_score": 1.1327, "inbatch_pos_score": 1.7207, "learning_rate": 3.0815789473684216e-05, "loss": 3.5091, "norm_diff": 0.1051, "norm_loss": 0.0, "num_token_doc": 66.7327, "num_token_overlap": 11.6506, "num_token_query": 31.3443, "num_token_union": 65.099, "num_word_context": 202.275, "num_word_doc": 49.8258, "num_word_query": 23.2884, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19785.0981, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1328, "query_norm": 1.613, "queue_k_norm": 1.7221, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3443, "sent_len_1": 66.7327, "sent_len_max_0": 127.995, "sent_len_max_1": 207.6925, "stdk": 0.0489, "stdq": 0.0426, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 82900 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.4949, "doc_norm": 1.7237, "encoder_q-embeddings": 10455.6396, "encoder_q-layer.0": 7291.1533, "encoder_q-layer.1": 7448.7632, "encoder_q-layer.10": 12128.0811, "encoder_q-layer.11": 30229.0, "encoder_q-layer.2": 8302.5723, "encoder_q-layer.3": 8424.9492, "encoder_q-layer.4": 8760.4521, "encoder_q-layer.5": 8753.7236, "encoder_q-layer.6": 10134.6074, "encoder_q-layer.7": 11186.541, "encoder_q-layer.8": 12996.6777, "encoder_q-layer.9": 12209.96, "epoch": 0.36, "inbatch_neg_score": 1.1372, "inbatch_pos_score": 1.7344, "learning_rate": 3.078947368421053e-05, "loss": 3.4949, "norm_diff": 0.0953, "norm_loss": 0.0, "num_token_doc": 66.8602, "num_token_overlap": 11.6973, "num_token_query": 31.4232, "num_token_union": 65.1891, "num_word_context": 202.4077, "num_word_doc": 49.8875, "num_word_query": 23.3419, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19173.4071, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1367, "query_norm": 1.6284, "queue_k_norm": 1.7215, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4232, "sent_len_1": 66.8602, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.73, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 83000 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.5005, "doc_norm": 1.7193, "encoder_q-embeddings": 10023.5, "encoder_q-layer.0": 6663.356, "encoder_q-layer.1": 7017.5562, "encoder_q-layer.10": 12747.0371, "encoder_q-layer.11": 30461.6113, "encoder_q-layer.2": 7943.2979, "encoder_q-layer.3": 8367.6035, "encoder_q-layer.4": 8675.083, "encoder_q-layer.5": 8655.9463, "encoder_q-layer.6": 9870.3057, "encoder_q-layer.7": 11647.29, "encoder_q-layer.8": 14271.4561, "encoder_q-layer.9": 12777.8057, "epoch": 0.36, "inbatch_neg_score": 1.1361, "inbatch_pos_score": 1.7188, "learning_rate": 3.0763157894736846e-05, "loss": 3.5005, "norm_diff": 0.102, "norm_loss": 0.0, "num_token_doc": 66.7006, "num_token_overlap": 11.6816, "num_token_query": 31.332, "num_token_union": 65.062, "num_word_context": 202.3078, "num_word_doc": 49.781, "num_word_query": 23.2629, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19380.3966, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1367, "query_norm": 1.6174, "queue_k_norm": 1.7213, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.332, "sent_len_1": 66.7006, "sent_len_max_0": 127.985, "sent_len_max_1": 208.9688, "stdk": 0.0489, "stdq": 0.0426, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 83100 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4936, "doc_norm": 1.719, "encoder_q-embeddings": 9922.7529, "encoder_q-layer.0": 6558.6724, "encoder_q-layer.1": 7007.8345, "encoder_q-layer.10": 12269.3828, "encoder_q-layer.11": 29551.8984, "encoder_q-layer.2": 7996.9761, "encoder_q-layer.3": 8167.8232, "encoder_q-layer.4": 8707.8926, "encoder_q-layer.5": 9340.8691, "encoder_q-layer.6": 10427.6514, "encoder_q-layer.7": 11301.1055, "encoder_q-layer.8": 12919.9219, "encoder_q-layer.9": 12066.959, "epoch": 0.36, "inbatch_neg_score": 1.1376, "inbatch_pos_score": 1.7363, "learning_rate": 3.073684210526316e-05, "loss": 3.4936, "norm_diff": 0.0883, "norm_loss": 0.0, "num_token_doc": 66.6338, "num_token_overlap": 11.6798, "num_token_query": 31.4168, "num_token_union": 65.0668, "num_word_context": 202.1823, "num_word_doc": 49.6914, "num_word_query": 23.3257, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18736.1738, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1367, "query_norm": 1.6307, "queue_k_norm": 1.723, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4168, "sent_len_1": 66.6338, "sent_len_max_0": 127.99, "sent_len_max_1": 209.1113, "stdk": 0.0488, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 83200 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4752, "doc_norm": 1.722, "encoder_q-embeddings": 9749.4893, "encoder_q-layer.0": 6989.396, "encoder_q-layer.1": 7335.6641, "encoder_q-layer.10": 12809.4717, "encoder_q-layer.11": 29430.0664, "encoder_q-layer.2": 8206.8887, "encoder_q-layer.3": 8372.3828, "encoder_q-layer.4": 8910.8984, "encoder_q-layer.5": 8824.6484, "encoder_q-layer.6": 10019.0674, "encoder_q-layer.7": 11361.292, "encoder_q-layer.8": 13615.8936, "encoder_q-layer.9": 12429.3672, "epoch": 0.36, "inbatch_neg_score": 1.1414, "inbatch_pos_score": 1.7422, "learning_rate": 3.0710526315789476e-05, "loss": 3.4752, "norm_diff": 0.0872, "norm_loss": 0.0, "num_token_doc": 66.9417, "num_token_overlap": 11.7002, "num_token_query": 31.3793, "num_token_union": 65.1949, "num_word_context": 202.4449, "num_word_doc": 49.9024, "num_word_query": 23.3092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18990.392, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1416, "query_norm": 1.6349, "queue_k_norm": 1.7228, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3793, "sent_len_1": 66.9417, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.33, "stdk": 0.049, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 83300 }, { "accuracy": 48.3398, "active_queue_size": 16384.0, "cl_loss": 3.4876, "doc_norm": 1.7231, "encoder_q-embeddings": 10708.4697, "encoder_q-layer.0": 7447.5947, "encoder_q-layer.1": 7823.042, "encoder_q-layer.10": 13026.3203, "encoder_q-layer.11": 30630.4375, "encoder_q-layer.2": 8610.7207, "encoder_q-layer.3": 9125.1826, "encoder_q-layer.4": 9458.002, "encoder_q-layer.5": 9827.9248, "encoder_q-layer.6": 10928.7354, "encoder_q-layer.7": 12664.7432, "encoder_q-layer.8": 14147.2588, "encoder_q-layer.9": 12484.1572, "epoch": 0.36, "inbatch_neg_score": 1.1471, "inbatch_pos_score": 1.7344, "learning_rate": 3.0684210526315795e-05, "loss": 3.4876, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.8017, "num_token_overlap": 11.6993, "num_token_query": 31.4753, "num_token_union": 65.2145, "num_word_context": 202.442, "num_word_doc": 49.8667, "num_word_query": 23.3942, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20151.3975, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1465, "query_norm": 1.6443, "queue_k_norm": 1.7244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4753, "sent_len_1": 66.8017, "sent_len_max_0": 127.995, "sent_len_max_1": 205.94, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 83400 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4984, "doc_norm": 1.7257, "encoder_q-embeddings": 11283.9863, "encoder_q-layer.0": 7613.3555, "encoder_q-layer.1": 8105.4678, "encoder_q-layer.10": 14859.6494, "encoder_q-layer.11": 30567.2559, "encoder_q-layer.2": 9050.0215, "encoder_q-layer.3": 9024.5605, "encoder_q-layer.4": 9638.4307, "encoder_q-layer.5": 9793.0781, "encoder_q-layer.6": 11026.0391, "encoder_q-layer.7": 12896.9297, "encoder_q-layer.8": 14935.4941, "encoder_q-layer.9": 13214.8223, "epoch": 0.36, "inbatch_neg_score": 1.1532, "inbatch_pos_score": 1.7471, "learning_rate": 3.0657894736842106e-05, "loss": 3.4984, "norm_diff": 0.0725, "norm_loss": 0.0, "num_token_doc": 66.8477, "num_token_overlap": 11.6639, "num_token_query": 31.4072, "num_token_union": 65.1924, "num_word_context": 202.402, "num_word_doc": 49.8883, "num_word_query": 23.3347, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20328.097, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1543, "query_norm": 1.6532, "queue_k_norm": 1.7241, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4072, "sent_len_1": 66.8477, "sent_len_max_0": 127.99, "sent_len_max_1": 208.615, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 83500 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.5013, "doc_norm": 1.7218, "encoder_q-embeddings": 11665.4385, "encoder_q-layer.0": 8445.0127, "encoder_q-layer.1": 8983.9492, "encoder_q-layer.10": 11930.6816, "encoder_q-layer.11": 28503.2656, "encoder_q-layer.2": 9865.8193, "encoder_q-layer.3": 9628.1045, "encoder_q-layer.4": 9409.7432, "encoder_q-layer.5": 9406.8809, "encoder_q-layer.6": 10394.8848, "encoder_q-layer.7": 11246.0146, "encoder_q-layer.8": 13171.1963, "encoder_q-layer.9": 12116.1611, "epoch": 0.36, "inbatch_neg_score": 1.1605, "inbatch_pos_score": 1.7715, "learning_rate": 3.0631578947368425e-05, "loss": 3.5013, "norm_diff": 0.0522, "norm_loss": 0.0, "num_token_doc": 66.7466, "num_token_overlap": 11.6663, "num_token_query": 31.3009, "num_token_union": 65.0305, "num_word_context": 202.1161, "num_word_doc": 49.7684, "num_word_query": 23.2384, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19390.6313, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1602, "query_norm": 1.6695, "queue_k_norm": 1.7251, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3009, "sent_len_1": 66.7466, "sent_len_max_0": 127.9488, "sent_len_max_1": 209.2587, "stdk": 0.0489, "stdq": 0.0438, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 83600 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4887, "doc_norm": 1.7268, "encoder_q-embeddings": 10397.4326, "encoder_q-layer.0": 7017.0361, "encoder_q-layer.1": 7607.7612, "encoder_q-layer.10": 12620.3525, "encoder_q-layer.11": 29851.791, "encoder_q-layer.2": 8381.4004, "encoder_q-layer.3": 8428.4551, "encoder_q-layer.4": 9074.0176, "encoder_q-layer.5": 9305.0488, "encoder_q-layer.6": 10155.2656, "encoder_q-layer.7": 11393.0674, "encoder_q-layer.8": 13699.4482, "encoder_q-layer.9": 12575.3115, "epoch": 0.36, "inbatch_neg_score": 1.1706, "inbatch_pos_score": 1.7598, "learning_rate": 3.0605263157894736e-05, "loss": 3.4887, "norm_diff": 0.0613, "norm_loss": 0.0, "num_token_doc": 66.6219, "num_token_overlap": 11.7067, "num_token_query": 31.4375, "num_token_union": 65.062, "num_word_context": 202.3065, "num_word_doc": 49.7257, "num_word_query": 23.3535, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19109.2273, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1709, "query_norm": 1.6655, "queue_k_norm": 1.7282, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4375, "sent_len_1": 66.6219, "sent_len_max_0": 127.985, "sent_len_max_1": 208.7438, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 83700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4898, "doc_norm": 1.7299, "encoder_q-embeddings": 10337.5684, "encoder_q-layer.0": 6907.9458, "encoder_q-layer.1": 7382.1396, "encoder_q-layer.10": 13765.6924, "encoder_q-layer.11": 29464.5137, "encoder_q-layer.2": 7992.9482, "encoder_q-layer.3": 8285.2285, "encoder_q-layer.4": 8816.3545, "encoder_q-layer.5": 9218.0342, "encoder_q-layer.6": 10520.2148, "encoder_q-layer.7": 11517.248, "encoder_q-layer.8": 13429.5039, "encoder_q-layer.9": 12094.8369, "epoch": 0.36, "inbatch_neg_score": 1.1857, "inbatch_pos_score": 1.7861, "learning_rate": 3.0578947368421054e-05, "loss": 3.4898, "norm_diff": 0.0566, "norm_loss": 0.0, "num_token_doc": 66.695, "num_token_overlap": 11.666, "num_token_query": 31.441, "num_token_union": 65.1364, "num_word_context": 202.4392, "num_word_doc": 49.7674, "num_word_query": 23.3579, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19254.5792, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1855, "query_norm": 1.6733, "queue_k_norm": 1.73, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.441, "sent_len_1": 66.695, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.8925, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 83800 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.4908, "doc_norm": 1.7305, "encoder_q-embeddings": 10324.0254, "encoder_q-layer.0": 7121.0713, "encoder_q-layer.1": 7383.2524, "encoder_q-layer.10": 13301.8555, "encoder_q-layer.11": 29504.2324, "encoder_q-layer.2": 8162.7837, "encoder_q-layer.3": 8446.6982, "encoder_q-layer.4": 9055.0479, "encoder_q-layer.5": 9129.5205, "encoder_q-layer.6": 10160.1445, "encoder_q-layer.7": 11047.082, "encoder_q-layer.8": 13568.2891, "encoder_q-layer.9": 12643.9756, "epoch": 0.36, "inbatch_neg_score": 1.1974, "inbatch_pos_score": 1.7939, "learning_rate": 3.055263157894737e-05, "loss": 3.4908, "norm_diff": 0.0341, "norm_loss": 0.0, "num_token_doc": 66.7571, "num_token_overlap": 11.6889, "num_token_query": 31.5432, "num_token_union": 65.2297, "num_word_context": 202.5017, "num_word_doc": 49.8001, "num_word_query": 23.4383, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19115.5574, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.1963, "query_norm": 1.6964, "queue_k_norm": 1.7286, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5432, "sent_len_1": 66.7571, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.2025, "stdk": 0.0491, "stdq": 0.0438, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 83900 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4956, "doc_norm": 1.73, "encoder_q-embeddings": 5600.3545, "encoder_q-layer.0": 3957.1104, "encoder_q-layer.1": 4265.436, "encoder_q-layer.10": 6481.6143, "encoder_q-layer.11": 14877.9629, "encoder_q-layer.2": 4686.1685, "encoder_q-layer.3": 5005.9805, "encoder_q-layer.4": 5284.7588, "encoder_q-layer.5": 5474.6382, "encoder_q-layer.6": 5696.1562, "encoder_q-layer.7": 6394.2192, "encoder_q-layer.8": 7246.041, "encoder_q-layer.9": 6425.3936, "epoch": 0.36, "inbatch_neg_score": 1.2133, "inbatch_pos_score": 1.8086, "learning_rate": 3.0526315789473684e-05, "loss": 3.4956, "norm_diff": 0.042, "norm_loss": 0.0, "num_token_doc": 66.9228, "num_token_overlap": 11.7031, "num_token_query": 31.4275, "num_token_union": 65.2471, "num_word_context": 202.5594, "num_word_doc": 49.9629, "num_word_query": 23.3453, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10194.4847, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2129, "query_norm": 1.688, "queue_k_norm": 1.7319, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4275, "sent_len_1": 66.9228, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.8525, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84000 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.4864, "doc_norm": 1.7318, "encoder_q-embeddings": 4758.4736, "encoder_q-layer.0": 3299.0527, "encoder_q-layer.1": 3466.5571, "encoder_q-layer.10": 6533.5884, "encoder_q-layer.11": 15058.8975, "encoder_q-layer.2": 3912.8081, "encoder_q-layer.3": 4025.4976, "encoder_q-layer.4": 4138.3892, "encoder_q-layer.5": 4340.0654, "encoder_q-layer.6": 5195.6118, "encoder_q-layer.7": 5884.8853, "encoder_q-layer.8": 7000.7495, "encoder_q-layer.9": 6359.2495, "epoch": 0.36, "inbatch_neg_score": 1.2234, "inbatch_pos_score": 1.8213, "learning_rate": 3.05e-05, "loss": 3.4864, "norm_diff": 0.037, "norm_loss": 0.0, "num_token_doc": 66.8287, "num_token_overlap": 11.6978, "num_token_query": 31.4344, "num_token_union": 65.195, "num_word_context": 202.3523, "num_word_doc": 49.8572, "num_word_query": 23.3423, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9609.2325, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2227, "query_norm": 1.6949, "queue_k_norm": 1.7351, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4344, "sent_len_1": 66.8287, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.5712, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 84100 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4849, "doc_norm": 1.738, "encoder_q-embeddings": 5286.5415, "encoder_q-layer.0": 3589.4514, "encoder_q-layer.1": 3799.5881, "encoder_q-layer.10": 7437.2788, "encoder_q-layer.11": 15307.3047, "encoder_q-layer.2": 4426.0, "encoder_q-layer.3": 4449.0176, "encoder_q-layer.4": 4685.0874, "encoder_q-layer.5": 4761.7222, "encoder_q-layer.6": 5214.6973, "encoder_q-layer.7": 5550.8389, "encoder_q-layer.8": 6909.895, "encoder_q-layer.9": 6450.7749, "epoch": 0.37, "inbatch_neg_score": 1.2343, "inbatch_pos_score": 1.8467, "learning_rate": 3.0473684210526314e-05, "loss": 3.4849, "norm_diff": 0.0449, "norm_loss": 0.0, "num_token_doc": 66.8712, "num_token_overlap": 11.6958, "num_token_query": 31.3881, "num_token_union": 65.1906, "num_word_context": 202.5092, "num_word_doc": 49.9142, "num_word_query": 23.311, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10006.9695, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2344, "query_norm": 1.6932, "queue_k_norm": 1.7361, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3881, "sent_len_1": 66.8712, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.0975, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84200 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4887, "doc_norm": 1.7401, "encoder_q-embeddings": 4881.1475, "encoder_q-layer.0": 3299.7808, "encoder_q-layer.1": 3499.3855, "encoder_q-layer.10": 6030.4282, "encoder_q-layer.11": 15610.0928, "encoder_q-layer.2": 3858.6357, "encoder_q-layer.3": 4018.2725, "encoder_q-layer.4": 4177.2173, "encoder_q-layer.5": 4277.1855, "encoder_q-layer.6": 4749.8281, "encoder_q-layer.7": 5345.187, "encoder_q-layer.8": 6354.021, "encoder_q-layer.9": 5921.0845, "epoch": 0.37, "inbatch_neg_score": 1.24, "inbatch_pos_score": 1.8242, "learning_rate": 3.0447368421052636e-05, "loss": 3.4887, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.8285, "num_token_overlap": 11.6612, "num_token_query": 31.2973, "num_token_union": 65.0864, "num_word_context": 202.4218, "num_word_doc": 49.8245, "num_word_query": 23.2499, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9422.3644, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2393, "query_norm": 1.6677, "queue_k_norm": 1.7371, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2973, "sent_len_1": 66.8285, "sent_len_max_0": 127.985, "sent_len_max_1": 211.9725, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84300 }, { "accuracy": 47.4609, "active_queue_size": 16384.0, "cl_loss": 3.4777, "doc_norm": 1.7378, "encoder_q-embeddings": 5219.6016, "encoder_q-layer.0": 3575.114, "encoder_q-layer.1": 3827.2351, "encoder_q-layer.10": 7085.8979, "encoder_q-layer.11": 16156.7568, "encoder_q-layer.2": 4280.6494, "encoder_q-layer.3": 4399.1035, "encoder_q-layer.4": 4524.957, "encoder_q-layer.5": 4698.9644, "encoder_q-layer.6": 5253.3291, "encoder_q-layer.7": 5823.2856, "encoder_q-layer.8": 7166.3472, "encoder_q-layer.9": 6849.665, "epoch": 0.37, "inbatch_neg_score": 1.2481, "inbatch_pos_score": 1.8301, "learning_rate": 3.042105263157895e-05, "loss": 3.4777, "norm_diff": 0.07, "norm_loss": 0.0, "num_token_doc": 66.8289, "num_token_overlap": 11.6867, "num_token_query": 31.3671, "num_token_union": 65.1316, "num_word_context": 202.0655, "num_word_doc": 49.8499, "num_word_query": 23.294, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9951.762, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.248, "query_norm": 1.6678, "queue_k_norm": 1.7395, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3671, "sent_len_1": 66.8289, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.7413, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84400 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4923, "doc_norm": 1.7351, "encoder_q-embeddings": 5223.9429, "encoder_q-layer.0": 3506.8196, "encoder_q-layer.1": 3835.2412, "encoder_q-layer.10": 6274.8022, "encoder_q-layer.11": 15915.4561, "encoder_q-layer.2": 4172.8408, "encoder_q-layer.3": 4169.9438, "encoder_q-layer.4": 4492.5322, "encoder_q-layer.5": 4485.6011, "encoder_q-layer.6": 5116.0845, "encoder_q-layer.7": 5659.1978, "encoder_q-layer.8": 6984.1943, "encoder_q-layer.9": 6194.2041, "epoch": 0.37, "inbatch_neg_score": 1.2536, "inbatch_pos_score": 1.8477, "learning_rate": 3.0394736842105266e-05, "loss": 3.4923, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.6968, "num_token_overlap": 11.6869, "num_token_query": 31.4051, "num_token_union": 65.1062, "num_word_context": 202.3032, "num_word_doc": 49.8022, "num_word_query": 23.3176, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9996.7673, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2529, "query_norm": 1.667, "queue_k_norm": 1.7417, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4051, "sent_len_1": 66.6968, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.12, "stdk": 0.0487, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84500 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.4926, "doc_norm": 1.7425, "encoder_q-embeddings": 5129.272, "encoder_q-layer.0": 3385.6704, "encoder_q-layer.1": 3551.8496, "encoder_q-layer.10": 7414.6958, "encoder_q-layer.11": 16139.7236, "encoder_q-layer.2": 3936.7766, "encoder_q-layer.3": 4119.5449, "encoder_q-layer.4": 4573.272, "encoder_q-layer.5": 4403.0361, "encoder_q-layer.6": 5042.6782, "encoder_q-layer.7": 5911.3989, "encoder_q-layer.8": 7476.2041, "encoder_q-layer.9": 6988.4287, "epoch": 0.37, "inbatch_neg_score": 1.256, "inbatch_pos_score": 1.8447, "learning_rate": 3.036842105263158e-05, "loss": 3.4926, "norm_diff": 0.0714, "norm_loss": 0.0, "num_token_doc": 66.8149, "num_token_overlap": 11.6753, "num_token_query": 31.3609, "num_token_union": 65.1635, "num_word_context": 202.327, "num_word_doc": 49.859, "num_word_query": 23.3049, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10099.8939, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2559, "query_norm": 1.6711, "queue_k_norm": 1.7448, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3609, "sent_len_1": 66.8149, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.7725, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 84600 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.4894, "doc_norm": 1.746, "encoder_q-embeddings": 5013.8188, "encoder_q-layer.0": 3554.051, "encoder_q-layer.1": 3905.2456, "encoder_q-layer.10": 6907.3438, "encoder_q-layer.11": 15121.9092, "encoder_q-layer.2": 4465.9609, "encoder_q-layer.3": 4509.9463, "encoder_q-layer.4": 4753.9033, "encoder_q-layer.5": 4784.9556, "encoder_q-layer.6": 5260.8882, "encoder_q-layer.7": 6527.4829, "encoder_q-layer.8": 7361.2974, "encoder_q-layer.9": 6692.1284, "epoch": 0.37, "inbatch_neg_score": 1.2577, "inbatch_pos_score": 1.8516, "learning_rate": 3.0342105263157893e-05, "loss": 3.4894, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.731, "num_token_overlap": 11.671, "num_token_query": 31.45, "num_token_union": 65.1653, "num_word_context": 202.2606, "num_word_doc": 49.8098, "num_word_query": 23.3631, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9783.1602, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2578, "query_norm": 1.6711, "queue_k_norm": 1.7441, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.45, "sent_len_1": 66.731, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.6325, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 84700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4748, "doc_norm": 1.7484, "encoder_q-embeddings": 5317.7881, "encoder_q-layer.0": 3580.3564, "encoder_q-layer.1": 3790.4521, "encoder_q-layer.10": 6719.3784, "encoder_q-layer.11": 16430.7344, "encoder_q-layer.2": 4238.54, "encoder_q-layer.3": 4359.9434, "encoder_q-layer.4": 4664.1777, "encoder_q-layer.5": 4916.3413, "encoder_q-layer.6": 5498.0757, "encoder_q-layer.7": 6407.1201, "encoder_q-layer.8": 7686.6309, "encoder_q-layer.9": 6691.3335, "epoch": 0.37, "inbatch_neg_score": 1.2561, "inbatch_pos_score": 1.8613, "learning_rate": 3.0315789473684214e-05, "loss": 3.4748, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.9489, "num_token_overlap": 11.679, "num_token_query": 31.4852, "num_token_union": 65.3048, "num_word_context": 202.5402, "num_word_doc": 49.9368, "num_word_query": 23.3866, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10266.1381, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2559, "query_norm": 1.6748, "queue_k_norm": 1.7472, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4852, "sent_len_1": 66.9489, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.9638, "stdk": 0.0491, "stdq": 0.0437, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84800 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.4783, "doc_norm": 1.7513, "encoder_q-embeddings": 4983.9468, "encoder_q-layer.0": 3505.0261, "encoder_q-layer.1": 3824.9539, "encoder_q-layer.10": 6921.3579, "encoder_q-layer.11": 16622.5762, "encoder_q-layer.2": 4482.6113, "encoder_q-layer.3": 4516.939, "encoder_q-layer.4": 4662.0991, "encoder_q-layer.5": 4665.9375, "encoder_q-layer.6": 4890.2417, "encoder_q-layer.7": 5606.23, "encoder_q-layer.8": 6838.248, "encoder_q-layer.9": 6230.9785, "epoch": 0.37, "inbatch_neg_score": 1.262, "inbatch_pos_score": 1.8594, "learning_rate": 3.028947368421053e-05, "loss": 3.4783, "norm_diff": 0.084, "norm_loss": 0.0, "num_token_doc": 66.7462, "num_token_overlap": 11.6491, "num_token_query": 31.3097, "num_token_union": 65.0637, "num_word_context": 202.1296, "num_word_doc": 49.8106, "num_word_query": 23.244, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10034.4415, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2607, "query_norm": 1.6673, "queue_k_norm": 1.7491, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3097, "sent_len_1": 66.7462, "sent_len_max_0": 127.945, "sent_len_max_1": 209.71, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 84900 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.4888, "doc_norm": 1.7509, "encoder_q-embeddings": 5138.4468, "encoder_q-layer.0": 3468.5017, "encoder_q-layer.1": 3706.8835, "encoder_q-layer.10": 6548.4443, "encoder_q-layer.11": 15744.6943, "encoder_q-layer.2": 4148.2607, "encoder_q-layer.3": 4469.2949, "encoder_q-layer.4": 4654.8462, "encoder_q-layer.5": 4748.5049, "encoder_q-layer.6": 5590.0073, "encoder_q-layer.7": 6478.0811, "encoder_q-layer.8": 7459.7241, "encoder_q-layer.9": 6410.7471, "epoch": 0.37, "inbatch_neg_score": 1.2625, "inbatch_pos_score": 1.8652, "learning_rate": 3.0263157894736844e-05, "loss": 3.4888, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.7477, "num_token_overlap": 11.6578, "num_token_query": 31.3657, "num_token_union": 65.1401, "num_word_context": 202.2721, "num_word_doc": 49.8314, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10080.4412, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2627, "query_norm": 1.681, "queue_k_norm": 1.7495, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3657, "sent_len_1": 66.7477, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.4038, "stdk": 0.0491, "stdq": 0.0439, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 85000 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.5095, "doc_norm": 1.7506, "encoder_q-embeddings": 5135.8838, "encoder_q-layer.0": 3581.5056, "encoder_q-layer.1": 3748.9204, "encoder_q-layer.10": 6084.835, "encoder_q-layer.11": 14399.0078, "encoder_q-layer.2": 4160.2661, "encoder_q-layer.3": 4384.9531, "encoder_q-layer.4": 4855.1934, "encoder_q-layer.5": 4614.3735, "encoder_q-layer.6": 5297.8374, "encoder_q-layer.7": 5842.9414, "encoder_q-layer.8": 6644.4775, "encoder_q-layer.9": 6010.7856, "epoch": 0.37, "inbatch_neg_score": 1.2632, "inbatch_pos_score": 1.8623, "learning_rate": 3.023684210526316e-05, "loss": 3.5095, "norm_diff": 0.0902, "norm_loss": 0.0, "num_token_doc": 66.8602, "num_token_overlap": 11.654, "num_token_query": 31.3227, "num_token_union": 65.1522, "num_word_context": 202.4573, "num_word_doc": 49.8844, "num_word_query": 23.2539, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9523.0496, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2637, "query_norm": 1.6604, "queue_k_norm": 1.7503, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3227, "sent_len_1": 66.8602, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.195, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 85100 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4884, "doc_norm": 1.7538, "encoder_q-embeddings": 4955.9873, "encoder_q-layer.0": 3573.8701, "encoder_q-layer.1": 3556.9258, "encoder_q-layer.10": 6612.2896, "encoder_q-layer.11": 15015.8135, "encoder_q-layer.2": 3914.4888, "encoder_q-layer.3": 3989.7056, "encoder_q-layer.4": 4269.1572, "encoder_q-layer.5": 4479.9189, "encoder_q-layer.6": 5074.3394, "encoder_q-layer.7": 6087.2388, "encoder_q-layer.8": 7279.8613, "encoder_q-layer.9": 6454.6675, "epoch": 0.37, "inbatch_neg_score": 1.2664, "inbatch_pos_score": 1.8535, "learning_rate": 3.021052631578947e-05, "loss": 3.4884, "norm_diff": 0.096, "norm_loss": 0.0, "num_token_doc": 66.7109, "num_token_overlap": 11.6668, "num_token_query": 31.3763, "num_token_union": 65.0884, "num_word_context": 202.3616, "num_word_doc": 49.7665, "num_word_query": 23.3111, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9664.2913, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2656, "query_norm": 1.6578, "queue_k_norm": 1.7521, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3763, "sent_len_1": 66.7109, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.5025, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 85200 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.4852, "doc_norm": 1.7501, "encoder_q-embeddings": 6325.1484, "encoder_q-layer.0": 4254.7046, "encoder_q-layer.1": 4756.2344, "encoder_q-layer.10": 6623.1455, "encoder_q-layer.11": 14434.4639, "encoder_q-layer.2": 5493.1543, "encoder_q-layer.3": 5773.6196, "encoder_q-layer.4": 6720.5068, "encoder_q-layer.5": 7539.2217, "encoder_q-layer.6": 8118.9736, "encoder_q-layer.7": 8367.6699, "encoder_q-layer.8": 8119.6665, "encoder_q-layer.9": 6514.2578, "epoch": 0.37, "inbatch_neg_score": 1.2728, "inbatch_pos_score": 1.8682, "learning_rate": 3.0184210526315793e-05, "loss": 3.4852, "norm_diff": 0.0817, "norm_loss": 0.0, "num_token_doc": 66.793, "num_token_overlap": 11.6915, "num_token_query": 31.4314, "num_token_union": 65.1567, "num_word_context": 202.182, "num_word_doc": 49.8458, "num_word_query": 23.3466, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11036.8303, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2725, "query_norm": 1.6684, "queue_k_norm": 1.7554, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4314, "sent_len_1": 66.793, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.745, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 85300 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4759, "doc_norm": 1.7565, "encoder_q-embeddings": 5138.2788, "encoder_q-layer.0": 3465.2451, "encoder_q-layer.1": 3678.1541, "encoder_q-layer.10": 6239.1006, "encoder_q-layer.11": 14534.6816, "encoder_q-layer.2": 4158.2896, "encoder_q-layer.3": 4205.7725, "encoder_q-layer.4": 4488.2305, "encoder_q-layer.5": 4592.5933, "encoder_q-layer.6": 5120.2119, "encoder_q-layer.7": 5544.8911, "encoder_q-layer.8": 6667.5537, "encoder_q-layer.9": 6239.1089, "epoch": 0.37, "inbatch_neg_score": 1.2721, "inbatch_pos_score": 1.8906, "learning_rate": 3.0157894736842108e-05, "loss": 3.4759, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.7689, "num_token_overlap": 11.6692, "num_token_query": 31.3422, "num_token_union": 65.1187, "num_word_context": 202.2991, "num_word_doc": 49.8213, "num_word_query": 23.287, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9517.4648, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2725, "query_norm": 1.6735, "queue_k_norm": 1.7566, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3422, "sent_len_1": 66.7689, "sent_len_max_0": 127.9488, "sent_len_max_1": 208.1687, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 85400 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4986, "doc_norm": 1.7532, "encoder_q-embeddings": 5801.6841, "encoder_q-layer.0": 3759.1196, "encoder_q-layer.1": 3968.0229, "encoder_q-layer.10": 6305.3145, "encoder_q-layer.11": 14659.9834, "encoder_q-layer.2": 4404.9136, "encoder_q-layer.3": 4494.4233, "encoder_q-layer.4": 4868.5879, "encoder_q-layer.5": 4930.6582, "encoder_q-layer.6": 5584.2119, "encoder_q-layer.7": 6154.7949, "encoder_q-layer.8": 6973.7344, "encoder_q-layer.9": 6167.3911, "epoch": 0.37, "inbatch_neg_score": 1.2755, "inbatch_pos_score": 1.8662, "learning_rate": 3.0131578947368423e-05, "loss": 3.4986, "norm_diff": 0.0858, "norm_loss": 0.0, "num_token_doc": 66.64, "num_token_overlap": 11.6657, "num_token_query": 31.3114, "num_token_union": 65.0239, "num_word_context": 202.148, "num_word_doc": 49.7295, "num_word_query": 23.2587, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9946.2215, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2754, "query_norm": 1.6674, "queue_k_norm": 1.7543, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3114, "sent_len_1": 66.64, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.2612, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 85500 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.487, "doc_norm": 1.7556, "encoder_q-embeddings": 5034.2183, "encoder_q-layer.0": 3555.3479, "encoder_q-layer.1": 3640.7793, "encoder_q-layer.10": 6299.8096, "encoder_q-layer.11": 14674.9678, "encoder_q-layer.2": 4138.7949, "encoder_q-layer.3": 4326.5752, "encoder_q-layer.4": 4585.749, "encoder_q-layer.5": 4465.2051, "encoder_q-layer.6": 5087.7466, "encoder_q-layer.7": 5357.6025, "encoder_q-layer.8": 6409.8237, "encoder_q-layer.9": 6013.729, "epoch": 0.37, "inbatch_neg_score": 1.2796, "inbatch_pos_score": 1.8701, "learning_rate": 3.0105263157894737e-05, "loss": 3.487, "norm_diff": 0.0938, "norm_loss": 0.0, "num_token_doc": 66.7195, "num_token_overlap": 11.6546, "num_token_query": 31.311, "num_token_union": 65.0642, "num_word_context": 202.0995, "num_word_doc": 49.7445, "num_word_query": 23.2562, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9475.9776, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2793, "query_norm": 1.6619, "queue_k_norm": 1.7589, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.311, "sent_len_1": 66.7195, "sent_len_max_0": 127.99, "sent_len_max_1": 210.5588, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 85600 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4783, "doc_norm": 1.7617, "encoder_q-embeddings": 5846.3984, "encoder_q-layer.0": 4100.6919, "encoder_q-layer.1": 4315.623, "encoder_q-layer.10": 6689.29, "encoder_q-layer.11": 15973.0781, "encoder_q-layer.2": 4873.7744, "encoder_q-layer.3": 4983.1987, "encoder_q-layer.4": 5180.4937, "encoder_q-layer.5": 5285.8594, "encoder_q-layer.6": 5962.7925, "encoder_q-layer.7": 6863.3926, "encoder_q-layer.8": 7595.7354, "encoder_q-layer.9": 6492.6655, "epoch": 0.37, "inbatch_neg_score": 1.2828, "inbatch_pos_score": 1.8916, "learning_rate": 3.0078947368421056e-05, "loss": 3.4783, "norm_diff": 0.0946, "norm_loss": 0.0, "num_token_doc": 66.9785, "num_token_overlap": 11.6896, "num_token_query": 31.3987, "num_token_union": 65.2586, "num_word_context": 202.6211, "num_word_doc": 49.9825, "num_word_query": 23.3148, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10687.9573, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2812, "query_norm": 1.6671, "queue_k_norm": 1.7581, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3987, "sent_len_1": 66.9785, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.1513, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 85700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4859, "doc_norm": 1.7602, "encoder_q-embeddings": 5197.1509, "encoder_q-layer.0": 3568.0874, "encoder_q-layer.1": 3691.2556, "encoder_q-layer.10": 6699.4717, "encoder_q-layer.11": 15557.8398, "encoder_q-layer.2": 4135.1309, "encoder_q-layer.3": 4200.0913, "encoder_q-layer.4": 4403.8711, "encoder_q-layer.5": 4629.7505, "encoder_q-layer.6": 5125.0654, "encoder_q-layer.7": 5743.4492, "encoder_q-layer.8": 6875.4917, "encoder_q-layer.9": 6532.9126, "epoch": 0.37, "inbatch_neg_score": 1.2897, "inbatch_pos_score": 1.9004, "learning_rate": 3.005263157894737e-05, "loss": 3.4859, "norm_diff": 0.0865, "norm_loss": 0.0, "num_token_doc": 66.9625, "num_token_overlap": 11.6861, "num_token_query": 31.3795, "num_token_union": 65.2356, "num_word_context": 202.5394, "num_word_doc": 49.9721, "num_word_query": 23.3096, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9882.0606, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2891, "query_norm": 1.6737, "queue_k_norm": 1.7614, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3795, "sent_len_1": 66.9625, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.6587, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 85800 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.4876, "doc_norm": 1.7632, "encoder_q-embeddings": 5203.6714, "encoder_q-layer.0": 3563.7754, "encoder_q-layer.1": 3903.3257, "encoder_q-layer.10": 7482.0439, "encoder_q-layer.11": 15455.4199, "encoder_q-layer.2": 4477.9814, "encoder_q-layer.3": 4629.3135, "encoder_q-layer.4": 4961.2974, "encoder_q-layer.5": 5174.6865, "encoder_q-layer.6": 5914.5537, "encoder_q-layer.7": 6413.9609, "encoder_q-layer.8": 7400.2539, "encoder_q-layer.9": 6868.8457, "epoch": 0.37, "inbatch_neg_score": 1.2887, "inbatch_pos_score": 1.8984, "learning_rate": 3.0026315789473686e-05, "loss": 3.4876, "norm_diff": 0.0987, "norm_loss": 0.0, "num_token_doc": 66.7787, "num_token_overlap": 11.6909, "num_token_query": 31.3857, "num_token_union": 65.1286, "num_word_context": 202.4756, "num_word_doc": 49.8048, "num_word_query": 23.3076, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10068.1412, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.2891, "query_norm": 1.6645, "queue_k_norm": 1.7619, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3857, "sent_len_1": 66.7787, "sent_len_max_0": 127.9613, "sent_len_max_1": 210.3175, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 85900 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.4793, "doc_norm": 1.7639, "encoder_q-embeddings": 9719.0342, "encoder_q-layer.0": 6830.7969, "encoder_q-layer.1": 7424.8472, "encoder_q-layer.10": 15082.7227, "encoder_q-layer.11": 31408.873, "encoder_q-layer.2": 8542.3525, "encoder_q-layer.3": 8734.832, "encoder_q-layer.4": 9779.8008, "encoder_q-layer.5": 10042.2197, "encoder_q-layer.6": 10462.2031, "encoder_q-layer.7": 11819.4453, "encoder_q-layer.8": 14125.043, "encoder_q-layer.9": 13492.957, "epoch": 0.37, "inbatch_neg_score": 1.2885, "inbatch_pos_score": 1.8975, "learning_rate": 3e-05, "loss": 3.4793, "norm_diff": 0.1009, "norm_loss": 0.0, "num_token_doc": 66.7691, "num_token_overlap": 11.7034, "num_token_query": 31.4366, "num_token_union": 65.1296, "num_word_context": 202.3877, "num_word_doc": 49.8239, "num_word_query": 23.3398, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19772.9231, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.29, "query_norm": 1.663, "queue_k_norm": 1.7629, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4366, "sent_len_1": 66.7691, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.3713, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 86000 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.4986, "doc_norm": 1.7667, "encoder_q-embeddings": 10135.2109, "encoder_q-layer.0": 6700.3179, "encoder_q-layer.1": 7170.3027, "encoder_q-layer.10": 13689.543, "encoder_q-layer.11": 30546.6738, "encoder_q-layer.2": 8043.2573, "encoder_q-layer.3": 8491.3428, "encoder_q-layer.4": 8931.5156, "encoder_q-layer.5": 9209.0488, "encoder_q-layer.6": 10308.9785, "encoder_q-layer.7": 11910.6973, "encoder_q-layer.8": 14205.583, "encoder_q-layer.9": 12563.9824, "epoch": 0.37, "inbatch_neg_score": 1.294, "inbatch_pos_score": 1.8984, "learning_rate": 2.9973684210526316e-05, "loss": 3.4986, "norm_diff": 0.0929, "norm_loss": 0.0, "num_token_doc": 66.764, "num_token_overlap": 11.6158, "num_token_query": 31.2795, "num_token_union": 65.0623, "num_word_context": 202.3215, "num_word_doc": 49.7936, "num_word_query": 23.2148, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19350.9238, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.2939, "query_norm": 1.6738, "queue_k_norm": 1.7619, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2795, "sent_len_1": 66.764, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.245, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 86100 }, { "accuracy": 48.291, "active_queue_size": 16384.0, "cl_loss": 3.4907, "doc_norm": 1.763, "encoder_q-embeddings": 10900.4775, "encoder_q-layer.0": 7251.2183, "encoder_q-layer.1": 7754.9878, "encoder_q-layer.10": 14538.2852, "encoder_q-layer.11": 30382.6719, "encoder_q-layer.2": 8949.5713, "encoder_q-layer.3": 9480.0664, "encoder_q-layer.4": 9908.4082, "encoder_q-layer.5": 10609.3232, "encoder_q-layer.6": 11318.667, "encoder_q-layer.7": 12379.0498, "encoder_q-layer.8": 14762.1934, "encoder_q-layer.9": 13771.3682, "epoch": 0.37, "inbatch_neg_score": 1.2993, "inbatch_pos_score": 1.8906, "learning_rate": 2.9947368421052634e-05, "loss": 3.4907, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.8675, "num_token_overlap": 11.6852, "num_token_query": 31.417, "num_token_union": 65.1959, "num_word_context": 202.5682, "num_word_doc": 49.9026, "num_word_query": 23.3578, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20215.289, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.2988, "query_norm": 1.6793, "queue_k_norm": 1.7655, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.417, "sent_len_1": 66.8675, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.5863, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 86200 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.486, "doc_norm": 1.7678, "encoder_q-embeddings": 9818.7725, "encoder_q-layer.0": 6589.8242, "encoder_q-layer.1": 6914.2583, "encoder_q-layer.10": 12304.2725, "encoder_q-layer.11": 29164.4785, "encoder_q-layer.2": 7802.4897, "encoder_q-layer.3": 7994.2666, "encoder_q-layer.4": 8497.2529, "encoder_q-layer.5": 8719.082, "encoder_q-layer.6": 9404.9746, "encoder_q-layer.7": 10517.8623, "encoder_q-layer.8": 12762.9873, "encoder_q-layer.9": 11890.8135, "epoch": 0.37, "inbatch_neg_score": 1.3057, "inbatch_pos_score": 1.9062, "learning_rate": 2.992105263157895e-05, "loss": 3.486, "norm_diff": 0.0893, "norm_loss": 0.0, "num_token_doc": 66.8492, "num_token_overlap": 11.6498, "num_token_query": 31.2721, "num_token_union": 65.1029, "num_word_context": 202.116, "num_word_doc": 49.8897, "num_word_query": 23.2209, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18130.6851, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.3047, "query_norm": 1.6785, "queue_k_norm": 1.765, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2721, "sent_len_1": 66.8492, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.66, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 86300 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.4838, "doc_norm": 1.7677, "encoder_q-embeddings": 10070.2783, "encoder_q-layer.0": 6887.356, "encoder_q-layer.1": 7373.1294, "encoder_q-layer.10": 13002.3926, "encoder_q-layer.11": 29677.0781, "encoder_q-layer.2": 8256.5811, "encoder_q-layer.3": 8394.2334, "encoder_q-layer.4": 8983.0088, "encoder_q-layer.5": 8875.4268, "encoder_q-layer.6": 10082.7793, "encoder_q-layer.7": 11865.0674, "encoder_q-layer.8": 14570.5967, "encoder_q-layer.9": 12516.8887, "epoch": 0.37, "inbatch_neg_score": 1.3088, "inbatch_pos_score": 1.9033, "learning_rate": 2.9894736842105264e-05, "loss": 3.4838, "norm_diff": 0.0885, "norm_loss": 0.0, "num_token_doc": 66.953, "num_token_overlap": 11.6626, "num_token_query": 31.3168, "num_token_union": 65.1853, "num_word_context": 202.561, "num_word_doc": 49.9634, "num_word_query": 23.2718, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19320.4941, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.3086, "query_norm": 1.6791, "queue_k_norm": 1.7658, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3168, "sent_len_1": 66.953, "sent_len_max_0": 127.97, "sent_len_max_1": 209.6037, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 86400 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.4792, "doc_norm": 1.7713, "encoder_q-embeddings": 10577.5234, "encoder_q-layer.0": 7329.5952, "encoder_q-layer.1": 7990.7466, "encoder_q-layer.10": 12342.6787, "encoder_q-layer.11": 28783.3125, "encoder_q-layer.2": 8976.2314, "encoder_q-layer.3": 9334.0654, "encoder_q-layer.4": 10058.0498, "encoder_q-layer.5": 10847.3594, "encoder_q-layer.6": 11706.4893, "encoder_q-layer.7": 11757.1201, "encoder_q-layer.8": 13089.2266, "encoder_q-layer.9": 12339.1396, "epoch": 0.38, "inbatch_neg_score": 1.3088, "inbatch_pos_score": 1.9131, "learning_rate": 2.986842105263158e-05, "loss": 3.4792, "norm_diff": 0.0981, "norm_loss": 0.0, "num_token_doc": 66.9027, "num_token_overlap": 11.6883, "num_token_query": 31.3255, "num_token_union": 65.1819, "num_word_context": 202.3231, "num_word_doc": 49.9446, "num_word_query": 23.261, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19490.0976, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.3096, "query_norm": 1.6732, "queue_k_norm": 1.7676, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3255, "sent_len_1": 66.9027, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.1637, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 86500 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.481, "doc_norm": 1.7691, "encoder_q-embeddings": 5586.4683, "encoder_q-layer.0": 3827.2588, "encoder_q-layer.1": 4112.0972, "encoder_q-layer.10": 7223.3022, "encoder_q-layer.11": 14870.2822, "encoder_q-layer.2": 4737.3345, "encoder_q-layer.3": 5041.6411, "encoder_q-layer.4": 5227.54, "encoder_q-layer.5": 5293.1416, "encoder_q-layer.6": 5767.7036, "encoder_q-layer.7": 6505.6816, "encoder_q-layer.8": 7582.3442, "encoder_q-layer.9": 6573.0879, "epoch": 0.38, "inbatch_neg_score": 1.3126, "inbatch_pos_score": 1.9043, "learning_rate": 2.9842105263157894e-05, "loss": 3.481, "norm_diff": 0.0861, "norm_loss": 0.0, "num_token_doc": 66.6265, "num_token_overlap": 11.665, "num_token_query": 31.3301, "num_token_union": 65.0222, "num_word_context": 202.0548, "num_word_doc": 49.7232, "num_word_query": 23.261, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10209.1567, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3125, "query_norm": 1.683, "queue_k_norm": 1.7703, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3301, "sent_len_1": 66.6265, "sent_len_max_0": 127.965, "sent_len_max_1": 206.5687, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 86600 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.4844, "doc_norm": 1.7693, "encoder_q-embeddings": 6873.7021, "encoder_q-layer.0": 4636.9214, "encoder_q-layer.1": 5278.2407, "encoder_q-layer.10": 6807.6235, "encoder_q-layer.11": 15516.1025, "encoder_q-layer.2": 6186.5898, "encoder_q-layer.3": 6402.9351, "encoder_q-layer.4": 6839.3325, "encoder_q-layer.5": 6911.5166, "encoder_q-layer.6": 7307.3862, "encoder_q-layer.7": 7131.2437, "encoder_q-layer.8": 7389.9702, "encoder_q-layer.9": 6440.0869, "epoch": 0.38, "inbatch_neg_score": 1.3145, "inbatch_pos_score": 1.9121, "learning_rate": 2.9815789473684212e-05, "loss": 3.4844, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.9464, "num_token_overlap": 11.6902, "num_token_query": 31.3767, "num_token_union": 65.1827, "num_word_context": 202.5868, "num_word_doc": 49.9503, "num_word_query": 23.3081, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11439.4201, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3145, "query_norm": 1.6862, "queue_k_norm": 1.771, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3767, "sent_len_1": 66.9464, "sent_len_max_0": 127.975, "sent_len_max_1": 209.8975, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 86700 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.4816, "doc_norm": 1.7723, "encoder_q-embeddings": 4974.3809, "encoder_q-layer.0": 3344.4114, "encoder_q-layer.1": 3496.7639, "encoder_q-layer.10": 5671.7383, "encoder_q-layer.11": 13983.168, "encoder_q-layer.2": 3928.886, "encoder_q-layer.3": 3913.4348, "encoder_q-layer.4": 4166.9414, "encoder_q-layer.5": 4189.3628, "encoder_q-layer.6": 4560.3237, "encoder_q-layer.7": 5138.0391, "encoder_q-layer.8": 6282.0659, "encoder_q-layer.9": 5558.0894, "epoch": 0.38, "inbatch_neg_score": 1.3146, "inbatch_pos_score": 1.9189, "learning_rate": 2.9789473684210527e-05, "loss": 3.4816, "norm_diff": 0.0972, "norm_loss": 0.0, "num_token_doc": 66.8303, "num_token_overlap": 11.7081, "num_token_query": 31.4528, "num_token_union": 65.1859, "num_word_context": 202.3531, "num_word_doc": 49.8683, "num_word_query": 23.3703, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9075.511, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3154, "query_norm": 1.6751, "queue_k_norm": 1.7711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4528, "sent_len_1": 66.8303, "sent_len_max_0": 128.0, "sent_len_max_1": 207.1575, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 86800 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.4923, "doc_norm": 1.7685, "encoder_q-embeddings": 5269.9785, "encoder_q-layer.0": 3601.5925, "encoder_q-layer.1": 3845.1323, "encoder_q-layer.10": 6941.0381, "encoder_q-layer.11": 15132.7061, "encoder_q-layer.2": 4378.3882, "encoder_q-layer.3": 4361.6523, "encoder_q-layer.4": 4609.2007, "encoder_q-layer.5": 4691.6807, "encoder_q-layer.6": 5326.334, "encoder_q-layer.7": 5919.354, "encoder_q-layer.8": 7166.2832, "encoder_q-layer.9": 6553.2432, "epoch": 0.38, "inbatch_neg_score": 1.3244, "inbatch_pos_score": 1.9316, "learning_rate": 2.9763157894736842e-05, "loss": 3.4923, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.8425, "num_token_overlap": 11.6758, "num_token_query": 31.3183, "num_token_union": 65.134, "num_word_context": 202.2495, "num_word_doc": 49.8959, "num_word_query": 23.2531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9849.0563, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3242, "query_norm": 1.6919, "queue_k_norm": 1.7726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3183, "sent_len_1": 66.8425, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.3262, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 86900 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.4797, "doc_norm": 1.7697, "encoder_q-embeddings": 4934.167, "encoder_q-layer.0": 3422.011, "encoder_q-layer.1": 3659.9431, "encoder_q-layer.10": 5949.2686, "encoder_q-layer.11": 14598.2852, "encoder_q-layer.2": 4092.4395, "encoder_q-layer.3": 4081.0374, "encoder_q-layer.4": 4325.2041, "encoder_q-layer.5": 4530.0264, "encoder_q-layer.6": 5119.9531, "encoder_q-layer.7": 5807.8516, "encoder_q-layer.8": 6709.7427, "encoder_q-layer.9": 5904.9688, "epoch": 0.38, "inbatch_neg_score": 1.3215, "inbatch_pos_score": 1.9082, "learning_rate": 2.9736842105263157e-05, "loss": 3.4797, "norm_diff": 0.091, "norm_loss": 0.0, "num_token_doc": 66.8433, "num_token_overlap": 11.701, "num_token_query": 31.4462, "num_token_union": 65.1564, "num_word_context": 202.1273, "num_word_doc": 49.8675, "num_word_query": 23.3598, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9418.788, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3213, "query_norm": 1.6787, "queue_k_norm": 1.7711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4462, "sent_len_1": 66.8433, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.9938, "stdk": 0.0489, "stdq": 0.0426, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 87000 }, { "accuracy": 47.8516, "active_queue_size": 16384.0, "cl_loss": 3.4717, "doc_norm": 1.7759, "encoder_q-embeddings": 5351.3726, "encoder_q-layer.0": 3634.3008, "encoder_q-layer.1": 3819.0925, "encoder_q-layer.10": 7040.4282, "encoder_q-layer.11": 15140.0166, "encoder_q-layer.2": 4270.1938, "encoder_q-layer.3": 4507.2461, "encoder_q-layer.4": 4705.4165, "encoder_q-layer.5": 4840.5352, "encoder_q-layer.6": 5514.0366, "encoder_q-layer.7": 6186.9414, "encoder_q-layer.8": 7489.1089, "encoder_q-layer.9": 6584.2158, "epoch": 0.38, "inbatch_neg_score": 1.3251, "inbatch_pos_score": 1.915, "learning_rate": 2.971052631578948e-05, "loss": 3.4717, "norm_diff": 0.0811, "norm_loss": 0.0, "num_token_doc": 66.7912, "num_token_overlap": 11.6985, "num_token_query": 31.4282, "num_token_union": 65.1453, "num_word_context": 202.3195, "num_word_doc": 49.8313, "num_word_query": 23.3355, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10002.5405, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3242, "query_norm": 1.6948, "queue_k_norm": 1.7737, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4282, "sent_len_1": 66.7912, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.7388, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87100 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.5014, "doc_norm": 1.7759, "encoder_q-embeddings": 4957.8325, "encoder_q-layer.0": 3437.3066, "encoder_q-layer.1": 3634.6836, "encoder_q-layer.10": 6467.7183, "encoder_q-layer.11": 14989.4385, "encoder_q-layer.2": 4130.7427, "encoder_q-layer.3": 4203.5991, "encoder_q-layer.4": 4396.5786, "encoder_q-layer.5": 4557.0723, "encoder_q-layer.6": 5086.1011, "encoder_q-layer.7": 5588.8809, "encoder_q-layer.8": 6599.5986, "encoder_q-layer.9": 6262.9082, "epoch": 0.38, "inbatch_neg_score": 1.3322, "inbatch_pos_score": 1.9463, "learning_rate": 2.968421052631579e-05, "loss": 3.5014, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.843, "num_token_overlap": 11.6421, "num_token_query": 31.316, "num_token_union": 65.1345, "num_word_context": 202.2344, "num_word_doc": 49.8797, "num_word_query": 23.2614, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9568.5138, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.332, "query_norm": 1.706, "queue_k_norm": 1.7742, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.316, "sent_len_1": 66.843, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.2337, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87200 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.4905, "doc_norm": 1.773, "encoder_q-embeddings": 4919.4756, "encoder_q-layer.0": 3401.675, "encoder_q-layer.1": 3635.9348, "encoder_q-layer.10": 7022.999, "encoder_q-layer.11": 15503.209, "encoder_q-layer.2": 3947.7432, "encoder_q-layer.3": 4093.2168, "encoder_q-layer.4": 4304.9473, "encoder_q-layer.5": 4377.415, "encoder_q-layer.6": 5104.0156, "encoder_q-layer.7": 5619.855, "encoder_q-layer.8": 6965.4893, "encoder_q-layer.9": 6620.9224, "epoch": 0.38, "inbatch_neg_score": 1.3407, "inbatch_pos_score": 1.9395, "learning_rate": 2.9657894736842106e-05, "loss": 3.4905, "norm_diff": 0.0587, "norm_loss": 0.0, "num_token_doc": 66.7496, "num_token_overlap": 11.6794, "num_token_query": 31.3881, "num_token_union": 65.0942, "num_word_context": 202.2342, "num_word_doc": 49.7911, "num_word_query": 23.301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9708.9595, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3408, "query_norm": 1.7143, "queue_k_norm": 1.7764, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3881, "sent_len_1": 66.7496, "sent_len_max_0": 127.9875, "sent_len_max_1": 211.08, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87300 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4861, "doc_norm": 1.7774, "encoder_q-embeddings": 5711.4224, "encoder_q-layer.0": 4021.9683, "encoder_q-layer.1": 4406.8242, "encoder_q-layer.10": 6315.2349, "encoder_q-layer.11": 14220.1113, "encoder_q-layer.2": 5042.0679, "encoder_q-layer.3": 4896.0547, "encoder_q-layer.4": 5307.5347, "encoder_q-layer.5": 5490.3252, "encoder_q-layer.6": 5779.2666, "encoder_q-layer.7": 6473.7114, "encoder_q-layer.8": 7347.5869, "encoder_q-layer.9": 5829.3315, "epoch": 0.38, "inbatch_neg_score": 1.3483, "inbatch_pos_score": 1.9453, "learning_rate": 2.963157894736842e-05, "loss": 3.4861, "norm_diff": 0.0572, "norm_loss": 0.0, "num_token_doc": 66.6616, "num_token_overlap": 11.6607, "num_token_query": 31.3622, "num_token_union": 65.0784, "num_word_context": 202.3335, "num_word_doc": 49.7734, "num_word_query": 23.3042, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9955.8338, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3477, "query_norm": 1.7202, "queue_k_norm": 1.777, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3622, "sent_len_1": 66.6616, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.1425, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87400 }, { "accuracy": 48.7305, "active_queue_size": 16384.0, "cl_loss": 3.485, "doc_norm": 1.7758, "encoder_q-embeddings": 5099.3394, "encoder_q-layer.0": 3422.6111, "encoder_q-layer.1": 3680.4399, "encoder_q-layer.10": 6264.4634, "encoder_q-layer.11": 14932.5312, "encoder_q-layer.2": 4057.4202, "encoder_q-layer.3": 4205.4194, "encoder_q-layer.4": 4511.4458, "encoder_q-layer.5": 4605.252, "encoder_q-layer.6": 5049.3154, "encoder_q-layer.7": 5713.7603, "encoder_q-layer.8": 6889.5103, "encoder_q-layer.9": 6174.3813, "epoch": 0.38, "inbatch_neg_score": 1.3629, "inbatch_pos_score": 1.9502, "learning_rate": 2.9605263157894735e-05, "loss": 3.485, "norm_diff": 0.0496, "norm_loss": 0.0, "num_token_doc": 66.6512, "num_token_overlap": 11.6782, "num_token_query": 31.3622, "num_token_union": 65.0387, "num_word_context": 202.2957, "num_word_doc": 49.7564, "num_word_query": 23.2863, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9570.6412, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3623, "query_norm": 1.7261, "queue_k_norm": 1.7792, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3622, "sent_len_1": 66.6512, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.3963, "stdk": 0.0489, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 87500 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4817, "doc_norm": 1.78, "encoder_q-embeddings": 5051.2969, "encoder_q-layer.0": 3399.853, "encoder_q-layer.1": 3582.6082, "encoder_q-layer.10": 6479.2959, "encoder_q-layer.11": 14846.04, "encoder_q-layer.2": 4040.3713, "encoder_q-layer.3": 4188.7305, "encoder_q-layer.4": 4448.5483, "encoder_q-layer.5": 4523.6582, "encoder_q-layer.6": 5305.3105, "encoder_q-layer.7": 5975.0205, "encoder_q-layer.8": 6732.8828, "encoder_q-layer.9": 6095.3721, "epoch": 0.38, "inbatch_neg_score": 1.3759, "inbatch_pos_score": 1.9648, "learning_rate": 2.9578947368421057e-05, "loss": 3.4817, "norm_diff": 0.0504, "norm_loss": 0.0, "num_token_doc": 66.6397, "num_token_overlap": 11.6718, "num_token_query": 31.3701, "num_token_union": 65.0234, "num_word_context": 202.1927, "num_word_doc": 49.7348, "num_word_query": 23.2876, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9555.0644, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.375, "query_norm": 1.7296, "queue_k_norm": 1.7783, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3701, "sent_len_1": 66.6397, "sent_len_max_0": 127.99, "sent_len_max_1": 206.6675, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.049, "stdqueue_q": 0.0, "step": 87600 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.466, "doc_norm": 1.7818, "encoder_q-embeddings": 5870.5415, "encoder_q-layer.0": 4100.2402, "encoder_q-layer.1": 4632.5513, "encoder_q-layer.10": 6567.8701, "encoder_q-layer.11": 15106.8984, "encoder_q-layer.2": 5293.5356, "encoder_q-layer.3": 5552.9346, "encoder_q-layer.4": 6045.8916, "encoder_q-layer.5": 6121.2808, "encoder_q-layer.6": 6855.8657, "encoder_q-layer.7": 7822.8447, "encoder_q-layer.8": 8049.3037, "encoder_q-layer.9": 6686.0684, "epoch": 0.38, "inbatch_neg_score": 1.3859, "inbatch_pos_score": 1.9883, "learning_rate": 2.9552631578947372e-05, "loss": 3.466, "norm_diff": 0.0363, "norm_loss": 0.0, "num_token_doc": 66.7374, "num_token_overlap": 11.7235, "num_token_query": 31.4737, "num_token_union": 65.121, "num_word_context": 202.2866, "num_word_doc": 49.7919, "num_word_query": 23.38, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10877.8603, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3857, "query_norm": 1.7455, "queue_k_norm": 1.7833, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4737, "sent_len_1": 66.7374, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.92, "stdk": 0.0491, "stdq": 0.044, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 87700 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4762, "doc_norm": 1.7801, "encoder_q-embeddings": 4821.4087, "encoder_q-layer.0": 3323.2153, "encoder_q-layer.1": 3597.9788, "encoder_q-layer.10": 6456.9453, "encoder_q-layer.11": 14808.334, "encoder_q-layer.2": 3913.6047, "encoder_q-layer.3": 3961.0676, "encoder_q-layer.4": 4200.8315, "encoder_q-layer.5": 4266.5708, "encoder_q-layer.6": 4725.5244, "encoder_q-layer.7": 5196.8818, "encoder_q-layer.8": 6659.5981, "encoder_q-layer.9": 5986.6929, "epoch": 0.38, "inbatch_neg_score": 1.3941, "inbatch_pos_score": 2.002, "learning_rate": 2.9526315789473684e-05, "loss": 3.4762, "norm_diff": 0.0533, "norm_loss": 0.0, "num_token_doc": 66.7698, "num_token_overlap": 11.6778, "num_token_query": 31.4024, "num_token_union": 65.1541, "num_word_context": 202.1183, "num_word_doc": 49.8147, "num_word_query": 23.312, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9305.8755, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.3945, "query_norm": 1.7268, "queue_k_norm": 1.7831, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4024, "sent_len_1": 66.7698, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.1825, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87800 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4742, "doc_norm": 1.7861, "encoder_q-embeddings": 4797.459, "encoder_q-layer.0": 3346.6736, "encoder_q-layer.1": 3613.1292, "encoder_q-layer.10": 6838.0571, "encoder_q-layer.11": 15659.1504, "encoder_q-layer.2": 3808.0818, "encoder_q-layer.3": 3966.7976, "encoder_q-layer.4": 4504.0972, "encoder_q-layer.5": 4544.418, "encoder_q-layer.6": 5128.5342, "encoder_q-layer.7": 6076.1582, "encoder_q-layer.8": 7025.543, "encoder_q-layer.9": 6407.0693, "epoch": 0.38, "inbatch_neg_score": 1.4046, "inbatch_pos_score": 2.0098, "learning_rate": 2.95e-05, "loss": 3.4742, "norm_diff": 0.0653, "norm_loss": 0.0, "num_token_doc": 66.8773, "num_token_overlap": 11.6879, "num_token_query": 31.3655, "num_token_union": 65.194, "num_word_context": 202.5532, "num_word_doc": 49.9047, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9765.5855, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4043, "query_norm": 1.7207, "queue_k_norm": 1.785, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3655, "sent_len_1": 66.8773, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.7587, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 87900 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4751, "doc_norm": 1.7872, "encoder_q-embeddings": 6946.9224, "encoder_q-layer.0": 5129.125, "encoder_q-layer.1": 5451.2075, "encoder_q-layer.10": 6541.3027, "encoder_q-layer.11": 14971.877, "encoder_q-layer.2": 6448.9561, "encoder_q-layer.3": 6656.7153, "encoder_q-layer.4": 7058.3154, "encoder_q-layer.5": 7015.4414, "encoder_q-layer.6": 7960.7075, "encoder_q-layer.7": 7897.21, "encoder_q-layer.8": 7493.5757, "encoder_q-layer.9": 6144.3042, "epoch": 0.38, "inbatch_neg_score": 1.4062, "inbatch_pos_score": 2.0098, "learning_rate": 2.9473684210526314e-05, "loss": 3.4751, "norm_diff": 0.0702, "norm_loss": 0.0, "num_token_doc": 66.994, "num_token_overlap": 11.6989, "num_token_query": 31.36, "num_token_union": 65.2466, "num_word_context": 202.5104, "num_word_doc": 49.9875, "num_word_query": 23.2773, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11470.0075, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4053, "query_norm": 1.717, "queue_k_norm": 1.7881, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.36, "sent_len_1": 66.994, "sent_len_max_0": 127.9363, "sent_len_max_1": 209.6538, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 88000 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4743, "doc_norm": 1.7892, "encoder_q-embeddings": 5488.1226, "encoder_q-layer.0": 3583.6377, "encoder_q-layer.1": 3800.9766, "encoder_q-layer.10": 7073.3496, "encoder_q-layer.11": 15352.4189, "encoder_q-layer.2": 4282.4478, "encoder_q-layer.3": 4395.0664, "encoder_q-layer.4": 4664.1958, "encoder_q-layer.5": 4906.792, "encoder_q-layer.6": 5746.4912, "encoder_q-layer.7": 6125.4893, "encoder_q-layer.8": 6804.2568, "encoder_q-layer.9": 6528.9985, "epoch": 0.38, "inbatch_neg_score": 1.4063, "inbatch_pos_score": 1.9922, "learning_rate": 2.9447368421052635e-05, "loss": 3.4743, "norm_diff": 0.0876, "norm_loss": 0.0, "num_token_doc": 66.7723, "num_token_overlap": 11.729, "num_token_query": 31.5316, "num_token_union": 65.1807, "num_word_context": 202.3509, "num_word_doc": 49.8307, "num_word_query": 23.4445, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9963.8603, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4072, "query_norm": 1.7016, "queue_k_norm": 1.7892, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5316, "sent_len_1": 66.7723, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.27, "stdk": 0.0491, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88100 }, { "accuracy": 47.9004, "active_queue_size": 16384.0, "cl_loss": 3.4792, "doc_norm": 1.7932, "encoder_q-embeddings": 4872.3828, "encoder_q-layer.0": 3532.2522, "encoder_q-layer.1": 3669.8359, "encoder_q-layer.10": 6581.8774, "encoder_q-layer.11": 15996.748, "encoder_q-layer.2": 4057.3369, "encoder_q-layer.3": 4063.0461, "encoder_q-layer.4": 4211.3091, "encoder_q-layer.5": 4268.3818, "encoder_q-layer.6": 4954.5474, "encoder_q-layer.7": 5688.4043, "encoder_q-layer.8": 6915.9141, "encoder_q-layer.9": 6323.0308, "epoch": 0.38, "inbatch_neg_score": 1.4083, "inbatch_pos_score": 1.9883, "learning_rate": 2.942105263157895e-05, "loss": 3.4792, "norm_diff": 0.0851, "norm_loss": 0.0, "num_token_doc": 66.656, "num_token_overlap": 11.6561, "num_token_query": 31.3304, "num_token_union": 65.0365, "num_word_context": 202.1442, "num_word_doc": 49.7216, "num_word_query": 23.2721, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9857.4175, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4102, "query_norm": 1.7081, "queue_k_norm": 1.7906, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3304, "sent_len_1": 66.656, "sent_len_max_0": 127.98, "sent_len_max_1": 209.825, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88200 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4826, "doc_norm": 1.7909, "encoder_q-embeddings": 4769.4497, "encoder_q-layer.0": 3228.4697, "encoder_q-layer.1": 3465.9822, "encoder_q-layer.10": 6971.1831, "encoder_q-layer.11": 15548.9092, "encoder_q-layer.2": 3838.9119, "encoder_q-layer.3": 4018.8306, "encoder_q-layer.4": 4267.6777, "encoder_q-layer.5": 4425.7334, "encoder_q-layer.6": 5122.7622, "encoder_q-layer.7": 5735.8311, "encoder_q-layer.8": 7345.7271, "encoder_q-layer.9": 6853.0977, "epoch": 0.38, "inbatch_neg_score": 1.4086, "inbatch_pos_score": 2.0, "learning_rate": 2.9394736842105265e-05, "loss": 3.4826, "norm_diff": 0.0979, "norm_loss": 0.0, "num_token_doc": 66.8351, "num_token_overlap": 11.6829, "num_token_query": 31.3061, "num_token_union": 65.121, "num_word_context": 202.3879, "num_word_doc": 49.8771, "num_word_query": 23.2455, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9793.1811, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4082, "query_norm": 1.693, "queue_k_norm": 1.7922, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3061, "sent_len_1": 66.8351, "sent_len_max_0": 127.99, "sent_len_max_1": 207.8113, "stdk": 0.049, "stdq": 0.0424, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88300 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.5014, "doc_norm": 1.7914, "encoder_q-embeddings": 2567.4346, "encoder_q-layer.0": 1739.3821, "encoder_q-layer.1": 1873.9484, "encoder_q-layer.10": 3041.2192, "encoder_q-layer.11": 7259.1836, "encoder_q-layer.2": 2079.4265, "encoder_q-layer.3": 2106.0676, "encoder_q-layer.4": 2217.3164, "encoder_q-layer.5": 2271.8867, "encoder_q-layer.6": 2563.0974, "encoder_q-layer.7": 2840.7959, "encoder_q-layer.8": 3241.8701, "encoder_q-layer.9": 2941.0527, "epoch": 0.38, "inbatch_neg_score": 1.409, "inbatch_pos_score": 1.998, "learning_rate": 2.9368421052631577e-05, "loss": 3.5014, "norm_diff": 0.092, "norm_loss": 0.0, "num_token_doc": 66.6404, "num_token_overlap": 11.6715, "num_token_query": 31.4758, "num_token_union": 65.1173, "num_word_context": 202.3276, "num_word_doc": 49.731, "num_word_query": 23.3801, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4758.0508, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4092, "query_norm": 1.6994, "queue_k_norm": 1.7931, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4758, "sent_len_1": 66.6404, "sent_len_max_0": 127.985, "sent_len_max_1": 209.2562, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88400 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.4563, "doc_norm": 1.7958, "encoder_q-embeddings": 2571.6011, "encoder_q-layer.0": 1731.1068, "encoder_q-layer.1": 1850.9205, "encoder_q-layer.10": 4095.1707, "encoder_q-layer.11": 8371.2656, "encoder_q-layer.2": 2038.7457, "encoder_q-layer.3": 2175.8557, "encoder_q-layer.4": 2381.5442, "encoder_q-layer.5": 2379.8408, "encoder_q-layer.6": 2718.8638, "encoder_q-layer.7": 3132.2837, "encoder_q-layer.8": 4019.2588, "encoder_q-layer.9": 3574.4407, "epoch": 0.38, "inbatch_neg_score": 1.4127, "inbatch_pos_score": 2.0156, "learning_rate": 2.93421052631579e-05, "loss": 3.4563, "norm_diff": 0.0746, "norm_loss": 0.0, "num_token_doc": 66.8128, "num_token_overlap": 11.7397, "num_token_query": 31.5407, "num_token_union": 65.1833, "num_word_context": 202.4958, "num_word_doc": 49.8646, "num_word_query": 23.4329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5259.7226, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4121, "query_norm": 1.7212, "queue_k_norm": 1.7956, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5407, "sent_len_1": 66.8128, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.1962, "stdk": 0.0491, "stdq": 0.0438, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 88500 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4705, "doc_norm": 1.7935, "encoder_q-embeddings": 2666.8025, "encoder_q-layer.0": 1937.1105, "encoder_q-layer.1": 1976.6722, "encoder_q-layer.10": 3282.6436, "encoder_q-layer.11": 7454.4287, "encoder_q-layer.2": 2073.3792, "encoder_q-layer.3": 2104.863, "encoder_q-layer.4": 2303.4141, "encoder_q-layer.5": 2345.0806, "encoder_q-layer.6": 2574.8599, "encoder_q-layer.7": 2790.0737, "encoder_q-layer.8": 3386.8384, "encoder_q-layer.9": 3050.8296, "epoch": 0.38, "inbatch_neg_score": 1.4158, "inbatch_pos_score": 1.9961, "learning_rate": 2.9315789473684214e-05, "loss": 3.4705, "norm_diff": 0.0944, "norm_loss": 0.0, "num_token_doc": 66.7494, "num_token_overlap": 11.6875, "num_token_query": 31.3994, "num_token_union": 65.1354, "num_word_context": 202.5026, "num_word_doc": 49.8438, "num_word_query": 23.3282, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4845.3482, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.416, "query_norm": 1.6992, "queue_k_norm": 1.7965, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3994, "sent_len_1": 66.7494, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.7463, "stdk": 0.049, "stdq": 0.0426, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88600 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.486, "doc_norm": 1.8025, "encoder_q-embeddings": 2491.9309, "encoder_q-layer.0": 1690.3734, "encoder_q-layer.1": 1820.5417, "encoder_q-layer.10": 3353.832, "encoder_q-layer.11": 7753.8467, "encoder_q-layer.2": 2054.7632, "encoder_q-layer.3": 2063.6956, "encoder_q-layer.4": 2277.3816, "encoder_q-layer.5": 2367.5181, "encoder_q-layer.6": 2534.178, "encoder_q-layer.7": 2824.2556, "encoder_q-layer.8": 3493.5076, "encoder_q-layer.9": 3156.1218, "epoch": 0.38, "inbatch_neg_score": 1.416, "inbatch_pos_score": 2.0117, "learning_rate": 2.928947368421053e-05, "loss": 3.486, "norm_diff": 0.1003, "norm_loss": 0.0, "num_token_doc": 66.6401, "num_token_overlap": 11.634, "num_token_query": 31.325, "num_token_union": 65.0453, "num_word_context": 202.3455, "num_word_doc": 49.7393, "num_word_query": 23.2665, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4911.9986, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.416, "query_norm": 1.7022, "queue_k_norm": 1.7987, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.325, "sent_len_1": 66.6401, "sent_len_max_0": 127.9788, "sent_len_max_1": 207.8688, "stdk": 0.0493, "stdq": 0.0428, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 88700 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4862, "doc_norm": 1.7991, "encoder_q-embeddings": 2751.6167, "encoder_q-layer.0": 1816.1799, "encoder_q-layer.1": 2040.8484, "encoder_q-layer.10": 3238.4512, "encoder_q-layer.11": 7611.1865, "encoder_q-layer.2": 2209.6545, "encoder_q-layer.3": 2317.0925, "encoder_q-layer.4": 2540.2368, "encoder_q-layer.5": 2720.0979, "encoder_q-layer.6": 2885.165, "encoder_q-layer.7": 3091.7151, "encoder_q-layer.8": 3599.8098, "encoder_q-layer.9": 3068.3687, "epoch": 0.39, "inbatch_neg_score": 1.4249, "inbatch_pos_score": 2.0176, "learning_rate": 2.9263157894736844e-05, "loss": 3.4862, "norm_diff": 0.088, "norm_loss": 0.0, "num_token_doc": 66.8215, "num_token_overlap": 11.636, "num_token_query": 31.3042, "num_token_union": 65.145, "num_word_context": 202.4014, "num_word_doc": 49.8608, "num_word_query": 23.2543, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5033.0837, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4238, "query_norm": 1.7111, "queue_k_norm": 1.7983, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3042, "sent_len_1": 66.8215, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.3487, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 88800 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4864, "doc_norm": 1.7991, "encoder_q-embeddings": 2505.4214, "encoder_q-layer.0": 1707.6963, "encoder_q-layer.1": 1785.1896, "encoder_q-layer.10": 3361.4746, "encoder_q-layer.11": 7516.1582, "encoder_q-layer.2": 1964.0583, "encoder_q-layer.3": 2014.5386, "encoder_q-layer.4": 2220.9448, "encoder_q-layer.5": 2222.1802, "encoder_q-layer.6": 2525.1631, "encoder_q-layer.7": 2840.8347, "encoder_q-layer.8": 3642.7781, "encoder_q-layer.9": 3177.2263, "epoch": 0.39, "inbatch_neg_score": 1.421, "inbatch_pos_score": 2.0273, "learning_rate": 2.9236842105263155e-05, "loss": 3.4864, "norm_diff": 0.0912, "norm_loss": 0.0, "num_token_doc": 66.9754, "num_token_overlap": 11.6479, "num_token_query": 31.2897, "num_token_union": 65.2238, "num_word_context": 202.3651, "num_word_doc": 49.965, "num_word_query": 23.2237, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4859.2682, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4219, "query_norm": 1.7079, "queue_k_norm": 1.8004, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2897, "sent_len_1": 66.9754, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.3812, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 88900 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.4788, "doc_norm": 1.8033, "encoder_q-embeddings": 2487.5337, "encoder_q-layer.0": 1666.3862, "encoder_q-layer.1": 1749.1814, "encoder_q-layer.10": 3844.3943, "encoder_q-layer.11": 8174.1689, "encoder_q-layer.2": 1964.5255, "encoder_q-layer.3": 1976.0615, "encoder_q-layer.4": 2056.3459, "encoder_q-layer.5": 2207.3357, "encoder_q-layer.6": 2512.7302, "encoder_q-layer.7": 2797.0068, "encoder_q-layer.8": 3729.217, "encoder_q-layer.9": 3541.5618, "epoch": 0.39, "inbatch_neg_score": 1.4273, "inbatch_pos_score": 2.0293, "learning_rate": 2.9210526315789477e-05, "loss": 3.4788, "norm_diff": 0.0963, "norm_loss": 0.0, "num_token_doc": 66.7617, "num_token_overlap": 11.7007, "num_token_query": 31.4964, "num_token_union": 65.1669, "num_word_context": 202.3071, "num_word_doc": 49.8195, "num_word_query": 23.401, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4949.4027, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4277, "query_norm": 1.707, "queue_k_norm": 1.8009, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4964, "sent_len_1": 66.7617, "sent_len_max_0": 128.0, "sent_len_max_1": 209.9988, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 89000 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4657, "doc_norm": 1.8049, "encoder_q-embeddings": 2747.0852, "encoder_q-layer.0": 1906.0916, "encoder_q-layer.1": 2064.2712, "encoder_q-layer.10": 3344.6331, "encoder_q-layer.11": 7546.1426, "encoder_q-layer.2": 2390.6997, "encoder_q-layer.3": 2515.7368, "encoder_q-layer.4": 2612.968, "encoder_q-layer.5": 2628.7261, "encoder_q-layer.6": 2911.0132, "encoder_q-layer.7": 3199.3433, "encoder_q-layer.8": 3687.2861, "encoder_q-layer.9": 3210.2197, "epoch": 0.39, "inbatch_neg_score": 1.4275, "inbatch_pos_score": 2.0293, "learning_rate": 2.9184210526315792e-05, "loss": 3.4657, "norm_diff": 0.1004, "norm_loss": 0.0, "num_token_doc": 66.913, "num_token_overlap": 11.7441, "num_token_query": 31.5394, "num_token_union": 65.2731, "num_word_context": 202.4196, "num_word_doc": 49.9399, "num_word_query": 23.4375, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5088.3421, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4287, "query_norm": 1.7044, "queue_k_norm": 1.8023, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5394, "sent_len_1": 66.913, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.1962, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 89100 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.4756, "doc_norm": 1.8052, "encoder_q-embeddings": 2702.282, "encoder_q-layer.0": 1790.5339, "encoder_q-layer.1": 1920.1677, "encoder_q-layer.10": 3175.4709, "encoder_q-layer.11": 7559.3467, "encoder_q-layer.2": 2108.0337, "encoder_q-layer.3": 2148.5388, "encoder_q-layer.4": 2265.8157, "encoder_q-layer.5": 2321.6912, "encoder_q-layer.6": 2606.4282, "encoder_q-layer.7": 2761.5034, "encoder_q-layer.8": 3339.4202, "encoder_q-layer.9": 3170.1448, "epoch": 0.39, "inbatch_neg_score": 1.4292, "inbatch_pos_score": 2.0195, "learning_rate": 2.9157894736842107e-05, "loss": 3.4756, "norm_diff": 0.0909, "norm_loss": 0.0, "num_token_doc": 66.8914, "num_token_overlap": 11.6519, "num_token_query": 31.2764, "num_token_union": 65.1451, "num_word_context": 202.4102, "num_word_doc": 49.9161, "num_word_query": 23.2218, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4902.1211, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4287, "query_norm": 1.7143, "queue_k_norm": 1.8031, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2764, "sent_len_1": 66.8914, "sent_len_max_0": 127.99, "sent_len_max_1": 209.1275, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 89200 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.4816, "doc_norm": 1.8048, "encoder_q-embeddings": 2753.428, "encoder_q-layer.0": 1850.8779, "encoder_q-layer.1": 1981.051, "encoder_q-layer.10": 3368.24, "encoder_q-layer.11": 7457.6494, "encoder_q-layer.2": 2220.7974, "encoder_q-layer.3": 2292.4673, "encoder_q-layer.4": 2437.5457, "encoder_q-layer.5": 2613.1042, "encoder_q-layer.6": 2844.1233, "encoder_q-layer.7": 3162.7676, "encoder_q-layer.8": 3540.2305, "encoder_q-layer.9": 3118.7634, "epoch": 0.39, "inbatch_neg_score": 1.4333, "inbatch_pos_score": 2.0391, "learning_rate": 2.9131578947368422e-05, "loss": 3.4816, "norm_diff": 0.0843, "norm_loss": 0.0, "num_token_doc": 66.7714, "num_token_overlap": 11.6654, "num_token_query": 31.3933, "num_token_union": 65.1406, "num_word_context": 202.6472, "num_word_doc": 49.827, "num_word_query": 23.3307, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4945.9668, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4326, "query_norm": 1.7205, "queue_k_norm": 1.8046, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3933, "sent_len_1": 66.7714, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.11, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 89300 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4828, "doc_norm": 1.8045, "encoder_q-embeddings": 2514.3484, "encoder_q-layer.0": 1702.6344, "encoder_q-layer.1": 1753.1611, "encoder_q-layer.10": 3219.5417, "encoder_q-layer.11": 7343.9233, "encoder_q-layer.2": 1987.3051, "encoder_q-layer.3": 2078.4297, "encoder_q-layer.4": 2231.8887, "encoder_q-layer.5": 2228.5276, "encoder_q-layer.6": 2436.4592, "encoder_q-layer.7": 2786.0603, "encoder_q-layer.8": 3275.7969, "encoder_q-layer.9": 3043.9712, "epoch": 0.39, "inbatch_neg_score": 1.4365, "inbatch_pos_score": 2.0312, "learning_rate": 2.910526315789474e-05, "loss": 3.4828, "norm_diff": 0.094, "norm_loss": 0.0, "num_token_doc": 66.743, "num_token_overlap": 11.6605, "num_token_query": 31.3616, "num_token_union": 65.1123, "num_word_context": 202.2967, "num_word_doc": 49.8208, "num_word_query": 23.2923, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4733.2894, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4355, "query_norm": 1.7105, "queue_k_norm": 1.8042, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3616, "sent_len_1": 66.743, "sent_len_max_0": 127.94, "sent_len_max_1": 207.9275, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 89400 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4772, "doc_norm": 1.8032, "encoder_q-embeddings": 2501.3616, "encoder_q-layer.0": 1714.926, "encoder_q-layer.1": 1883.0825, "encoder_q-layer.10": 3057.5933, "encoder_q-layer.11": 7186.8252, "encoder_q-layer.2": 2113.1455, "encoder_q-layer.3": 2179.355, "encoder_q-layer.4": 2263.7292, "encoder_q-layer.5": 2273.502, "encoder_q-layer.6": 2479.1379, "encoder_q-layer.7": 2735.519, "encoder_q-layer.8": 3223.9941, "encoder_q-layer.9": 2939.5332, "epoch": 0.39, "inbatch_neg_score": 1.442, "inbatch_pos_score": 2.0371, "learning_rate": 2.9078947368421055e-05, "loss": 3.4772, "norm_diff": 0.0944, "norm_loss": 0.0, "num_token_doc": 66.7759, "num_token_overlap": 11.7059, "num_token_query": 31.4023, "num_token_union": 65.0869, "num_word_context": 202.0892, "num_word_doc": 49.797, "num_word_query": 23.3135, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4714.9702, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4414, "query_norm": 1.7089, "queue_k_norm": 1.8049, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4023, "sent_len_1": 66.7759, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.455, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 89500 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4819, "doc_norm": 1.8059, "encoder_q-embeddings": 2489.0608, "encoder_q-layer.0": 1772.5, "encoder_q-layer.1": 1882.976, "encoder_q-layer.10": 3068.9285, "encoder_q-layer.11": 7456.874, "encoder_q-layer.2": 2123.3796, "encoder_q-layer.3": 2160.9492, "encoder_q-layer.4": 2260.1057, "encoder_q-layer.5": 2320.5144, "encoder_q-layer.6": 2581.4741, "encoder_q-layer.7": 3012.0352, "encoder_q-layer.8": 3556.9163, "encoder_q-layer.9": 3046.1152, "epoch": 0.39, "inbatch_neg_score": 1.4452, "inbatch_pos_score": 2.041, "learning_rate": 2.905263157894737e-05, "loss": 3.4819, "norm_diff": 0.0938, "norm_loss": 0.0, "num_token_doc": 66.6603, "num_token_overlap": 11.6863, "num_token_query": 31.3902, "num_token_union": 65.0174, "num_word_context": 202.2605, "num_word_doc": 49.7371, "num_word_query": 23.3112, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4892.6533, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4453, "query_norm": 1.7121, "queue_k_norm": 1.8067, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3902, "sent_len_1": 66.6603, "sent_len_max_0": 127.98, "sent_len_max_1": 209.95, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 89600 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.4862, "doc_norm": 1.8048, "encoder_q-embeddings": 2648.145, "encoder_q-layer.0": 1781.8135, "encoder_q-layer.1": 1869.5403, "encoder_q-layer.10": 3271.8865, "encoder_q-layer.11": 7564.1377, "encoder_q-layer.2": 2028.6206, "encoder_q-layer.3": 2094.658, "encoder_q-layer.4": 2188.0723, "encoder_q-layer.5": 2170.1421, "encoder_q-layer.6": 2415.0073, "encoder_q-layer.7": 2793.3679, "encoder_q-layer.8": 3385.4822, "encoder_q-layer.9": 3183.2581, "epoch": 0.39, "inbatch_neg_score": 1.4496, "inbatch_pos_score": 2.0527, "learning_rate": 2.9026315789473685e-05, "loss": 3.4862, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.8233, "num_token_overlap": 11.6255, "num_token_query": 31.2821, "num_token_union": 65.1346, "num_word_context": 202.4255, "num_word_doc": 49.8322, "num_word_query": 23.2336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4842.8066, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4492, "query_norm": 1.7313, "queue_k_norm": 1.808, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2821, "sent_len_1": 66.8233, "sent_len_max_0": 127.965, "sent_len_max_1": 210.3038, "stdk": 0.049, "stdq": 0.0437, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 89700 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4794, "doc_norm": 1.8127, "encoder_q-embeddings": 4552.1514, "encoder_q-layer.0": 3268.3774, "encoder_q-layer.1": 3900.4065, "encoder_q-layer.10": 3327.0957, "encoder_q-layer.11": 7635.1758, "encoder_q-layer.2": 4714.979, "encoder_q-layer.3": 4113.7422, "encoder_q-layer.4": 4286.1284, "encoder_q-layer.5": 4499.5269, "encoder_q-layer.6": 4881.644, "encoder_q-layer.7": 5093.1963, "encoder_q-layer.8": 4159.6045, "encoder_q-layer.9": 3266.0508, "epoch": 0.39, "inbatch_neg_score": 1.4514, "inbatch_pos_score": 2.0469, "learning_rate": 2.9e-05, "loss": 3.4794, "norm_diff": 0.0879, "norm_loss": 0.0, "num_token_doc": 66.8045, "num_token_overlap": 11.686, "num_token_query": 31.3689, "num_token_union": 65.1339, "num_word_context": 202.0765, "num_word_doc": 49.8421, "num_word_query": 23.2819, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6956.4064, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4521, "query_norm": 1.7248, "queue_k_norm": 1.8077, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3689, "sent_len_1": 66.8045, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.3262, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 89800 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4855, "doc_norm": 1.8121, "encoder_q-embeddings": 2588.7922, "encoder_q-layer.0": 1765.6331, "encoder_q-layer.1": 1920.4985, "encoder_q-layer.10": 3138.2844, "encoder_q-layer.11": 7245.8896, "encoder_q-layer.2": 2096.1062, "encoder_q-layer.3": 2122.0095, "encoder_q-layer.4": 2266.2341, "encoder_q-layer.5": 2367.3984, "encoder_q-layer.6": 2586.6948, "encoder_q-layer.7": 2859.5632, "encoder_q-layer.8": 3438.0488, "encoder_q-layer.9": 3185.4202, "epoch": 0.39, "inbatch_neg_score": 1.4521, "inbatch_pos_score": 2.0605, "learning_rate": 2.897368421052632e-05, "loss": 3.4855, "norm_diff": 0.0865, "norm_loss": 0.0, "num_token_doc": 66.7283, "num_token_overlap": 11.6563, "num_token_query": 31.3016, "num_token_union": 65.0598, "num_word_context": 202.1654, "num_word_doc": 49.7862, "num_word_query": 23.2388, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4850.5969, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4531, "query_norm": 1.7257, "queue_k_norm": 1.8093, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3016, "sent_len_1": 66.7283, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.6213, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 89900 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.4713, "doc_norm": 1.8078, "encoder_q-embeddings": 2657.6746, "encoder_q-layer.0": 1769.1694, "encoder_q-layer.1": 1955.954, "encoder_q-layer.10": 3326.4695, "encoder_q-layer.11": 7494.7593, "encoder_q-layer.2": 2176.2539, "encoder_q-layer.3": 2216.3618, "encoder_q-layer.4": 2339.4214, "encoder_q-layer.5": 2385.2375, "encoder_q-layer.6": 2695.3389, "encoder_q-layer.7": 3135.1514, "encoder_q-layer.8": 3790.1211, "encoder_q-layer.9": 3220.6079, "epoch": 0.39, "inbatch_neg_score": 1.4615, "inbatch_pos_score": 2.0547, "learning_rate": 2.8947368421052634e-05, "loss": 3.4713, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.7316, "num_token_overlap": 11.6662, "num_token_query": 31.3777, "num_token_union": 65.1041, "num_word_context": 202.5039, "num_word_doc": 49.7928, "num_word_query": 23.2948, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5014.8358, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4619, "query_norm": 1.729, "queue_k_norm": 1.8101, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3777, "sent_len_1": 66.7316, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.7725, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 90000 }, { "dev_runtime": 28.3472, "dev_samples_per_second": 1.129, "dev_steps_per_second": 0.035, "epoch": 0.39, "step": 90000, "test_accuracy": 93.8720703125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3485654592514038, "test_doc_norm": 1.7890076637268066, "test_inbatch_neg_score": 1.7462849617004395, "test_inbatch_pos_score": 2.7596383094787598, "test_loss": 0.3485654592514038, "test_loss_align": 1.081956148147583, "test_loss_unif": -0.08808627724647522, "test_loss_unif_q@queue": -0.08808627724647522, "test_norm_diff": 0.04971891641616821, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.4559266567230225, "test_query_norm": 1.8387267589569092, "test_queue_k_norm": 1.8101754188537598, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04388260096311569, "test_stdq": 0.04431743919849396, "test_stdqueue_k": 0.04911280795931816, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.3472, "dev_samples_per_second": 1.129, "dev_steps_per_second": 0.035, "epoch": 0.39, "eval_beir-arguana_ndcg@10": 0.38683, "eval_beir-arguana_recall@10": 0.65718, "eval_beir-arguana_recall@100": 0.93243, "eval_beir-arguana_recall@20": 0.78094, "eval_beir-avg_ndcg@10": 0.3865496666666667, "eval_beir-avg_recall@10": 0.4567119166666667, "eval_beir-avg_recall@100": 0.6391509166666667, "eval_beir-avg_recall@20": 0.5160510833333334, "eval_beir-cqadupstack_ndcg@10": 0.29077666666666663, "eval_beir-cqadupstack_recall@10": 0.3908891666666667, "eval_beir-cqadupstack_recall@100": 0.6240891666666667, "eval_beir-cqadupstack_recall@20": 0.45659083333333333, "eval_beir-fiqa_ndcg@10": 0.25897, "eval_beir-fiqa_recall@10": 0.31623, "eval_beir-fiqa_recall@100": 0.60657, "eval_beir-fiqa_recall@20": 0.38271, "eval_beir-nfcorpus_ndcg@10": 0.31729, "eval_beir-nfcorpus_recall@10": 0.15852, "eval_beir-nfcorpus_recall@100": 0.29411, "eval_beir-nfcorpus_recall@20": 0.18922, "eval_beir-nq_ndcg@10": 0.30862, "eval_beir-nq_recall@10": 0.49739, "eval_beir-nq_recall@100": 0.82298, "eval_beir-nq_recall@20": 0.61076, "eval_beir-quora_ndcg@10": 0.78726, "eval_beir-quora_recall@10": 0.89324, "eval_beir-quora_recall@100": 0.97966, "eval_beir-quora_recall@20": 0.9347, "eval_beir-scidocs_ndcg@10": 0.15904, "eval_beir-scidocs_recall@10": 0.16827, "eval_beir-scidocs_recall@100": 0.377, "eval_beir-scidocs_recall@20": 0.22807, "eval_beir-scifact_ndcg@10": 0.65137, "eval_beir-scifact_recall@10": 0.793, "eval_beir-scifact_recall@100": 0.90156, "eval_beir-scifact_recall@20": 0.83411, "eval_beir-trec-covid_ndcg@10": 0.54559, "eval_beir-trec-covid_recall@10": 0.574, "eval_beir-trec-covid_recall@100": 0.4432, "eval_beir-trec-covid_recall@20": 0.552, "eval_beir-webis-touche2020_ndcg@10": 0.15975, "eval_beir-webis-touche2020_recall@10": 0.1184, "eval_beir-webis-touche2020_recall@100": 0.40991, "eval_beir-webis-touche2020_recall@20": 0.19141, "eval_senteval-avg_sts": 0.7420996777382923, "eval_senteval-sickr_spearman": 0.7310734698577754, "eval_senteval-stsb_spearman": 0.7531258856188092, "step": 90000, "test_accuracy": 93.8720703125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3485654592514038, "test_doc_norm": 1.7890076637268066, "test_inbatch_neg_score": 1.7462849617004395, "test_inbatch_pos_score": 2.7596383094787598, "test_loss": 0.3485654592514038, "test_loss_align": 1.081956148147583, "test_loss_unif": -0.08808627724647522, "test_loss_unif_q@queue": -0.08808627724647522, "test_norm_diff": 0.04971891641616821, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.4559266567230225, "test_query_norm": 1.8387267589569092, "test_queue_k_norm": 1.8101754188537598, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04388260096311569, "test_stdq": 0.04431743919849396, "test_stdqueue_k": 0.04911280795931816, "test_stdqueue_q": 0.0 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.4697, "doc_norm": 1.8122, "encoder_q-embeddings": 2658.2603, "encoder_q-layer.0": 1839.2935, "encoder_q-layer.1": 2038.3943, "encoder_q-layer.10": 3011.7632, "encoder_q-layer.11": 7428.6221, "encoder_q-layer.2": 2210.9097, "encoder_q-layer.3": 2270.8745, "encoder_q-layer.4": 2406.3254, "encoder_q-layer.5": 2413.2727, "encoder_q-layer.6": 2734.5122, "encoder_q-layer.7": 3107.8865, "encoder_q-layer.8": 3522.0574, "encoder_q-layer.9": 3041.9512, "epoch": 0.39, "inbatch_neg_score": 1.4605, "inbatch_pos_score": 2.0605, "learning_rate": 2.892105263157895e-05, "loss": 3.4697, "norm_diff": 0.082, "norm_loss": 0.0, "num_token_doc": 66.7471, "num_token_overlap": 11.6818, "num_token_query": 31.3155, "num_token_union": 65.0466, "num_word_context": 202.1395, "num_word_doc": 49.806, "num_word_query": 23.2579, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4947.8352, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.459, "query_norm": 1.7302, "queue_k_norm": 1.8107, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3155, "sent_len_1": 66.7471, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.3713, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 90100 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4763, "doc_norm": 1.8139, "encoder_q-embeddings": 2650.4102, "encoder_q-layer.0": 1786.6178, "encoder_q-layer.1": 1940.314, "encoder_q-layer.10": 3166.8096, "encoder_q-layer.11": 7416.9072, "encoder_q-layer.2": 2187.9722, "encoder_q-layer.3": 2350.7297, "encoder_q-layer.4": 2514.4282, "encoder_q-layer.5": 2470.3123, "encoder_q-layer.6": 2803.7373, "encoder_q-layer.7": 3261.8418, "encoder_q-layer.8": 3814.6353, "encoder_q-layer.9": 3271.9058, "epoch": 0.39, "inbatch_neg_score": 1.4655, "inbatch_pos_score": 2.0625, "learning_rate": 2.8894736842105263e-05, "loss": 3.4763, "norm_diff": 0.0822, "norm_loss": 0.0, "num_token_doc": 66.8344, "num_token_overlap": 11.6431, "num_token_query": 31.3341, "num_token_union": 65.1793, "num_word_context": 202.4297, "num_word_doc": 49.8791, "num_word_query": 23.2755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5024.903, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4648, "query_norm": 1.7317, "queue_k_norm": 1.8139, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3341, "sent_len_1": 66.8344, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.515, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 90200 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.4691, "doc_norm": 1.8124, "encoder_q-embeddings": 2810.4067, "encoder_q-layer.0": 1972.6239, "encoder_q-layer.1": 2089.252, "encoder_q-layer.10": 3438.3569, "encoder_q-layer.11": 7353.145, "encoder_q-layer.2": 2362.6548, "encoder_q-layer.3": 2474.3916, "encoder_q-layer.4": 2711.1211, "encoder_q-layer.5": 2781.9648, "encoder_q-layer.6": 2900.4973, "encoder_q-layer.7": 3184.2026, "encoder_q-layer.8": 3620.5569, "encoder_q-layer.9": 2971.6042, "epoch": 0.39, "inbatch_neg_score": 1.4707, "inbatch_pos_score": 2.0742, "learning_rate": 2.886842105263158e-05, "loss": 3.4691, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.9929, "num_token_overlap": 11.7151, "num_token_query": 31.4795, "num_token_union": 65.2804, "num_word_context": 202.4495, "num_word_doc": 49.965, "num_word_query": 23.3878, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5043.4536, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 1.4707, "query_norm": 1.747, "queue_k_norm": 1.8144, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4795, "sent_len_1": 66.9929, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.7488, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 90300 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4731, "doc_norm": 1.8155, "encoder_q-embeddings": 6360.1289, "encoder_q-layer.0": 4497.7852, "encoder_q-layer.1": 4933.8335, "encoder_q-layer.10": 6281.1338, "encoder_q-layer.11": 14923.1523, "encoder_q-layer.2": 5785.0186, "encoder_q-layer.3": 6183.6914, "encoder_q-layer.4": 6453.8291, "encoder_q-layer.5": 6789.7896, "encoder_q-layer.6": 7418.417, "encoder_q-layer.7": 7866.8457, "encoder_q-layer.8": 8480.8008, "encoder_q-layer.9": 6185.6836, "epoch": 0.39, "inbatch_neg_score": 1.4806, "inbatch_pos_score": 2.0781, "learning_rate": 2.8842105263157897e-05, "loss": 3.4731, "norm_diff": 0.0623, "norm_loss": 0.0, "num_token_doc": 66.7756, "num_token_overlap": 11.6756, "num_token_query": 31.3794, "num_token_union": 65.1348, "num_word_context": 202.3564, "num_word_doc": 49.858, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11197.0744, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4814, "query_norm": 1.7531, "queue_k_norm": 1.8144, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3794, "sent_len_1": 66.7756, "sent_len_max_0": 128.0, "sent_len_max_1": 207.8013, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 90400 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.4595, "doc_norm": 1.8159, "encoder_q-embeddings": 5078.0337, "encoder_q-layer.0": 3516.8281, "encoder_q-layer.1": 3609.2102, "encoder_q-layer.10": 6497.793, "encoder_q-layer.11": 15783.9414, "encoder_q-layer.2": 4094.9695, "encoder_q-layer.3": 4307.5908, "encoder_q-layer.4": 4475.5903, "encoder_q-layer.5": 4691.6431, "encoder_q-layer.6": 5293.5205, "encoder_q-layer.7": 5763.728, "encoder_q-layer.8": 6983.2954, "encoder_q-layer.9": 6334.5356, "epoch": 0.39, "inbatch_neg_score": 1.4897, "inbatch_pos_score": 2.0801, "learning_rate": 2.8815789473684212e-05, "loss": 3.4595, "norm_diff": 0.0636, "norm_loss": 0.0, "num_token_doc": 66.775, "num_token_overlap": 11.6724, "num_token_query": 31.4237, "num_token_union": 65.1368, "num_word_context": 202.5464, "num_word_doc": 49.8156, "num_word_query": 23.3387, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9952.6754, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.4893, "query_norm": 1.7522, "queue_k_norm": 1.8165, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4237, "sent_len_1": 66.775, "sent_len_max_0": 127.9988, "sent_len_max_1": 210.2612, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 90500 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.4728, "doc_norm": 1.8161, "encoder_q-embeddings": 4948.8125, "encoder_q-layer.0": 3303.2102, "encoder_q-layer.1": 3604.2585, "encoder_q-layer.10": 6574.2529, "encoder_q-layer.11": 15125.21, "encoder_q-layer.2": 3950.7693, "encoder_q-layer.3": 4188.1289, "encoder_q-layer.4": 4357.3936, "encoder_q-layer.5": 4356.1738, "encoder_q-layer.6": 4788.811, "encoder_q-layer.7": 5611.4492, "encoder_q-layer.8": 6928.7778, "encoder_q-layer.9": 6260.6265, "epoch": 0.39, "inbatch_neg_score": 1.5005, "inbatch_pos_score": 2.0938, "learning_rate": 2.8789473684210527e-05, "loss": 3.4728, "norm_diff": 0.0451, "norm_loss": 0.0, "num_token_doc": 66.8131, "num_token_overlap": 11.6708, "num_token_query": 31.368, "num_token_union": 65.1331, "num_word_context": 202.449, "num_word_doc": 49.853, "num_word_query": 23.3126, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9537.5066, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5, "query_norm": 1.771, "queue_k_norm": 1.8151, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.368, "sent_len_1": 66.8131, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.045, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 90600 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.481, "doc_norm": 1.8229, "encoder_q-embeddings": 5066.3184, "encoder_q-layer.0": 3434.2549, "encoder_q-layer.1": 3640.2644, "encoder_q-layer.10": 6421.2642, "encoder_q-layer.11": 14724.1182, "encoder_q-layer.2": 4056.3533, "encoder_q-layer.3": 4111.2939, "encoder_q-layer.4": 4333.9297, "encoder_q-layer.5": 4445.9658, "encoder_q-layer.6": 4986.293, "encoder_q-layer.7": 5803.3477, "encoder_q-layer.8": 7015.9868, "encoder_q-layer.9": 6322.7412, "epoch": 0.39, "inbatch_neg_score": 1.5136, "inbatch_pos_score": 2.1211, "learning_rate": 2.876315789473684e-05, "loss": 3.481, "norm_diff": 0.0479, "norm_loss": 0.0, "num_token_doc": 66.5441, "num_token_overlap": 11.6577, "num_token_query": 31.3471, "num_token_union": 65.0162, "num_word_context": 202.2156, "num_word_doc": 49.68, "num_word_query": 23.2829, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9614.8436, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5137, "query_norm": 1.775, "queue_k_norm": 1.8182, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3471, "sent_len_1": 66.5441, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.5762, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 90700 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4658, "doc_norm": 1.8223, "encoder_q-embeddings": 4925.3652, "encoder_q-layer.0": 3334.522, "encoder_q-layer.1": 3580.3508, "encoder_q-layer.10": 7098.3887, "encoder_q-layer.11": 15333.2646, "encoder_q-layer.2": 3997.6431, "encoder_q-layer.3": 4166.2695, "encoder_q-layer.4": 4413.8486, "encoder_q-layer.5": 4626.4951, "encoder_q-layer.6": 5017.7124, "encoder_q-layer.7": 5893.6919, "encoder_q-layer.8": 6948.8291, "encoder_q-layer.9": 6539.8389, "epoch": 0.39, "inbatch_neg_score": 1.5262, "inbatch_pos_score": 2.123, "learning_rate": 2.8736842105263163e-05, "loss": 3.4658, "norm_diff": 0.0615, "norm_loss": 0.0, "num_token_doc": 66.7565, "num_token_overlap": 11.6796, "num_token_query": 31.4097, "num_token_union": 65.1128, "num_word_context": 202.2745, "num_word_doc": 49.8013, "num_word_query": 23.329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9807.1835, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5264, "query_norm": 1.7607, "queue_k_norm": 1.8203, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4097, "sent_len_1": 66.7565, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.2025, "stdk": 0.0492, "stdq": 0.0425, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 90800 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4758, "doc_norm": 1.8225, "encoder_q-embeddings": 4952.7949, "encoder_q-layer.0": 3392.0054, "encoder_q-layer.1": 3588.5083, "encoder_q-layer.10": 5902.1729, "encoder_q-layer.11": 14936.0918, "encoder_q-layer.2": 4019.0688, "encoder_q-layer.3": 4026.0649, "encoder_q-layer.4": 4252.4375, "encoder_q-layer.5": 4497.6299, "encoder_q-layer.6": 4976.1714, "encoder_q-layer.7": 5619.4629, "encoder_q-layer.8": 6374.0894, "encoder_q-layer.9": 5876.7886, "epoch": 0.39, "inbatch_neg_score": 1.5373, "inbatch_pos_score": 2.1309, "learning_rate": 2.8710526315789475e-05, "loss": 3.4758, "norm_diff": 0.056, "norm_loss": 0.0, "num_token_doc": 66.8676, "num_token_overlap": 11.6562, "num_token_query": 31.3821, "num_token_union": 65.2063, "num_word_context": 202.4462, "num_word_doc": 49.9018, "num_word_query": 23.3038, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9269.1735, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5361, "query_norm": 1.7665, "queue_k_norm": 1.8227, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3821, "sent_len_1": 66.8676, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.915, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 90900 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.4661, "doc_norm": 1.8201, "encoder_q-embeddings": 5196.2427, "encoder_q-layer.0": 3521.0093, "encoder_q-layer.1": 3830.1885, "encoder_q-layer.10": 6643.1069, "encoder_q-layer.11": 14502.0293, "encoder_q-layer.2": 4400.5947, "encoder_q-layer.3": 4532.71, "encoder_q-layer.4": 4857.4258, "encoder_q-layer.5": 4965.748, "encoder_q-layer.6": 5443.1372, "encoder_q-layer.7": 5769.4614, "encoder_q-layer.8": 6649.9453, "encoder_q-layer.9": 6507.4741, "epoch": 0.39, "inbatch_neg_score": 1.5493, "inbatch_pos_score": 2.1387, "learning_rate": 2.868421052631579e-05, "loss": 3.4661, "norm_diff": 0.0484, "norm_loss": 0.0, "num_token_doc": 66.767, "num_token_overlap": 11.6813, "num_token_query": 31.3899, "num_token_union": 65.1245, "num_word_context": 202.2975, "num_word_doc": 49.8564, "num_word_query": 23.3194, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9635.276, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5488, "query_norm": 1.7717, "queue_k_norm": 1.8236, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3899, "sent_len_1": 66.767, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.6062, "stdk": 0.0489, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 91000 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.4767, "doc_norm": 1.8259, "encoder_q-embeddings": 5220.1807, "encoder_q-layer.0": 3427.512, "encoder_q-layer.1": 3524.499, "encoder_q-layer.10": 6601.1128, "encoder_q-layer.11": 15537.2881, "encoder_q-layer.2": 3927.1948, "encoder_q-layer.3": 4054.0635, "encoder_q-layer.4": 4246.896, "encoder_q-layer.5": 4491.2119, "encoder_q-layer.6": 4997.4551, "encoder_q-layer.7": 5673.4648, "encoder_q-layer.8": 6990.5371, "encoder_q-layer.9": 6566.1021, "epoch": 0.4, "inbatch_neg_score": 1.5529, "inbatch_pos_score": 2.1465, "learning_rate": 2.8657894736842105e-05, "loss": 3.4767, "norm_diff": 0.0677, "norm_loss": 0.0, "num_token_doc": 66.9135, "num_token_overlap": 11.6809, "num_token_query": 31.4104, "num_token_union": 65.2433, "num_word_context": 202.4329, "num_word_doc": 49.9554, "num_word_query": 23.3372, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9804.0884, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5527, "query_norm": 1.7583, "queue_k_norm": 1.8277, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4104, "sent_len_1": 66.9135, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.6463, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91100 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4758, "doc_norm": 1.8294, "encoder_q-embeddings": 4926.9614, "encoder_q-layer.0": 3359.2737, "encoder_q-layer.1": 3510.5984, "encoder_q-layer.10": 6571.8286, "encoder_q-layer.11": 15756.3516, "encoder_q-layer.2": 3950.9724, "encoder_q-layer.3": 4243.5039, "encoder_q-layer.4": 4511.9531, "encoder_q-layer.5": 4806.4365, "encoder_q-layer.6": 5264.0864, "encoder_q-layer.7": 5808.4116, "encoder_q-layer.8": 6884.4473, "encoder_q-layer.9": 6233.9302, "epoch": 0.4, "inbatch_neg_score": 1.5542, "inbatch_pos_score": 2.1582, "learning_rate": 2.863157894736842e-05, "loss": 3.4758, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.7456, "num_token_overlap": 11.6368, "num_token_query": 31.2977, "num_token_union": 65.0955, "num_word_context": 202.3833, "num_word_doc": 49.7933, "num_word_query": 23.2345, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9746.5673, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5547, "query_norm": 1.7537, "queue_k_norm": 1.8284, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2977, "sent_len_1": 66.7456, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.405, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91200 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4785, "doc_norm": 1.8325, "encoder_q-embeddings": 5213.4316, "encoder_q-layer.0": 3541.0288, "encoder_q-layer.1": 3824.0088, "encoder_q-layer.10": 6660.375, "encoder_q-layer.11": 15142.7354, "encoder_q-layer.2": 4355.084, "encoder_q-layer.3": 4577.9834, "encoder_q-layer.4": 4847.3579, "encoder_q-layer.5": 5105.8013, "encoder_q-layer.6": 5520.4434, "encoder_q-layer.7": 6121.3721, "encoder_q-layer.8": 6741.4263, "encoder_q-layer.9": 6087.7959, "epoch": 0.4, "inbatch_neg_score": 1.5601, "inbatch_pos_score": 2.1523, "learning_rate": 2.860526315789474e-05, "loss": 3.4785, "norm_diff": 0.083, "norm_loss": 0.0, "num_token_doc": 66.6881, "num_token_overlap": 11.6481, "num_token_query": 31.3223, "num_token_union": 65.0702, "num_word_context": 202.1216, "num_word_doc": 49.7586, "num_word_query": 23.2401, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9916.3123, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5596, "query_norm": 1.7495, "queue_k_norm": 1.8303, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3223, "sent_len_1": 66.6881, "sent_len_max_0": 127.96, "sent_len_max_1": 208.7388, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91300 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.477, "doc_norm": 1.8325, "encoder_q-embeddings": 6118.4453, "encoder_q-layer.0": 4350.3936, "encoder_q-layer.1": 4583.0137, "encoder_q-layer.10": 6255.2183, "encoder_q-layer.11": 14997.3398, "encoder_q-layer.2": 5872.1675, "encoder_q-layer.3": 6176.5356, "encoder_q-layer.4": 5233.2891, "encoder_q-layer.5": 5101.3066, "encoder_q-layer.6": 5742.8896, "encoder_q-layer.7": 6086.3408, "encoder_q-layer.8": 6896.502, "encoder_q-layer.9": 6049.3877, "epoch": 0.4, "inbatch_neg_score": 1.5627, "inbatch_pos_score": 2.1758, "learning_rate": 2.8578947368421057e-05, "loss": 3.477, "norm_diff": 0.0739, "norm_loss": 0.0, "num_token_doc": 66.9312, "num_token_overlap": 11.6837, "num_token_query": 31.4065, "num_token_union": 65.211, "num_word_context": 202.6884, "num_word_doc": 49.8846, "num_word_query": 23.3153, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10440.295, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5625, "query_norm": 1.7586, "queue_k_norm": 1.8335, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4065, "sent_len_1": 66.9312, "sent_len_max_0": 127.99, "sent_len_max_1": 210.8325, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91400 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4694, "doc_norm": 1.8352, "encoder_q-embeddings": 4926.4009, "encoder_q-layer.0": 3373.7664, "encoder_q-layer.1": 3573.0923, "encoder_q-layer.10": 6287.2686, "encoder_q-layer.11": 14636.1826, "encoder_q-layer.2": 4013.2524, "encoder_q-layer.3": 4097.8188, "encoder_q-layer.4": 4438.231, "encoder_q-layer.5": 4414.5684, "encoder_q-layer.6": 5103.7153, "encoder_q-layer.7": 5433.209, "encoder_q-layer.8": 6604.1797, "encoder_q-layer.9": 6080.0356, "epoch": 0.4, "inbatch_neg_score": 1.5658, "inbatch_pos_score": 2.168, "learning_rate": 2.8552631578947368e-05, "loss": 3.4694, "norm_diff": 0.0828, "norm_loss": 0.0, "num_token_doc": 66.7928, "num_token_overlap": 11.6619, "num_token_query": 31.3263, "num_token_union": 65.1149, "num_word_context": 202.3213, "num_word_doc": 49.8273, "num_word_query": 23.264, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9494.28, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5664, "query_norm": 1.7523, "queue_k_norm": 1.8325, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3263, "sent_len_1": 66.7928, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.2512, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 91500 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.4578, "doc_norm": 1.8373, "encoder_q-embeddings": 4973.4648, "encoder_q-layer.0": 3410.5007, "encoder_q-layer.1": 3537.5215, "encoder_q-layer.10": 6131.1113, "encoder_q-layer.11": 14513.2588, "encoder_q-layer.2": 3831.3142, "encoder_q-layer.3": 3864.5945, "encoder_q-layer.4": 4108.2334, "encoder_q-layer.5": 4163.1064, "encoder_q-layer.6": 4748.5078, "encoder_q-layer.7": 5513.917, "encoder_q-layer.8": 6336.2061, "encoder_q-layer.9": 5822.1323, "epoch": 0.4, "inbatch_neg_score": 1.5691, "inbatch_pos_score": 2.1719, "learning_rate": 2.8526315789473683e-05, "loss": 3.4578, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.7041, "num_token_overlap": 11.6593, "num_token_query": 31.295, "num_token_union": 65.0973, "num_word_context": 202.3862, "num_word_doc": 49.8059, "num_word_query": 23.2472, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9129.9927, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5684, "query_norm": 1.7527, "queue_k_norm": 1.8358, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.295, "sent_len_1": 66.7041, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.0462, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91600 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.458, "doc_norm": 1.8411, "encoder_q-embeddings": 5010.1411, "encoder_q-layer.0": 3383.7639, "encoder_q-layer.1": 3583.5742, "encoder_q-layer.10": 6419.3706, "encoder_q-layer.11": 15259.666, "encoder_q-layer.2": 3986.7532, "encoder_q-layer.3": 4021.6265, "encoder_q-layer.4": 4432.457, "encoder_q-layer.5": 4488.5039, "encoder_q-layer.6": 5008.7964, "encoder_q-layer.7": 5479.3442, "encoder_q-layer.8": 6601.9375, "encoder_q-layer.9": 5980.7456, "epoch": 0.4, "inbatch_neg_score": 1.5668, "inbatch_pos_score": 2.168, "learning_rate": 2.8499999999999998e-05, "loss": 3.458, "norm_diff": 0.0872, "norm_loss": 0.0, "num_token_doc": 66.7874, "num_token_overlap": 11.6847, "num_token_query": 31.4036, "num_token_union": 65.1449, "num_word_context": 202.27, "num_word_doc": 49.8318, "num_word_query": 23.3215, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9692.3552, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5664, "query_norm": 1.7538, "queue_k_norm": 1.8369, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4036, "sent_len_1": 66.7874, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.3325, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 91700 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4803, "doc_norm": 1.8391, "encoder_q-embeddings": 5605.8477, "encoder_q-layer.0": 3752.5051, "encoder_q-layer.1": 4102.8354, "encoder_q-layer.10": 6365.4199, "encoder_q-layer.11": 14774.5674, "encoder_q-layer.2": 4834.9365, "encoder_q-layer.3": 4900.752, "encoder_q-layer.4": 5170.3896, "encoder_q-layer.5": 5023.6973, "encoder_q-layer.6": 5568.0923, "encoder_q-layer.7": 6448.0034, "encoder_q-layer.8": 7291.834, "encoder_q-layer.9": 6406.7114, "epoch": 0.4, "inbatch_neg_score": 1.5702, "inbatch_pos_score": 2.168, "learning_rate": 2.847368421052632e-05, "loss": 3.4803, "norm_diff": 0.0862, "norm_loss": 0.0, "num_token_doc": 66.833, "num_token_overlap": 11.6429, "num_token_query": 31.3286, "num_token_union": 65.1325, "num_word_context": 202.3805, "num_word_doc": 49.8449, "num_word_query": 23.2753, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10057.8908, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5703, "query_norm": 1.753, "queue_k_norm": 1.8375, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3286, "sent_len_1": 66.833, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.0375, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 91800 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4686, "doc_norm": 1.8398, "encoder_q-embeddings": 5137.7783, "encoder_q-layer.0": 3452.2493, "encoder_q-layer.1": 3650.6223, "encoder_q-layer.10": 7210.0308, "encoder_q-layer.11": 15947.3086, "encoder_q-layer.2": 4154.9033, "encoder_q-layer.3": 4597.5425, "encoder_q-layer.4": 4806.2358, "encoder_q-layer.5": 4902.3989, "encoder_q-layer.6": 5466.3936, "encoder_q-layer.7": 6208.0415, "encoder_q-layer.8": 7354.4468, "encoder_q-layer.9": 6808.7876, "epoch": 0.4, "inbatch_neg_score": 1.5764, "inbatch_pos_score": 2.1914, "learning_rate": 2.8447368421052635e-05, "loss": 3.4686, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.621, "num_token_overlap": 11.7097, "num_token_query": 31.4575, "num_token_union": 65.0673, "num_word_context": 202.0738, "num_word_doc": 49.725, "num_word_query": 23.3691, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10057.7586, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5752, "query_norm": 1.7739, "queue_k_norm": 1.8379, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4575, "sent_len_1": 66.621, "sent_len_max_0": 127.97, "sent_len_max_1": 207.73, "stdk": 0.0491, "stdq": 0.044, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 91900 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.4691, "doc_norm": 1.8427, "encoder_q-embeddings": 16328.9297, "encoder_q-layer.0": 11363.0615, "encoder_q-layer.1": 12721.1035, "encoder_q-layer.10": 6112.8491, "encoder_q-layer.11": 15094.8867, "encoder_q-layer.2": 15105.9102, "encoder_q-layer.3": 17918.4219, "encoder_q-layer.4": 20370.9492, "encoder_q-layer.5": 24882.4355, "encoder_q-layer.6": 25260.9902, "encoder_q-layer.7": 22390.4688, "encoder_q-layer.8": 18693.4258, "encoder_q-layer.9": 8487.4043, "epoch": 0.4, "inbatch_neg_score": 1.582, "inbatch_pos_score": 2.1797, "learning_rate": 2.842105263157895e-05, "loss": 3.4691, "norm_diff": 0.093, "norm_loss": 0.0, "num_token_doc": 66.8034, "num_token_overlap": 11.7162, "num_token_query": 31.4485, "num_token_union": 65.1547, "num_word_context": 202.2235, "num_word_doc": 49.8513, "num_word_query": 23.3693, "postclip_grad_norm": 1.0, "preclip_grad_norm": 25840.5651, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.582, "query_norm": 1.7497, "queue_k_norm": 1.8391, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4485, "sent_len_1": 66.8034, "sent_len_max_0": 128.0, "sent_len_max_1": 209.575, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 92000 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4676, "doc_norm": 1.8421, "encoder_q-embeddings": 5399.4263, "encoder_q-layer.0": 3514.5376, "encoder_q-layer.1": 3840.3328, "encoder_q-layer.10": 7021.4092, "encoder_q-layer.11": 15953.7959, "encoder_q-layer.2": 4311.9375, "encoder_q-layer.3": 4579.0298, "encoder_q-layer.4": 4983.6309, "encoder_q-layer.5": 5261.2979, "encoder_q-layer.6": 5739.6875, "encoder_q-layer.7": 6308.3022, "encoder_q-layer.8": 7729.1104, "encoder_q-layer.9": 6642.2739, "epoch": 0.4, "inbatch_neg_score": 1.5833, "inbatch_pos_score": 2.1855, "learning_rate": 2.839473684210526e-05, "loss": 3.4676, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.8951, "num_token_overlap": 11.6779, "num_token_query": 31.3706, "num_token_union": 65.1846, "num_word_context": 202.3991, "num_word_doc": 49.9228, "num_word_query": 23.2945, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10594.8242, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.583, "query_norm": 1.7597, "queue_k_norm": 1.8418, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3706, "sent_len_1": 66.8951, "sent_len_max_0": 127.9425, "sent_len_max_1": 209.015, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92100 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.4709, "doc_norm": 1.8468, "encoder_q-embeddings": 7362.1509, "encoder_q-layer.0": 4974.4517, "encoder_q-layer.1": 5723.0322, "encoder_q-layer.10": 6045.5801, "encoder_q-layer.11": 14666.9141, "encoder_q-layer.2": 6896.3369, "encoder_q-layer.3": 7492.0488, "encoder_q-layer.4": 7797.8589, "encoder_q-layer.5": 7706.3789, "encoder_q-layer.6": 8907.9785, "encoder_q-layer.7": 9105.1309, "encoder_q-layer.8": 8411.5186, "encoder_q-layer.9": 6226.3267, "epoch": 0.4, "inbatch_neg_score": 1.5854, "inbatch_pos_score": 2.1953, "learning_rate": 2.8368421052631583e-05, "loss": 3.4709, "norm_diff": 0.0871, "norm_loss": 0.0, "num_token_doc": 66.752, "num_token_overlap": 11.6708, "num_token_query": 31.3327, "num_token_union": 65.0766, "num_word_context": 202.4571, "num_word_doc": 49.8009, "num_word_query": 23.2486, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12241.6851, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.584, "query_norm": 1.7597, "queue_k_norm": 1.8443, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3327, "sent_len_1": 66.752, "sent_len_max_0": 127.975, "sent_len_max_1": 209.14, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92200 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.4661, "doc_norm": 1.8425, "encoder_q-embeddings": 5014.1812, "encoder_q-layer.0": 3429.2698, "encoder_q-layer.1": 3779.9314, "encoder_q-layer.10": 6606.1318, "encoder_q-layer.11": 15506.335, "encoder_q-layer.2": 4300.5034, "encoder_q-layer.3": 4329.3008, "encoder_q-layer.4": 4588.2373, "encoder_q-layer.5": 4815.0425, "encoder_q-layer.6": 5115.2432, "encoder_q-layer.7": 5844.1606, "encoder_q-layer.8": 6964.0342, "encoder_q-layer.9": 6190.6963, "epoch": 0.4, "inbatch_neg_score": 1.5852, "inbatch_pos_score": 2.1992, "learning_rate": 2.8342105263157898e-05, "loss": 3.4661, "norm_diff": 0.0803, "norm_loss": 0.0, "num_token_doc": 66.7315, "num_token_overlap": 11.6958, "num_token_query": 31.477, "num_token_union": 65.1838, "num_word_context": 202.4278, "num_word_doc": 49.7905, "num_word_query": 23.3873, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9978.4787, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.585, "query_norm": 1.7622, "queue_k_norm": 1.844, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.477, "sent_len_1": 66.7315, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.2612, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 92300 }, { "accuracy": 47.998, "active_queue_size": 16384.0, "cl_loss": 3.472, "doc_norm": 1.846, "encoder_q-embeddings": 10666.2188, "encoder_q-layer.0": 7444.2803, "encoder_q-layer.1": 8307.5703, "encoder_q-layer.10": 12718.5557, "encoder_q-layer.11": 29548.6328, "encoder_q-layer.2": 9694.376, "encoder_q-layer.3": 10346.9482, "encoder_q-layer.4": 11075.9668, "encoder_q-layer.5": 11183.4326, "encoder_q-layer.6": 11753.4531, "encoder_q-layer.7": 12109.3291, "encoder_q-layer.8": 13555.334, "encoder_q-layer.9": 12360.0088, "epoch": 0.4, "inbatch_neg_score": 1.5885, "inbatch_pos_score": 2.1719, "learning_rate": 2.8315789473684213e-05, "loss": 3.472, "norm_diff": 0.1024, "norm_loss": 0.0, "num_token_doc": 66.5823, "num_token_overlap": 11.6211, "num_token_query": 31.3049, "num_token_union": 65.0143, "num_word_context": 202.1602, "num_word_doc": 49.6776, "num_word_query": 23.249, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20128.2174, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5889, "query_norm": 1.7437, "queue_k_norm": 1.847, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3049, "sent_len_1": 66.5823, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.5662, "stdk": 0.0492, "stdq": 0.0424, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92400 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.4745, "doc_norm": 1.8454, "encoder_q-embeddings": 11909.9707, "encoder_q-layer.0": 8559.084, "encoder_q-layer.1": 8752.8652, "encoder_q-layer.10": 12298.9834, "encoder_q-layer.11": 29135.2539, "encoder_q-layer.2": 10214.9971, "encoder_q-layer.3": 10315.6045, "encoder_q-layer.4": 11290.7246, "encoder_q-layer.5": 11130.0684, "encoder_q-layer.6": 11685.5664, "encoder_q-layer.7": 12186.3574, "encoder_q-layer.8": 13565.4727, "encoder_q-layer.9": 11880.625, "epoch": 0.4, "inbatch_neg_score": 1.5868, "inbatch_pos_score": 2.1797, "learning_rate": 2.8289473684210528e-05, "loss": 3.4745, "norm_diff": 0.0963, "norm_loss": 0.0, "num_token_doc": 66.9445, "num_token_overlap": 11.7039, "num_token_query": 31.3289, "num_token_union": 65.1741, "num_word_context": 202.5993, "num_word_doc": 49.9579, "num_word_query": 23.2661, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20367.3161, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5859, "query_norm": 1.7492, "queue_k_norm": 1.8495, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3289, "sent_len_1": 66.9445, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.8725, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 92500 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.4734, "doc_norm": 1.8493, "encoder_q-embeddings": 9289.7236, "encoder_q-layer.0": 6379.4375, "encoder_q-layer.1": 6999.3955, "encoder_q-layer.10": 13512.1729, "encoder_q-layer.11": 30316.1777, "encoder_q-layer.2": 7476.5972, "encoder_q-layer.3": 7743.644, "encoder_q-layer.4": 8144.6919, "encoder_q-layer.5": 8600.7656, "encoder_q-layer.6": 9823.2998, "encoder_q-layer.7": 11473.335, "encoder_q-layer.8": 13191.1582, "encoder_q-layer.9": 12324.127, "epoch": 0.4, "inbatch_neg_score": 1.5885, "inbatch_pos_score": 2.1914, "learning_rate": 2.826315789473684e-05, "loss": 3.4734, "norm_diff": 0.0925, "norm_loss": 0.0, "num_token_doc": 66.5882, "num_token_overlap": 11.632, "num_token_query": 31.2764, "num_token_union": 64.9621, "num_word_context": 201.8795, "num_word_doc": 49.6621, "num_word_query": 23.2234, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18848.5176, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5879, "query_norm": 1.7567, "queue_k_norm": 1.8473, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2764, "sent_len_1": 66.5882, "sent_len_max_0": 127.9975, "sent_len_max_1": 211.11, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92600 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4639, "doc_norm": 1.8525, "encoder_q-embeddings": 9661.6709, "encoder_q-layer.0": 6866.5015, "encoder_q-layer.1": 7212.6299, "encoder_q-layer.10": 12264.2236, "encoder_q-layer.11": 28273.1172, "encoder_q-layer.2": 8169.9385, "encoder_q-layer.3": 8139.1387, "encoder_q-layer.4": 8806.6045, "encoder_q-layer.5": 8787.3594, "encoder_q-layer.6": 9788.9814, "encoder_q-layer.7": 10848.7539, "encoder_q-layer.8": 12617.1758, "encoder_q-layer.9": 12025.5615, "epoch": 0.4, "inbatch_neg_score": 1.5913, "inbatch_pos_score": 2.1914, "learning_rate": 2.823684210526316e-05, "loss": 3.4639, "norm_diff": 0.1001, "norm_loss": 0.0, "num_token_doc": 66.8868, "num_token_overlap": 11.711, "num_token_query": 31.3453, "num_token_union": 65.1689, "num_word_context": 202.4202, "num_word_doc": 49.917, "num_word_query": 23.2776, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18444.536, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5908, "query_norm": 1.7524, "queue_k_norm": 1.8502, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3453, "sent_len_1": 66.8868, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.2925, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 92700 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.4785, "doc_norm": 1.8536, "encoder_q-embeddings": 14967.1924, "encoder_q-layer.0": 10696.1104, "encoder_q-layer.1": 11485.2012, "encoder_q-layer.10": 13068.0127, "encoder_q-layer.11": 29016.7598, "encoder_q-layer.2": 12948.543, "encoder_q-layer.3": 14146.0176, "encoder_q-layer.4": 15780.6143, "encoder_q-layer.5": 17301.8203, "encoder_q-layer.6": 18149.998, "encoder_q-layer.7": 15855.4521, "encoder_q-layer.8": 14758.0518, "encoder_q-layer.9": 12220.8857, "epoch": 0.4, "inbatch_neg_score": 1.5907, "inbatch_pos_score": 2.1836, "learning_rate": 2.8210526315789476e-05, "loss": 3.4785, "norm_diff": 0.1056, "norm_loss": 0.0, "num_token_doc": 66.8724, "num_token_overlap": 11.6632, "num_token_query": 31.3373, "num_token_union": 65.1453, "num_word_context": 202.1924, "num_word_doc": 49.8625, "num_word_query": 23.2747, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23942.9673, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5918, "query_norm": 1.7479, "queue_k_norm": 1.8503, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3373, "sent_len_1": 66.8724, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.065, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92800 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.4931, "doc_norm": 1.8542, "encoder_q-embeddings": 10343.7881, "encoder_q-layer.0": 6941.729, "encoder_q-layer.1": 7399.0239, "encoder_q-layer.10": 14289.8613, "encoder_q-layer.11": 30378.9844, "encoder_q-layer.2": 8471.0479, "encoder_q-layer.3": 8592.0518, "encoder_q-layer.4": 9248.6562, "encoder_q-layer.5": 9378.5938, "encoder_q-layer.6": 10786.1289, "encoder_q-layer.7": 11737.8848, "encoder_q-layer.8": 13945.9092, "encoder_q-layer.9": 12947.3428, "epoch": 0.4, "inbatch_neg_score": 1.5913, "inbatch_pos_score": 2.1855, "learning_rate": 2.818421052631579e-05, "loss": 3.4931, "norm_diff": 0.0968, "norm_loss": 0.0, "num_token_doc": 66.7049, "num_token_overlap": 11.6248, "num_token_query": 31.296, "num_token_union": 65.0995, "num_word_context": 202.2762, "num_word_doc": 49.7916, "num_word_query": 23.2471, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19833.562, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5918, "query_norm": 1.7574, "queue_k_norm": 1.85, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.296, "sent_len_1": 66.7049, "sent_len_max_0": 127.9938, "sent_len_max_1": 206.6488, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 92900 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.467, "doc_norm": 1.8483, "encoder_q-embeddings": 9843.3037, "encoder_q-layer.0": 6772.1807, "encoder_q-layer.1": 6950.9722, "encoder_q-layer.10": 12839.1152, "encoder_q-layer.11": 30671.1562, "encoder_q-layer.2": 7693.1621, "encoder_q-layer.3": 8022.1772, "encoder_q-layer.4": 8491.1641, "encoder_q-layer.5": 8600.9922, "encoder_q-layer.6": 9919.2686, "encoder_q-layer.7": 11941.4951, "encoder_q-layer.8": 13171.6406, "encoder_q-layer.9": 12361.2783, "epoch": 0.4, "inbatch_neg_score": 1.5949, "inbatch_pos_score": 2.1895, "learning_rate": 2.8157894736842106e-05, "loss": 3.467, "norm_diff": 0.0972, "norm_loss": 0.0, "num_token_doc": 66.6374, "num_token_overlap": 11.644, "num_token_query": 31.2547, "num_token_union": 65.0126, "num_word_context": 202.1688, "num_word_doc": 49.7549, "num_word_query": 23.2092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19209.8001, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.5957, "query_norm": 1.751, "queue_k_norm": 1.8507, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2547, "sent_len_1": 66.6374, "sent_len_max_0": 127.9775, "sent_len_max_1": 206.2688, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 93000 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4735, "doc_norm": 1.8498, "encoder_q-embeddings": 4556.8481, "encoder_q-layer.0": 3200.667, "encoder_q-layer.1": 3420.78, "encoder_q-layer.10": 6129.8999, "encoder_q-layer.11": 14371.4932, "encoder_q-layer.2": 3777.9438, "encoder_q-layer.3": 3976.3354, "encoder_q-layer.4": 4229.8364, "encoder_q-layer.5": 4278.396, "encoder_q-layer.6": 4742.5869, "encoder_q-layer.7": 5121.9702, "encoder_q-layer.8": 6462.7373, "encoder_q-layer.9": 6038.7358, "epoch": 0.4, "inbatch_neg_score": 1.5987, "inbatch_pos_score": 2.1953, "learning_rate": 2.813157894736842e-05, "loss": 3.4735, "norm_diff": 0.1036, "norm_loss": 0.0, "num_token_doc": 66.7029, "num_token_overlap": 11.6553, "num_token_query": 31.3125, "num_token_union": 65.0598, "num_word_context": 202.1176, "num_word_doc": 49.7894, "num_word_query": 23.2531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9176.0851, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5996, "query_norm": 1.7462, "queue_k_norm": 1.852, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3125, "sent_len_1": 66.7029, "sent_len_max_0": 127.985, "sent_len_max_1": 207.8512, "stdk": 0.0491, "stdq": 0.0424, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 93100 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4712, "doc_norm": 1.8488, "encoder_q-embeddings": 5019.3159, "encoder_q-layer.0": 3430.6348, "encoder_q-layer.1": 3768.0342, "encoder_q-layer.10": 5998.9248, "encoder_q-layer.11": 14473.4102, "encoder_q-layer.2": 4196.7046, "encoder_q-layer.3": 4129.9907, "encoder_q-layer.4": 4340.7231, "encoder_q-layer.5": 4542.2861, "encoder_q-layer.6": 4979.7842, "encoder_q-layer.7": 5517.9248, "encoder_q-layer.8": 6371.4019, "encoder_q-layer.9": 6027.1367, "epoch": 0.4, "inbatch_neg_score": 1.5995, "inbatch_pos_score": 2.2051, "learning_rate": 2.810526315789474e-05, "loss": 3.4712, "norm_diff": 0.0889, "norm_loss": 0.0, "num_token_doc": 66.7856, "num_token_overlap": 11.6782, "num_token_query": 31.3499, "num_token_union": 65.1127, "num_word_context": 202.1858, "num_word_doc": 49.8091, "num_word_query": 23.293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9340.5922, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.5996, "query_norm": 1.7599, "queue_k_norm": 1.8536, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3499, "sent_len_1": 66.7856, "sent_len_max_0": 127.96, "sent_len_max_1": 209.9375, "stdk": 0.049, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 93200 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4824, "doc_norm": 1.8554, "encoder_q-embeddings": 5293.1865, "encoder_q-layer.0": 3479.5903, "encoder_q-layer.1": 3722.2036, "encoder_q-layer.10": 6808.8477, "encoder_q-layer.11": 15758.7773, "encoder_q-layer.2": 4210.7544, "encoder_q-layer.3": 4409.7129, "encoder_q-layer.4": 4635.4893, "encoder_q-layer.5": 4621.0933, "encoder_q-layer.6": 5229.9307, "encoder_q-layer.7": 6014.231, "encoder_q-layer.8": 6858.5957, "encoder_q-layer.9": 6329.1343, "epoch": 0.4, "inbatch_neg_score": 1.6047, "inbatch_pos_score": 2.1973, "learning_rate": 2.8078947368421055e-05, "loss": 3.4824, "norm_diff": 0.0953, "norm_loss": 0.0, "num_token_doc": 66.4641, "num_token_overlap": 11.6362, "num_token_query": 31.2979, "num_token_union": 64.9101, "num_word_context": 201.9445, "num_word_doc": 49.5772, "num_word_query": 23.2344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10053.6176, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6055, "query_norm": 1.7601, "queue_k_norm": 1.8522, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2979, "sent_len_1": 66.4641, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.5775, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 93300 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.4628, "doc_norm": 1.8552, "encoder_q-embeddings": 5112.4668, "encoder_q-layer.0": 3507.5273, "encoder_q-layer.1": 3684.3293, "encoder_q-layer.10": 6379.7808, "encoder_q-layer.11": 14678.8672, "encoder_q-layer.2": 4170.1479, "encoder_q-layer.3": 4274.1953, "encoder_q-layer.4": 4638.0332, "encoder_q-layer.5": 4606.1221, "encoder_q-layer.6": 5172.7598, "encoder_q-layer.7": 5579.248, "encoder_q-layer.8": 6572.7554, "encoder_q-layer.9": 5969.874, "epoch": 0.41, "inbatch_neg_score": 1.6082, "inbatch_pos_score": 2.207, "learning_rate": 2.805263157894737e-05, "loss": 3.4628, "norm_diff": 0.0759, "norm_loss": 0.0, "num_token_doc": 66.8497, "num_token_overlap": 11.6529, "num_token_query": 31.2637, "num_token_union": 65.1183, "num_word_context": 202.2252, "num_word_doc": 49.8615, "num_word_query": 23.2073, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9540.353, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6074, "query_norm": 1.7793, "queue_k_norm": 1.8546, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2637, "sent_len_1": 66.8497, "sent_len_max_0": 127.9912, "sent_len_max_1": 211.8113, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 93400 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.467, "doc_norm": 1.8577, "encoder_q-embeddings": 5011.9927, "encoder_q-layer.0": 3323.0103, "encoder_q-layer.1": 3490.4963, "encoder_q-layer.10": 6597.6577, "encoder_q-layer.11": 15052.79, "encoder_q-layer.2": 3882.7688, "encoder_q-layer.3": 3966.832, "encoder_q-layer.4": 4353.2563, "encoder_q-layer.5": 4205.3555, "encoder_q-layer.6": 4799.3735, "encoder_q-layer.7": 5605.3208, "encoder_q-layer.8": 6674.8496, "encoder_q-layer.9": 6125.417, "epoch": 0.41, "inbatch_neg_score": 1.6163, "inbatch_pos_score": 2.2168, "learning_rate": 2.8026315789473685e-05, "loss": 3.467, "norm_diff": 0.0943, "norm_loss": 0.0, "num_token_doc": 66.6811, "num_token_overlap": 11.6945, "num_token_query": 31.4107, "num_token_union": 65.0524, "num_word_context": 201.9212, "num_word_doc": 49.7503, "num_word_query": 23.3359, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9547.072, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6162, "query_norm": 1.7634, "queue_k_norm": 1.8532, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4107, "sent_len_1": 66.6811, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.9025, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 93500 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4539, "doc_norm": 1.8505, "encoder_q-embeddings": 5148.7837, "encoder_q-layer.0": 3582.1572, "encoder_q-layer.1": 3796.292, "encoder_q-layer.10": 6602.9741, "encoder_q-layer.11": 15171.3164, "encoder_q-layer.2": 4167.2012, "encoder_q-layer.3": 4331.8574, "encoder_q-layer.4": 4537.5557, "encoder_q-layer.5": 4539.4771, "encoder_q-layer.6": 5046.7437, "encoder_q-layer.7": 5918.021, "encoder_q-layer.8": 7084.2158, "encoder_q-layer.9": 6438.7534, "epoch": 0.41, "inbatch_neg_score": 1.6218, "inbatch_pos_score": 2.2227, "learning_rate": 2.8000000000000003e-05, "loss": 3.4539, "norm_diff": 0.0635, "norm_loss": 0.0, "num_token_doc": 66.9195, "num_token_overlap": 11.7028, "num_token_query": 31.4453, "num_token_union": 65.2342, "num_word_context": 202.5921, "num_word_doc": 49.9702, "num_word_query": 23.3441, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9850.1557, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6211, "query_norm": 1.787, "queue_k_norm": 1.8566, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4453, "sent_len_1": 66.9195, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.1875, "stdk": 0.0489, "stdq": 0.0436, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 93600 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4631, "doc_norm": 1.8555, "encoder_q-embeddings": 5213.667, "encoder_q-layer.0": 3503.3076, "encoder_q-layer.1": 3773.6792, "encoder_q-layer.10": 6905.1128, "encoder_q-layer.11": 15246.1016, "encoder_q-layer.2": 4240.8115, "encoder_q-layer.3": 4577.8989, "encoder_q-layer.4": 4851.9214, "encoder_q-layer.5": 5047.2007, "encoder_q-layer.6": 5342.5488, "encoder_q-layer.7": 6020.085, "encoder_q-layer.8": 7060.3765, "encoder_q-layer.9": 6576.9004, "epoch": 0.41, "inbatch_neg_score": 1.6262, "inbatch_pos_score": 2.2441, "learning_rate": 2.7973684210526318e-05, "loss": 3.4631, "norm_diff": 0.0583, "norm_loss": 0.0, "num_token_doc": 66.7318, "num_token_overlap": 11.6968, "num_token_query": 31.3969, "num_token_union": 65.063, "num_word_context": 202.023, "num_word_doc": 49.8077, "num_word_query": 23.3249, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9851.3457, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.626, "query_norm": 1.7972, "queue_k_norm": 1.8551, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3969, "sent_len_1": 66.7318, "sent_len_max_0": 127.985, "sent_len_max_1": 207.6725, "stdk": 0.0491, "stdq": 0.0439, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 93700 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.4683, "doc_norm": 1.8567, "encoder_q-embeddings": 5169.1689, "encoder_q-layer.0": 3569.332, "encoder_q-layer.1": 3736.3647, "encoder_q-layer.10": 6371.2686, "encoder_q-layer.11": 14791.5947, "encoder_q-layer.2": 4437.3286, "encoder_q-layer.3": 4599.228, "encoder_q-layer.4": 4839.6006, "encoder_q-layer.5": 4991.7935, "encoder_q-layer.6": 5690.6104, "encoder_q-layer.7": 6252.8477, "encoder_q-layer.8": 7008.5552, "encoder_q-layer.9": 6169.6357, "epoch": 0.41, "inbatch_neg_score": 1.6397, "inbatch_pos_score": 2.2344, "learning_rate": 2.7947368421052633e-05, "loss": 3.4683, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7462, "num_token_overlap": 11.7022, "num_token_query": 31.4976, "num_token_union": 65.1682, "num_word_context": 202.4295, "num_word_doc": 49.7954, "num_word_query": 23.4059, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9687.5858, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6406, "query_norm": 1.7872, "queue_k_norm": 1.8571, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4976, "sent_len_1": 66.7462, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.4, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 93800 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.4678, "doc_norm": 1.8591, "encoder_q-embeddings": 5046.8745, "encoder_q-layer.0": 3376.7537, "encoder_q-layer.1": 3562.0269, "encoder_q-layer.10": 6239.8101, "encoder_q-layer.11": 14693.6758, "encoder_q-layer.2": 4044.314, "encoder_q-layer.3": 4176.3169, "encoder_q-layer.4": 4497.0337, "encoder_q-layer.5": 4632.7661, "encoder_q-layer.6": 5154.8359, "encoder_q-layer.7": 5569.5088, "encoder_q-layer.8": 6727.4624, "encoder_q-layer.9": 6079.5381, "epoch": 0.41, "inbatch_neg_score": 1.6497, "inbatch_pos_score": 2.2598, "learning_rate": 2.7921052631578948e-05, "loss": 3.4678, "norm_diff": 0.0546, "norm_loss": 0.0, "num_token_doc": 66.7083, "num_token_overlap": 11.6592, "num_token_query": 31.3607, "num_token_union": 65.0721, "num_word_context": 202.3509, "num_word_doc": 49.7646, "num_word_query": 23.2777, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9505.9797, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6494, "query_norm": 1.8045, "queue_k_norm": 1.8571, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3607, "sent_len_1": 66.7083, "sent_len_max_0": 127.975, "sent_len_max_1": 208.4325, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 93900 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.453, "doc_norm": 1.8587, "encoder_q-embeddings": 4996.4541, "encoder_q-layer.0": 3376.5381, "encoder_q-layer.1": 3544.772, "encoder_q-layer.10": 6088.2246, "encoder_q-layer.11": 14443.4727, "encoder_q-layer.2": 3970.4243, "encoder_q-layer.3": 4059.7375, "encoder_q-layer.4": 4273.2554, "encoder_q-layer.5": 4350.4839, "encoder_q-layer.6": 4840.2988, "encoder_q-layer.7": 5365.687, "encoder_q-layer.8": 6412.5361, "encoder_q-layer.9": 5797.3071, "epoch": 0.41, "inbatch_neg_score": 1.6603, "inbatch_pos_score": 2.2617, "learning_rate": 2.7894736842105263e-05, "loss": 3.453, "norm_diff": 0.053, "norm_loss": 0.0, "num_token_doc": 66.699, "num_token_overlap": 11.6988, "num_token_query": 31.4051, "num_token_union": 65.0901, "num_word_context": 202.3549, "num_word_doc": 49.7991, "num_word_query": 23.3258, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9286.3795, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6602, "query_norm": 1.8057, "queue_k_norm": 1.8585, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4051, "sent_len_1": 66.699, "sent_len_max_0": 127.9712, "sent_len_max_1": 207.6987, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 94000 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.4857, "doc_norm": 1.8643, "encoder_q-embeddings": 4885.3286, "encoder_q-layer.0": 3410.0645, "encoder_q-layer.1": 3556.9753, "encoder_q-layer.10": 6791.3457, "encoder_q-layer.11": 14888.959, "encoder_q-layer.2": 3982.7192, "encoder_q-layer.3": 4206.4795, "encoder_q-layer.4": 4670.1924, "encoder_q-layer.5": 4542.6562, "encoder_q-layer.6": 5074.144, "encoder_q-layer.7": 5701.8594, "encoder_q-layer.8": 7008.4365, "encoder_q-layer.9": 6286.1133, "epoch": 0.41, "inbatch_neg_score": 1.6762, "inbatch_pos_score": 2.2539, "learning_rate": 2.786842105263158e-05, "loss": 3.4857, "norm_diff": 0.0537, "norm_loss": 0.0, "num_token_doc": 66.8783, "num_token_overlap": 11.6599, "num_token_query": 31.3385, "num_token_union": 65.1852, "num_word_context": 202.1811, "num_word_doc": 49.91, "num_word_query": 23.2866, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9726.7922, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6758, "query_norm": 1.8106, "queue_k_norm": 1.8604, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3385, "sent_len_1": 66.8783, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.3512, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 94100 }, { "accuracy": 48.1445, "active_queue_size": 16384.0, "cl_loss": 3.4539, "doc_norm": 1.8589, "encoder_q-embeddings": 5403.0078, "encoder_q-layer.0": 3746.3469, "encoder_q-layer.1": 4183.2573, "encoder_q-layer.10": 6385.6909, "encoder_q-layer.11": 14778.7061, "encoder_q-layer.2": 4610.5068, "encoder_q-layer.3": 4668.0161, "encoder_q-layer.4": 4722.6318, "encoder_q-layer.5": 4837.2676, "encoder_q-layer.6": 5633.563, "encoder_q-layer.7": 6117.8652, "encoder_q-layer.8": 6782.1372, "encoder_q-layer.9": 6177.3496, "epoch": 0.41, "inbatch_neg_score": 1.683, "inbatch_pos_score": 2.2676, "learning_rate": 2.7842105263157896e-05, "loss": 3.4539, "norm_diff": 0.0416, "norm_loss": 0.0, "num_token_doc": 66.6495, "num_token_overlap": 11.6704, "num_token_query": 31.427, "num_token_union": 65.0991, "num_word_context": 202.2508, "num_word_doc": 49.7585, "num_word_query": 23.3461, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9920.2799, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6826, "query_norm": 1.8173, "queue_k_norm": 1.8618, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.427, "sent_len_1": 66.6495, "sent_len_max_0": 127.99, "sent_len_max_1": 207.715, "stdk": 0.0489, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 94200 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4661, "doc_norm": 1.8655, "encoder_q-embeddings": 5045.3276, "encoder_q-layer.0": 3544.1357, "encoder_q-layer.1": 3686.9873, "encoder_q-layer.10": 6396.8779, "encoder_q-layer.11": 14844.7979, "encoder_q-layer.2": 4129.7383, "encoder_q-layer.3": 4296.1675, "encoder_q-layer.4": 4425.729, "encoder_q-layer.5": 4621.3813, "encoder_q-layer.6": 5102.9976, "encoder_q-layer.7": 5739.4443, "encoder_q-layer.8": 6595.9985, "encoder_q-layer.9": 6109.8535, "epoch": 0.41, "inbatch_neg_score": 1.6976, "inbatch_pos_score": 2.3008, "learning_rate": 2.781578947368421e-05, "loss": 3.4661, "norm_diff": 0.0417, "norm_loss": 0.0, "num_token_doc": 66.7868, "num_token_overlap": 11.6364, "num_token_query": 31.25, "num_token_union": 65.0731, "num_word_context": 202.262, "num_word_doc": 49.8227, "num_word_query": 23.2069, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9661.2871, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.6973, "query_norm": 1.8239, "queue_k_norm": 1.863, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.25, "sent_len_1": 66.7868, "sent_len_max_0": 127.9938, "sent_len_max_1": 211.2988, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 94300 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4609, "doc_norm": 1.8665, "encoder_q-embeddings": 5018.7515, "encoder_q-layer.0": 3447.78, "encoder_q-layer.1": 3687.1658, "encoder_q-layer.10": 6560.8447, "encoder_q-layer.11": 14840.1299, "encoder_q-layer.2": 4054.4646, "encoder_q-layer.3": 4173.3428, "encoder_q-layer.4": 4637.3115, "encoder_q-layer.5": 4754.4038, "encoder_q-layer.6": 5115.0459, "encoder_q-layer.7": 5882.3521, "encoder_q-layer.8": 6790.2681, "encoder_q-layer.9": 6218.5684, "epoch": 0.41, "inbatch_neg_score": 1.7084, "inbatch_pos_score": 2.3203, "learning_rate": 2.7789473684210526e-05, "loss": 3.4609, "norm_diff": 0.0458, "norm_loss": 0.0, "num_token_doc": 66.7517, "num_token_overlap": 11.6914, "num_token_query": 31.3709, "num_token_union": 65.0904, "num_word_context": 202.2954, "num_word_doc": 49.8137, "num_word_query": 23.2974, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9628.4674, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.708, "query_norm": 1.8207, "queue_k_norm": 1.8673, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3709, "sent_len_1": 66.7517, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.365, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94400 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.468, "doc_norm": 1.8689, "encoder_q-embeddings": 4924.6421, "encoder_q-layer.0": 3362.2185, "encoder_q-layer.1": 3661.3638, "encoder_q-layer.10": 6191.0215, "encoder_q-layer.11": 15251.4229, "encoder_q-layer.2": 4171.208, "encoder_q-layer.3": 4224.3555, "encoder_q-layer.4": 4443.9536, "encoder_q-layer.5": 4422.8037, "encoder_q-layer.6": 5021.1323, "encoder_q-layer.7": 5541.0063, "encoder_q-layer.8": 6902.1279, "encoder_q-layer.9": 6206.3096, "epoch": 0.41, "inbatch_neg_score": 1.7134, "inbatch_pos_score": 2.3066, "learning_rate": 2.776315789473684e-05, "loss": 3.468, "norm_diff": 0.0573, "norm_loss": 0.0, "num_token_doc": 66.6069, "num_token_overlap": 11.6387, "num_token_query": 31.3076, "num_token_union": 64.9966, "num_word_context": 202.3556, "num_word_doc": 49.7098, "num_word_query": 23.2491, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9645.4778, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7129, "query_norm": 1.8116, "queue_k_norm": 1.8693, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3076, "sent_len_1": 66.6069, "sent_len_max_0": 127.985, "sent_len_max_1": 207.5062, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94500 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4732, "doc_norm": 1.8687, "encoder_q-embeddings": 9908.29, "encoder_q-layer.0": 7262.3154, "encoder_q-layer.1": 8417.0225, "encoder_q-layer.10": 6199.9536, "encoder_q-layer.11": 14905.9873, "encoder_q-layer.2": 9595.167, "encoder_q-layer.3": 9604.3916, "encoder_q-layer.4": 10117.9453, "encoder_q-layer.5": 11199.9277, "encoder_q-layer.6": 11482.8193, "encoder_q-layer.7": 13160.2549, "encoder_q-layer.8": 8657.7344, "encoder_q-layer.9": 6273.6475, "epoch": 0.41, "inbatch_neg_score": 1.716, "inbatch_pos_score": 2.3066, "learning_rate": 2.773684210526316e-05, "loss": 3.4732, "norm_diff": 0.0709, "norm_loss": 0.0, "num_token_doc": 66.7885, "num_token_overlap": 11.6548, "num_token_query": 31.2962, "num_token_union": 65.1141, "num_word_context": 202.0119, "num_word_doc": 49.8323, "num_word_query": 23.2371, "postclip_grad_norm": 1.0, "preclip_grad_norm": 15283.2156, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7158, "query_norm": 1.7977, "queue_k_norm": 1.8703, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2962, "sent_len_1": 66.7885, "sent_len_max_0": 127.9688, "sent_len_max_1": 211.4638, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94600 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.4591, "doc_norm": 1.8695, "encoder_q-embeddings": 5221.1313, "encoder_q-layer.0": 3592.3835, "encoder_q-layer.1": 3819.6333, "encoder_q-layer.10": 6826.2061, "encoder_q-layer.11": 15670.708, "encoder_q-layer.2": 4284.7324, "encoder_q-layer.3": 4411.9883, "encoder_q-layer.4": 4805.4316, "encoder_q-layer.5": 4691.6753, "encoder_q-layer.6": 5187.645, "encoder_q-layer.7": 6056.498, "encoder_q-layer.8": 6867.8467, "encoder_q-layer.9": 6645.751, "epoch": 0.41, "inbatch_neg_score": 1.7239, "inbatch_pos_score": 2.3281, "learning_rate": 2.7710526315789474e-05, "loss": 3.4591, "norm_diff": 0.0587, "norm_loss": 0.0, "num_token_doc": 66.877, "num_token_overlap": 11.6729, "num_token_query": 31.3438, "num_token_union": 65.1647, "num_word_context": 202.1634, "num_word_doc": 49.8936, "num_word_query": 23.2776, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10133.1391, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7236, "query_norm": 1.8108, "queue_k_norm": 1.873, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3438, "sent_len_1": 66.877, "sent_len_max_0": 127.9887, "sent_len_max_1": 211.9187, "stdk": 0.049, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94700 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4637, "doc_norm": 1.8744, "encoder_q-embeddings": 4749.064, "encoder_q-layer.0": 3267.1006, "encoder_q-layer.1": 3410.7405, "encoder_q-layer.10": 6323.5059, "encoder_q-layer.11": 15017.3174, "encoder_q-layer.2": 3788.6497, "encoder_q-layer.3": 3887.3628, "encoder_q-layer.4": 4164.9375, "encoder_q-layer.5": 4222.5771, "encoder_q-layer.6": 4896.2744, "encoder_q-layer.7": 5390.1235, "encoder_q-layer.8": 6663.4116, "encoder_q-layer.9": 6336.7271, "epoch": 0.41, "inbatch_neg_score": 1.7258, "inbatch_pos_score": 2.3301, "learning_rate": 2.768421052631579e-05, "loss": 3.4637, "norm_diff": 0.0675, "norm_loss": 0.0, "num_token_doc": 66.7775, "num_token_overlap": 11.6836, "num_token_query": 31.3651, "num_token_union": 65.1048, "num_word_context": 202.357, "num_word_doc": 49.8358, "num_word_query": 23.2994, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9487.5997, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7256, "query_norm": 1.8069, "queue_k_norm": 1.8742, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3651, "sent_len_1": 66.7775, "sent_len_max_0": 127.995, "sent_len_max_1": 207.5125, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94800 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.4619, "doc_norm": 1.8763, "encoder_q-embeddings": 5006.1885, "encoder_q-layer.0": 3339.0222, "encoder_q-layer.1": 3624.4268, "encoder_q-layer.10": 6082.4814, "encoder_q-layer.11": 14438.4414, "encoder_q-layer.2": 4013.8545, "encoder_q-layer.3": 4040.8179, "encoder_q-layer.4": 4313.4985, "encoder_q-layer.5": 4488.1328, "encoder_q-layer.6": 4814.625, "encoder_q-layer.7": 5426.0063, "encoder_q-layer.8": 6653.2661, "encoder_q-layer.9": 6065.6899, "epoch": 0.41, "inbatch_neg_score": 1.7263, "inbatch_pos_score": 2.3516, "learning_rate": 2.7657894736842104e-05, "loss": 3.4619, "norm_diff": 0.0625, "norm_loss": 0.0, "num_token_doc": 66.779, "num_token_overlap": 11.6456, "num_token_query": 31.2733, "num_token_union": 65.0559, "num_word_context": 202.0011, "num_word_doc": 49.8048, "num_word_query": 23.22, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9396.5235, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7256, "query_norm": 1.8138, "queue_k_norm": 1.8765, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2733, "sent_len_1": 66.779, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.4112, "stdk": 0.0491, "stdq": 0.0439, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 94900 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4657, "doc_norm": 1.8801, "encoder_q-embeddings": 4819.624, "encoder_q-layer.0": 3309.1956, "encoder_q-layer.1": 3468.3364, "encoder_q-layer.10": 6341.0923, "encoder_q-layer.11": 14884.9375, "encoder_q-layer.2": 3984.1599, "encoder_q-layer.3": 3963.6387, "encoder_q-layer.4": 4069.8914, "encoder_q-layer.5": 4239.4346, "encoder_q-layer.6": 4794.8242, "encoder_q-layer.7": 5463.3691, "encoder_q-layer.8": 6473.978, "encoder_q-layer.9": 5995.7998, "epoch": 0.41, "inbatch_neg_score": 1.726, "inbatch_pos_score": 2.334, "learning_rate": 2.7631578947368426e-05, "loss": 3.4657, "norm_diff": 0.0738, "norm_loss": 0.0, "num_token_doc": 66.8186, "num_token_overlap": 11.6522, "num_token_query": 31.2552, "num_token_union": 65.1151, "num_word_context": 202.1604, "num_word_doc": 49.8419, "num_word_query": 23.2137, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9368.9569, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.7266, "query_norm": 1.8063, "queue_k_norm": 1.8784, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2552, "sent_len_1": 66.8186, "sent_len_max_0": 127.985, "sent_len_max_1": 209.5375, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95000 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.4676, "doc_norm": 1.8761, "encoder_q-embeddings": 10291.2168, "encoder_q-layer.0": 6883.6465, "encoder_q-layer.1": 7200.6562, "encoder_q-layer.10": 12251.9453, "encoder_q-layer.11": 30036.2734, "encoder_q-layer.2": 7965.2056, "encoder_q-layer.3": 8459.1387, "encoder_q-layer.4": 8870.9473, "encoder_q-layer.5": 8636.8242, "encoder_q-layer.6": 9849.6211, "encoder_q-layer.7": 11684.3262, "encoder_q-layer.8": 13625.1348, "encoder_q-layer.9": 12128.3506, "epoch": 0.41, "inbatch_neg_score": 1.7321, "inbatch_pos_score": 2.332, "learning_rate": 2.760526315789474e-05, "loss": 3.4676, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.8228, "num_token_overlap": 11.6635, "num_token_query": 31.3852, "num_token_union": 65.1528, "num_word_context": 202.2132, "num_word_doc": 49.8551, "num_word_query": 23.2968, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19418.2694, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7324, "query_norm": 1.7988, "queue_k_norm": 1.8787, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3852, "sent_len_1": 66.8228, "sent_len_max_0": 127.985, "sent_len_max_1": 207.4563, "stdk": 0.049, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 95100 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4669, "doc_norm": 1.8805, "encoder_q-embeddings": 11639.9111, "encoder_q-layer.0": 8121.8574, "encoder_q-layer.1": 9074.2188, "encoder_q-layer.10": 13088.2646, "encoder_q-layer.11": 29643.2637, "encoder_q-layer.2": 10170.8428, "encoder_q-layer.3": 10375.9844, "encoder_q-layer.4": 11030.3809, "encoder_q-layer.5": 11048.9414, "encoder_q-layer.6": 11731.5723, "encoder_q-layer.7": 13050.0801, "encoder_q-layer.8": 13870.3467, "encoder_q-layer.9": 12426.8018, "epoch": 0.41, "inbatch_neg_score": 1.7335, "inbatch_pos_score": 2.334, "learning_rate": 2.7578947368421053e-05, "loss": 3.4669, "norm_diff": 0.0743, "norm_loss": 0.0, "num_token_doc": 66.7944, "num_token_overlap": 11.6823, "num_token_query": 31.3672, "num_token_union": 65.1039, "num_word_context": 202.3333, "num_word_doc": 49.8386, "num_word_query": 23.2996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20419.7245, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7324, "query_norm": 1.8062, "queue_k_norm": 1.8802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3672, "sent_len_1": 66.7944, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.585, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 95200 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.4641, "doc_norm": 1.8836, "encoder_q-embeddings": 10637.4707, "encoder_q-layer.0": 6978.1812, "encoder_q-layer.1": 7375.71, "encoder_q-layer.10": 12911.501, "encoder_q-layer.11": 31731.3867, "encoder_q-layer.2": 8353.5615, "encoder_q-layer.3": 8620.3818, "encoder_q-layer.4": 9150.1182, "encoder_q-layer.5": 9353.2451, "encoder_q-layer.6": 10122.709, "encoder_q-layer.7": 11651.3096, "encoder_q-layer.8": 13510.8945, "encoder_q-layer.9": 12401.3027, "epoch": 0.41, "inbatch_neg_score": 1.7313, "inbatch_pos_score": 2.3379, "learning_rate": 2.7552631578947368e-05, "loss": 3.4641, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.8452, "num_token_overlap": 11.6549, "num_token_query": 31.2764, "num_token_union": 65.1327, "num_word_context": 202.2948, "num_word_doc": 49.856, "num_word_query": 23.2211, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19603.5011, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7314, "query_norm": 1.8056, "queue_k_norm": 1.8825, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2764, "sent_len_1": 66.8452, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.155, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95300 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.456, "doc_norm": 1.8821, "encoder_q-embeddings": 10172.5479, "encoder_q-layer.0": 7007.5557, "encoder_q-layer.1": 7422.7871, "encoder_q-layer.10": 12062.3594, "encoder_q-layer.11": 28298.0371, "encoder_q-layer.2": 8349.875, "encoder_q-layer.3": 8436.9971, "encoder_q-layer.4": 8875.4277, "encoder_q-layer.5": 8978.1211, "encoder_q-layer.6": 9854.7139, "encoder_q-layer.7": 11286.3418, "encoder_q-layer.8": 13099.334, "encoder_q-layer.9": 11819.8271, "epoch": 0.41, "inbatch_neg_score": 1.7363, "inbatch_pos_score": 2.334, "learning_rate": 2.7526315789473683e-05, "loss": 3.456, "norm_diff": 0.0791, "norm_loss": 0.0, "num_token_doc": 66.8431, "num_token_overlap": 11.678, "num_token_query": 31.3674, "num_token_union": 65.1686, "num_word_context": 202.3704, "num_word_doc": 49.8786, "num_word_query": 23.2961, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18724.5484, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7363, "query_norm": 1.803, "queue_k_norm": 1.8847, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3674, "sent_len_1": 66.8431, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.4712, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95400 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.462, "doc_norm": 1.884, "encoder_q-embeddings": 10678.248, "encoder_q-layer.0": 7279.7632, "encoder_q-layer.1": 7589.1377, "encoder_q-layer.10": 12701.9521, "encoder_q-layer.11": 31791.3418, "encoder_q-layer.2": 8354.6748, "encoder_q-layer.3": 8482.8652, "encoder_q-layer.4": 9158.499, "encoder_q-layer.5": 9442.2402, "encoder_q-layer.6": 10416.7256, "encoder_q-layer.7": 11869.7236, "encoder_q-layer.8": 13667.5488, "encoder_q-layer.9": 12620.7607, "epoch": 0.41, "inbatch_neg_score": 1.7401, "inbatch_pos_score": 2.3457, "learning_rate": 2.7500000000000004e-05, "loss": 3.462, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.7185, "num_token_overlap": 11.681, "num_token_query": 31.3049, "num_token_union": 65.0159, "num_word_context": 202.0837, "num_word_doc": 49.7689, "num_word_query": 23.2365, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19819.7534, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7402, "query_norm": 1.8125, "queue_k_norm": 1.8855, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3049, "sent_len_1": 66.7185, "sent_len_max_0": 127.995, "sent_len_max_1": 210.33, "stdk": 0.0491, "stdq": 0.0437, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95500 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.464, "doc_norm": 1.8872, "encoder_q-embeddings": 11000.0713, "encoder_q-layer.0": 7700.9448, "encoder_q-layer.1": 8067.79, "encoder_q-layer.10": 14010.6328, "encoder_q-layer.11": 30466.7852, "encoder_q-layer.2": 9146.5537, "encoder_q-layer.3": 9496.1758, "encoder_q-layer.4": 9900.6953, "encoder_q-layer.5": 9619.2178, "encoder_q-layer.6": 10871.707, "encoder_q-layer.7": 12027.5576, "encoder_q-layer.8": 13526.2744, "encoder_q-layer.9": 12179.5498, "epoch": 0.41, "inbatch_neg_score": 1.742, "inbatch_pos_score": 2.3496, "learning_rate": 2.747368421052632e-05, "loss": 3.464, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.868, "num_token_overlap": 11.704, "num_token_query": 31.4089, "num_token_union": 65.1744, "num_word_context": 202.4257, "num_word_doc": 49.8756, "num_word_query": 23.3238, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19999.1958, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7422, "query_norm": 1.8149, "queue_k_norm": 1.8855, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4089, "sent_len_1": 66.868, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.3725, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95600 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.4553, "doc_norm": 1.8836, "encoder_q-embeddings": 10766.75, "encoder_q-layer.0": 7531.9629, "encoder_q-layer.1": 8166.7163, "encoder_q-layer.10": 12752.4336, "encoder_q-layer.11": 31342.002, "encoder_q-layer.2": 9165.8184, "encoder_q-layer.3": 9668.7754, "encoder_q-layer.4": 10364.6797, "encoder_q-layer.5": 10637.7666, "encoder_q-layer.6": 11378.6973, "encoder_q-layer.7": 11864.5928, "encoder_q-layer.8": 13419.958, "encoder_q-layer.9": 12057.208, "epoch": 0.42, "inbatch_neg_score": 1.7445, "inbatch_pos_score": 2.3574, "learning_rate": 2.7447368421052634e-05, "loss": 3.4553, "norm_diff": 0.0834, "norm_loss": 0.0, "num_token_doc": 66.7895, "num_token_overlap": 11.6783, "num_token_query": 31.3481, "num_token_union": 65.1087, "num_word_context": 202.1509, "num_word_doc": 49.8315, "num_word_query": 23.2785, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20189.3425, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7451, "query_norm": 1.8002, "queue_k_norm": 1.886, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3481, "sent_len_1": 66.7895, "sent_len_max_0": 127.9525, "sent_len_max_1": 209.8963, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 95700 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.4644, "doc_norm": 1.8904, "encoder_q-embeddings": 10696.9795, "encoder_q-layer.0": 7348.0024, "encoder_q-layer.1": 8139.584, "encoder_q-layer.10": 12407.7334, "encoder_q-layer.11": 31241.5137, "encoder_q-layer.2": 9081.2852, "encoder_q-layer.3": 9485.0605, "encoder_q-layer.4": 9652.5166, "encoder_q-layer.5": 9987.2617, "encoder_q-layer.6": 10798.7207, "encoder_q-layer.7": 12021.0635, "encoder_q-layer.8": 13394.3955, "encoder_q-layer.9": 12335.7891, "epoch": 0.42, "inbatch_neg_score": 1.7518, "inbatch_pos_score": 2.3496, "learning_rate": 2.7421052631578946e-05, "loss": 3.4644, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.6842, "num_token_overlap": 11.6406, "num_token_query": 31.3454, "num_token_union": 65.0649, "num_word_context": 201.9831, "num_word_doc": 49.7154, "num_word_query": 23.2647, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20098.732, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.752, "query_norm": 1.8106, "queue_k_norm": 1.8876, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3454, "sent_len_1": 66.6842, "sent_len_max_0": 127.9912, "sent_len_max_1": 212.205, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95800 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.4473, "doc_norm": 1.8894, "encoder_q-embeddings": 11629.9639, "encoder_q-layer.0": 7977.3003, "encoder_q-layer.1": 8756.3418, "encoder_q-layer.10": 12242.374, "encoder_q-layer.11": 29038.5039, "encoder_q-layer.2": 9123.3838, "encoder_q-layer.3": 9142.2422, "encoder_q-layer.4": 9328.041, "encoder_q-layer.5": 9112.8389, "encoder_q-layer.6": 9819.6846, "encoder_q-layer.7": 11076.1621, "encoder_q-layer.8": 12448.543, "encoder_q-layer.9": 11490.3359, "epoch": 0.42, "inbatch_neg_score": 1.7551, "inbatch_pos_score": 2.3691, "learning_rate": 2.739473684210526e-05, "loss": 3.4473, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.7781, "num_token_overlap": 11.684, "num_token_query": 31.4097, "num_token_union": 65.1394, "num_word_context": 202.0125, "num_word_doc": 49.8258, "num_word_query": 23.3451, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19317.6725, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7549, "query_norm": 1.8136, "queue_k_norm": 1.8894, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4097, "sent_len_1": 66.7781, "sent_len_max_0": 127.9488, "sent_len_max_1": 210.9338, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 95900 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.4582, "doc_norm": 1.8878, "encoder_q-embeddings": 12263.3369, "encoder_q-layer.0": 8312.7646, "encoder_q-layer.1": 9249.1777, "encoder_q-layer.10": 13474.6123, "encoder_q-layer.11": 31856.834, "encoder_q-layer.2": 10388.6494, "encoder_q-layer.3": 11236.3389, "encoder_q-layer.4": 11957.8311, "encoder_q-layer.5": 12311.6729, "encoder_q-layer.6": 12920.3828, "encoder_q-layer.7": 13690.6836, "encoder_q-layer.8": 14460.5996, "encoder_q-layer.9": 13038.0527, "epoch": 0.42, "inbatch_neg_score": 1.7614, "inbatch_pos_score": 2.3789, "learning_rate": 2.7368421052631583e-05, "loss": 3.4582, "norm_diff": 0.0561, "norm_loss": 0.0, "num_token_doc": 66.7071, "num_token_overlap": 11.6697, "num_token_query": 31.3449, "num_token_union": 65.0542, "num_word_context": 202.3259, "num_word_doc": 49.7633, "num_word_query": 23.2753, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22006.9205, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7607, "query_norm": 1.8316, "queue_k_norm": 1.8917, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3449, "sent_len_1": 66.7071, "sent_len_max_0": 127.995, "sent_len_max_1": 210.63, "stdk": 0.049, "stdq": 0.0442, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96000 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4683, "doc_norm": 1.8979, "encoder_q-embeddings": 9506.0361, "encoder_q-layer.0": 6680.9697, "encoder_q-layer.1": 7083.0522, "encoder_q-layer.10": 13195.3203, "encoder_q-layer.11": 30504.1191, "encoder_q-layer.2": 7627.3081, "encoder_q-layer.3": 7936.9521, "encoder_q-layer.4": 8245.3447, "encoder_q-layer.5": 8520.9102, "encoder_q-layer.6": 9621.0947, "encoder_q-layer.7": 11240.1309, "encoder_q-layer.8": 12911.8652, "encoder_q-layer.9": 12948.7627, "epoch": 0.42, "inbatch_neg_score": 1.763, "inbatch_pos_score": 2.3594, "learning_rate": 2.7342105263157898e-05, "loss": 3.4683, "norm_diff": 0.0817, "norm_loss": 0.0, "num_token_doc": 66.8143, "num_token_overlap": 11.6536, "num_token_query": 31.2992, "num_token_union": 65.1147, "num_word_context": 202.1441, "num_word_doc": 49.8462, "num_word_query": 23.2293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19038.166, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7627, "query_norm": 1.8162, "queue_k_norm": 1.8923, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2992, "sent_len_1": 66.8143, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.8025, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96100 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.4476, "doc_norm": 1.8936, "encoder_q-embeddings": 9714.8271, "encoder_q-layer.0": 6699.0142, "encoder_q-layer.1": 6983.0229, "encoder_q-layer.10": 12002.2559, "encoder_q-layer.11": 28278.5879, "encoder_q-layer.2": 7691.2334, "encoder_q-layer.3": 8076.4727, "encoder_q-layer.4": 8223.5391, "encoder_q-layer.5": 8581.1279, "encoder_q-layer.6": 9566.2695, "encoder_q-layer.7": 10929.3877, "encoder_q-layer.8": 12828.0566, "encoder_q-layer.9": 11945.2949, "epoch": 0.42, "inbatch_neg_score": 1.7699, "inbatch_pos_score": 2.3828, "learning_rate": 2.7315789473684213e-05, "loss": 3.4476, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7017, "num_token_overlap": 11.7099, "num_token_query": 31.4527, "num_token_union": 65.0998, "num_word_context": 202.3252, "num_word_doc": 49.76, "num_word_query": 23.3592, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18417.7466, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7695, "query_norm": 1.8241, "queue_k_norm": 1.8941, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4527, "sent_len_1": 66.7017, "sent_len_max_0": 127.99, "sent_len_max_1": 210.1238, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96200 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4576, "doc_norm": 1.8942, "encoder_q-embeddings": 11281.9492, "encoder_q-layer.0": 7854.9604, "encoder_q-layer.1": 8423.9922, "encoder_q-layer.10": 12236.6377, "encoder_q-layer.11": 29214.4551, "encoder_q-layer.2": 9507.8184, "encoder_q-layer.3": 9742.3857, "encoder_q-layer.4": 10022.4707, "encoder_q-layer.5": 10150.416, "encoder_q-layer.6": 10531.4258, "encoder_q-layer.7": 11012.8555, "encoder_q-layer.8": 13019.2129, "encoder_q-layer.9": 12274.7607, "epoch": 0.42, "inbatch_neg_score": 1.7708, "inbatch_pos_score": 2.3887, "learning_rate": 2.7289473684210528e-05, "loss": 3.4576, "norm_diff": 0.0625, "norm_loss": 0.0, "num_token_doc": 66.741, "num_token_overlap": 11.6875, "num_token_query": 31.464, "num_token_union": 65.1734, "num_word_context": 202.2096, "num_word_doc": 49.8198, "num_word_query": 23.3669, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19649.4346, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7715, "query_norm": 1.8317, "queue_k_norm": 1.896, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.464, "sent_len_1": 66.741, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.1813, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 96300 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.4629, "doc_norm": 1.8975, "encoder_q-embeddings": 9549.1211, "encoder_q-layer.0": 6586.1973, "encoder_q-layer.1": 7035.918, "encoder_q-layer.10": 13379.9463, "encoder_q-layer.11": 29818.2422, "encoder_q-layer.2": 7735.3379, "encoder_q-layer.3": 8198.5732, "encoder_q-layer.4": 8732.4053, "encoder_q-layer.5": 8854.1387, "encoder_q-layer.6": 9848.4092, "encoder_q-layer.7": 11190.3398, "encoder_q-layer.8": 13149.8896, "encoder_q-layer.9": 12126.9307, "epoch": 0.42, "inbatch_neg_score": 1.7785, "inbatch_pos_score": 2.3887, "learning_rate": 2.7263157894736846e-05, "loss": 3.4629, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.7323, "num_token_overlap": 11.6898, "num_token_query": 31.3432, "num_token_union": 65.0745, "num_word_context": 202.0056, "num_word_doc": 49.8132, "num_word_query": 23.2692, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18845.8531, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7783, "query_norm": 1.8262, "queue_k_norm": 1.8965, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3432, "sent_len_1": 66.7323, "sent_len_max_0": 127.99, "sent_len_max_1": 208.2225, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96400 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.4563, "doc_norm": 1.8933, "encoder_q-embeddings": 9237.9395, "encoder_q-layer.0": 6564.4912, "encoder_q-layer.1": 6783.5347, "encoder_q-layer.10": 12389.7695, "encoder_q-layer.11": 28084.1172, "encoder_q-layer.2": 7627.1924, "encoder_q-layer.3": 7651.041, "encoder_q-layer.4": 8083.5146, "encoder_q-layer.5": 8200.0146, "encoder_q-layer.6": 9190.7363, "encoder_q-layer.7": 10738.1768, "encoder_q-layer.8": 12634.8574, "encoder_q-layer.9": 11817.7051, "epoch": 0.42, "inbatch_neg_score": 1.7892, "inbatch_pos_score": 2.3984, "learning_rate": 2.723684210526316e-05, "loss": 3.4563, "norm_diff": 0.0533, "norm_loss": 0.0, "num_token_doc": 66.8027, "num_token_overlap": 11.6957, "num_token_query": 31.4239, "num_token_union": 65.129, "num_word_context": 202.3844, "num_word_doc": 49.8136, "num_word_query": 23.32, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18067.9971, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.791, "query_norm": 1.84, "queue_k_norm": 1.8981, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4239, "sent_len_1": 66.8027, "sent_len_max_0": 127.9638, "sent_len_max_1": 207.2713, "stdk": 0.049, "stdq": 0.0436, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 96500 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4589, "doc_norm": 1.8987, "encoder_q-embeddings": 9805.5576, "encoder_q-layer.0": 6635.5757, "encoder_q-layer.1": 7188.3208, "encoder_q-layer.10": 12755.9902, "encoder_q-layer.11": 29591.1367, "encoder_q-layer.2": 7922.7324, "encoder_q-layer.3": 8076.7534, "encoder_q-layer.4": 8638.4541, "encoder_q-layer.5": 8807.6621, "encoder_q-layer.6": 10151.665, "encoder_q-layer.7": 11108.0098, "encoder_q-layer.8": 13475.8203, "encoder_q-layer.9": 12156.6484, "epoch": 0.42, "inbatch_neg_score": 1.7958, "inbatch_pos_score": 2.4004, "learning_rate": 2.7210526315789476e-05, "loss": 3.4589, "norm_diff": 0.0601, "norm_loss": 0.0, "num_token_doc": 66.8456, "num_token_overlap": 11.6533, "num_token_query": 31.1917, "num_token_union": 65.0498, "num_word_context": 202.1857, "num_word_doc": 49.8817, "num_word_query": 23.1532, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19063.1616, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.7959, "query_norm": 1.8385, "queue_k_norm": 1.8989, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.1917, "sent_len_1": 66.8456, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.905, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96600 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.463, "doc_norm": 1.8998, "encoder_q-embeddings": 22062.3906, "encoder_q-layer.0": 15166.6826, "encoder_q-layer.1": 18423.748, "encoder_q-layer.10": 13420.8105, "encoder_q-layer.11": 31006.5977, "encoder_q-layer.2": 21845.6484, "encoder_q-layer.3": 22262.0078, "encoder_q-layer.4": 21787.2949, "encoder_q-layer.5": 24542.6094, "encoder_q-layer.6": 20952.1992, "encoder_q-layer.7": 19780.0742, "encoder_q-layer.8": 15542.8438, "encoder_q-layer.9": 12706.373, "epoch": 0.42, "inbatch_neg_score": 1.807, "inbatch_pos_score": 2.4043, "learning_rate": 2.718421052631579e-05, "loss": 3.463, "norm_diff": 0.0555, "norm_loss": 0.0, "num_token_doc": 66.7427, "num_token_overlap": 11.6384, "num_token_query": 31.2762, "num_token_union": 65.0791, "num_word_context": 202.3777, "num_word_doc": 49.8141, "num_word_query": 23.2366, "postclip_grad_norm": 1.0, "preclip_grad_norm": 30814.9396, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 1.8076, "query_norm": 1.8444, "queue_k_norm": 1.8996, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2762, "sent_len_1": 66.7427, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.7988, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 96700 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4585, "doc_norm": 1.8979, "encoder_q-embeddings": 11760.0273, "encoder_q-layer.0": 8729.8242, "encoder_q-layer.1": 9782.7529, "encoder_q-layer.10": 12206.5215, "encoder_q-layer.11": 28408.6992, "encoder_q-layer.2": 11221.1387, "encoder_q-layer.3": 11892.6152, "encoder_q-layer.4": 12594.4541, "encoder_q-layer.5": 12562.2432, "encoder_q-layer.6": 12086.0498, "encoder_q-layer.7": 12623.8281, "encoder_q-layer.8": 13421.7588, "encoder_q-layer.9": 12239.0986, "epoch": 0.42, "inbatch_neg_score": 1.8206, "inbatch_pos_score": 2.418, "learning_rate": 2.7157894736842106e-05, "loss": 3.4585, "norm_diff": 0.0443, "norm_loss": 0.0, "num_token_doc": 66.7485, "num_token_overlap": 11.6775, "num_token_query": 31.4172, "num_token_union": 65.1086, "num_word_context": 202.1448, "num_word_doc": 49.7946, "num_word_query": 23.3233, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20486.3383, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.8203, "query_norm": 1.8536, "queue_k_norm": 1.8995, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4172, "sent_len_1": 66.7485, "sent_len_max_0": 127.9838, "sent_len_max_1": 211.2862, "stdk": 0.049, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 96800 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4646, "doc_norm": 1.9014, "encoder_q-embeddings": 10718.6631, "encoder_q-layer.0": 7705.1543, "encoder_q-layer.1": 8481.3682, "encoder_q-layer.10": 13394.5547, "encoder_q-layer.11": 30782.5273, "encoder_q-layer.2": 9460.4336, "encoder_q-layer.3": 9653.0957, "encoder_q-layer.4": 10198.4287, "encoder_q-layer.5": 10128.6738, "encoder_q-layer.6": 10916.3115, "encoder_q-layer.7": 12218.8008, "encoder_q-layer.8": 13695.3525, "encoder_q-layer.9": 12698.9902, "epoch": 0.42, "inbatch_neg_score": 1.8312, "inbatch_pos_score": 2.4082, "learning_rate": 2.7131578947368424e-05, "loss": 3.4646, "norm_diff": 0.0512, "norm_loss": 0.0, "num_token_doc": 66.7243, "num_token_overlap": 11.6429, "num_token_query": 31.2889, "num_token_union": 65.0838, "num_word_context": 202.327, "num_word_doc": 49.791, "num_word_query": 23.2301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20181.4219, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.8301, "query_norm": 1.8502, "queue_k_norm": 1.9005, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2889, "sent_len_1": 66.7243, "sent_len_max_0": 127.9437, "sent_len_max_1": 210.9737, "stdk": 0.0491, "stdq": 0.0426, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 96900 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.4657, "doc_norm": 1.9037, "encoder_q-embeddings": 10106.7451, "encoder_q-layer.0": 7029.6821, "encoder_q-layer.1": 7093.4131, "encoder_q-layer.10": 13127.2598, "encoder_q-layer.11": 31074.6016, "encoder_q-layer.2": 7706.2842, "encoder_q-layer.3": 7848.4941, "encoder_q-layer.4": 8361.1191, "encoder_q-layer.5": 8527.6094, "encoder_q-layer.6": 9706.1143, "encoder_q-layer.7": 11182.4375, "encoder_q-layer.8": 13445.9785, "encoder_q-layer.9": 12369.3037, "epoch": 0.42, "inbatch_neg_score": 1.8469, "inbatch_pos_score": 2.4414, "learning_rate": 2.710526315789474e-05, "loss": 3.4657, "norm_diff": 0.0394, "norm_loss": 0.0, "num_token_doc": 66.762, "num_token_overlap": 11.673, "num_token_query": 31.382, "num_token_union": 65.1189, "num_word_context": 202.1257, "num_word_doc": 49.8176, "num_word_query": 23.3185, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19559.2028, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.8467, "query_norm": 1.8643, "queue_k_norm": 1.9039, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.382, "sent_len_1": 66.762, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.2663, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97000 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4747, "doc_norm": 1.9115, "encoder_q-embeddings": 20136.7656, "encoder_q-layer.0": 13796.7373, "encoder_q-layer.1": 14426.0361, "encoder_q-layer.10": 24993.3457, "encoder_q-layer.11": 57678.2812, "encoder_q-layer.2": 15737.2129, "encoder_q-layer.3": 16186.5859, "encoder_q-layer.4": 17125.8203, "encoder_q-layer.5": 17702.4863, "encoder_q-layer.6": 21020.1992, "encoder_q-layer.7": 22129.8281, "encoder_q-layer.8": 26638.3457, "encoder_q-layer.9": 23547.293, "epoch": 0.42, "inbatch_neg_score": 1.8576, "inbatch_pos_score": 2.4688, "learning_rate": 2.7078947368421054e-05, "loss": 3.4747, "norm_diff": 0.0355, "norm_loss": 0.0, "num_token_doc": 66.7172, "num_token_overlap": 11.6494, "num_token_query": 31.3002, "num_token_union": 65.0719, "num_word_context": 202.0459, "num_word_doc": 49.7983, "num_word_query": 23.2494, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37422.2044, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 1.8574, "query_norm": 1.876, "queue_k_norm": 1.9057, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3002, "sent_len_1": 66.7172, "sent_len_max_0": 127.9712, "sent_len_max_1": 206.6012, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97100 }, { "accuracy": 48.877, "active_queue_size": 16384.0, "cl_loss": 3.4632, "doc_norm": 1.9093, "encoder_q-embeddings": 20230.7051, "encoder_q-layer.0": 13859.1719, "encoder_q-layer.1": 14969.916, "encoder_q-layer.10": 26766.7324, "encoder_q-layer.11": 61235.9531, "encoder_q-layer.2": 16638.4434, "encoder_q-layer.3": 16696.4062, "encoder_q-layer.4": 17832.5918, "encoder_q-layer.5": 18112.1055, "encoder_q-layer.6": 20520.5391, "encoder_q-layer.7": 24623.2969, "encoder_q-layer.8": 29683.502, "encoder_q-layer.9": 26603.6621, "epoch": 0.42, "inbatch_neg_score": 1.8727, "inbatch_pos_score": 2.4648, "learning_rate": 2.705263157894737e-05, "loss": 3.4632, "norm_diff": 0.0331, "norm_loss": 0.0, "num_token_doc": 66.7422, "num_token_overlap": 11.669, "num_token_query": 31.3515, "num_token_union": 65.0673, "num_word_context": 202.2017, "num_word_doc": 49.7982, "num_word_query": 23.2895, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39814.8251, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.8721, "query_norm": 1.8761, "queue_k_norm": 1.9082, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3515, "sent_len_1": 66.7422, "sent_len_max_0": 127.99, "sent_len_max_1": 209.8088, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97200 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4591, "doc_norm": 1.907, "encoder_q-embeddings": 18606.0215, "encoder_q-layer.0": 13187.9258, "encoder_q-layer.1": 13779.9541, "encoder_q-layer.10": 23954.0547, "encoder_q-layer.11": 61559.6055, "encoder_q-layer.2": 15311.8252, "encoder_q-layer.3": 15294.6191, "encoder_q-layer.4": 16646.2266, "encoder_q-layer.5": 16486.7637, "encoder_q-layer.6": 19109.457, "encoder_q-layer.7": 20666.1367, "encoder_q-layer.8": 25507.6113, "encoder_q-layer.9": 23147.4082, "epoch": 0.42, "inbatch_neg_score": 1.888, "inbatch_pos_score": 2.4707, "learning_rate": 2.7026315789473684e-05, "loss": 3.4591, "norm_diff": 0.0409, "norm_loss": 0.0, "num_token_doc": 66.6649, "num_token_overlap": 11.6243, "num_token_query": 31.3365, "num_token_union": 65.0788, "num_word_context": 202.192, "num_word_doc": 49.7309, "num_word_query": 23.2699, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37010.959, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 1.8877, "query_norm": 1.8662, "queue_k_norm": 1.9101, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3365, "sent_len_1": 66.6649, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.6725, "stdk": 0.0489, "stdq": 0.0428, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 97300 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.46, "doc_norm": 1.9158, "encoder_q-embeddings": 20430.9609, "encoder_q-layer.0": 14374.1982, "encoder_q-layer.1": 15534.1865, "encoder_q-layer.10": 24629.0039, "encoder_q-layer.11": 57757.8438, "encoder_q-layer.2": 17382.7754, "encoder_q-layer.3": 17177.0703, "encoder_q-layer.4": 18098.4531, "encoder_q-layer.5": 18640.7305, "encoder_q-layer.6": 20610.0352, "encoder_q-layer.7": 23246.5977, "encoder_q-layer.8": 26992.7344, "encoder_q-layer.9": 23853.5566, "epoch": 0.42, "inbatch_neg_score": 1.8915, "inbatch_pos_score": 2.498, "learning_rate": 2.7000000000000002e-05, "loss": 3.46, "norm_diff": 0.0446, "norm_loss": 0.0, "num_token_doc": 66.8688, "num_token_overlap": 11.69, "num_token_query": 31.4928, "num_token_union": 65.2221, "num_word_context": 202.5954, "num_word_doc": 49.8856, "num_word_query": 23.3933, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38339.2646, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.8906, "query_norm": 1.8712, "queue_k_norm": 1.9122, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4928, "sent_len_1": 66.8688, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.06, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 97400 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.4577, "doc_norm": 1.9081, "encoder_q-embeddings": 21466.1523, "encoder_q-layer.0": 14649.2354, "encoder_q-layer.1": 16111.2188, "encoder_q-layer.10": 27305.1211, "encoder_q-layer.11": 64160.3867, "encoder_q-layer.2": 17912.957, "encoder_q-layer.3": 17717.6719, "encoder_q-layer.4": 18602.4863, "encoder_q-layer.5": 18906.0117, "encoder_q-layer.6": 21107.5039, "encoder_q-layer.7": 23482.709, "encoder_q-layer.8": 29283.9062, "encoder_q-layer.9": 28127.6465, "epoch": 0.42, "inbatch_neg_score": 1.9054, "inbatch_pos_score": 2.4863, "learning_rate": 2.6973684210526317e-05, "loss": 3.4577, "norm_diff": 0.0487, "norm_loss": 0.0, "num_token_doc": 66.6995, "num_token_overlap": 11.6687, "num_token_query": 31.347, "num_token_union": 65.0909, "num_word_context": 202.2717, "num_word_doc": 49.7872, "num_word_query": 23.2772, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40962.4864, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.9043, "query_norm": 1.8594, "queue_k_norm": 1.9154, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.347, "sent_len_1": 66.6995, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.6763, "stdk": 0.0487, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97500 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.441, "doc_norm": 1.9203, "encoder_q-embeddings": 10249.7275, "encoder_q-layer.0": 7043.5601, "encoder_q-layer.1": 7670.8179, "encoder_q-layer.10": 13160.3779, "encoder_q-layer.11": 29533.7324, "encoder_q-layer.2": 8736.5771, "encoder_q-layer.3": 8921.6162, "encoder_q-layer.4": 9040.5781, "encoder_q-layer.5": 9366.0186, "encoder_q-layer.6": 10512.668, "encoder_q-layer.7": 11065.8096, "encoder_q-layer.8": 13764.9688, "encoder_q-layer.9": 12181.1729, "epoch": 0.42, "inbatch_neg_score": 1.9095, "inbatch_pos_score": 2.5234, "learning_rate": 2.6947368421052632e-05, "loss": 3.441, "norm_diff": 0.0544, "norm_loss": 0.0, "num_token_doc": 66.9737, "num_token_overlap": 11.7094, "num_token_query": 31.3993, "num_token_union": 65.1938, "num_word_context": 202.2387, "num_word_doc": 49.9678, "num_word_query": 23.3329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19441.8171, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9082, "query_norm": 1.8659, "queue_k_norm": 1.9185, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3993, "sent_len_1": 66.9737, "sent_len_max_0": 128.0, "sent_len_max_1": 208.805, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97600 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.4637, "doc_norm": 1.9229, "encoder_q-embeddings": 9438.8105, "encoder_q-layer.0": 6524.1816, "encoder_q-layer.1": 6913.4976, "encoder_q-layer.10": 13466.2178, "encoder_q-layer.11": 30339.1582, "encoder_q-layer.2": 7599.8159, "encoder_q-layer.3": 7851.2832, "encoder_q-layer.4": 8335.2842, "encoder_q-layer.5": 8452.1162, "encoder_q-layer.6": 9293.0361, "encoder_q-layer.7": 11056.7012, "encoder_q-layer.8": 13223.7803, "encoder_q-layer.9": 12116.4248, "epoch": 0.42, "inbatch_neg_score": 1.9109, "inbatch_pos_score": 2.5059, "learning_rate": 2.6921052631578947e-05, "loss": 3.4637, "norm_diff": 0.0623, "norm_loss": 0.0, "num_token_doc": 66.8155, "num_token_overlap": 11.6901, "num_token_query": 31.4311, "num_token_union": 65.1748, "num_word_context": 202.4018, "num_word_doc": 49.8413, "num_word_query": 23.3387, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18899.7958, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9102, "query_norm": 1.8606, "queue_k_norm": 1.9192, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4311, "sent_len_1": 66.8155, "sent_len_max_0": 127.9912, "sent_len_max_1": 211.1275, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 97700 }, { "accuracy": 48.1445, "active_queue_size": 16384.0, "cl_loss": 3.4639, "doc_norm": 1.9245, "encoder_q-embeddings": 10841.1768, "encoder_q-layer.0": 7516.9722, "encoder_q-layer.1": 8310.9248, "encoder_q-layer.10": 13132.1113, "encoder_q-layer.11": 30156.0957, "encoder_q-layer.2": 9245.6895, "encoder_q-layer.3": 9359.2402, "encoder_q-layer.4": 9546.415, "encoder_q-layer.5": 9864.8633, "encoder_q-layer.6": 10614.2812, "encoder_q-layer.7": 11714.0078, "encoder_q-layer.8": 13888.9824, "encoder_q-layer.9": 12382.8359, "epoch": 0.42, "inbatch_neg_score": 1.9167, "inbatch_pos_score": 2.5078, "learning_rate": 2.6894736842105266e-05, "loss": 3.4639, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.6977, "num_token_overlap": 11.6359, "num_token_query": 31.3076, "num_token_union": 65.0768, "num_word_context": 202.3496, "num_word_doc": 49.7499, "num_word_query": 23.2487, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20064.2836, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.917, "query_norm": 1.8558, "queue_k_norm": 1.9212, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3076, "sent_len_1": 66.6977, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.5, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 97800 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4516, "doc_norm": 1.9279, "encoder_q-embeddings": 14508.4111, "encoder_q-layer.0": 11042.7842, "encoder_q-layer.1": 11629.752, "encoder_q-layer.10": 12661.6113, "encoder_q-layer.11": 30248.0723, "encoder_q-layer.2": 13638.4229, "encoder_q-layer.3": 14078.9248, "encoder_q-layer.4": 14620.3691, "encoder_q-layer.5": 13980.4209, "encoder_q-layer.6": 13682.3682, "encoder_q-layer.7": 12844.5264, "encoder_q-layer.8": 14460.6992, "encoder_q-layer.9": 12624.1719, "epoch": 0.42, "inbatch_neg_score": 1.92, "inbatch_pos_score": 2.5156, "learning_rate": 2.686842105263158e-05, "loss": 3.4516, "norm_diff": 0.0747, "norm_loss": 0.0, "num_token_doc": 66.5286, "num_token_overlap": 11.6973, "num_token_query": 31.4918, "num_token_union": 65.0316, "num_word_context": 202.1837, "num_word_doc": 49.6346, "num_word_query": 23.3971, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23187.2392, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9199, "query_norm": 1.8531, "queue_k_norm": 1.9252, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4918, "sent_len_1": 66.5286, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.855, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 97900 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.451, "doc_norm": 1.9253, "encoder_q-embeddings": 9798.5508, "encoder_q-layer.0": 6944.8774, "encoder_q-layer.1": 7254.0063, "encoder_q-layer.10": 12807.1074, "encoder_q-layer.11": 29025.4883, "encoder_q-layer.2": 8056.9199, "encoder_q-layer.3": 8321.2715, "encoder_q-layer.4": 9001.917, "encoder_q-layer.5": 9266.0664, "encoder_q-layer.6": 9896.3486, "encoder_q-layer.7": 10968.3301, "encoder_q-layer.8": 13132.1543, "encoder_q-layer.9": 12616.9521, "epoch": 0.43, "inbatch_neg_score": 1.9216, "inbatch_pos_score": 2.5352, "learning_rate": 2.6842105263157896e-05, "loss": 3.451, "norm_diff": 0.0639, "norm_loss": 0.0, "num_token_doc": 66.7664, "num_token_overlap": 11.6914, "num_token_query": 31.382, "num_token_union": 65.0886, "num_word_context": 202.2798, "num_word_doc": 49.8167, "num_word_query": 23.3075, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18767.5185, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9209, "query_norm": 1.8614, "queue_k_norm": 1.926, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.382, "sent_len_1": 66.7664, "sent_len_max_0": 127.9762, "sent_len_max_1": 211.045, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98000 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4592, "doc_norm": 1.9277, "encoder_q-embeddings": 10087.8867, "encoder_q-layer.0": 7091.4966, "encoder_q-layer.1": 7477.9492, "encoder_q-layer.10": 13435.3369, "encoder_q-layer.11": 31472.8086, "encoder_q-layer.2": 8331.0488, "encoder_q-layer.3": 8969.1943, "encoder_q-layer.4": 9426.7168, "encoder_q-layer.5": 9672.7197, "encoder_q-layer.6": 10560.7754, "encoder_q-layer.7": 12498.7715, "encoder_q-layer.8": 14623.1484, "encoder_q-layer.9": 13156.7871, "epoch": 0.43, "inbatch_neg_score": 1.9232, "inbatch_pos_score": 2.5195, "learning_rate": 2.681578947368421e-05, "loss": 3.4592, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.915, "num_token_overlap": 11.6737, "num_token_query": 31.4256, "num_token_union": 65.2327, "num_word_context": 202.4426, "num_word_doc": 49.8772, "num_word_query": 23.3427, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20184.026, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9229, "query_norm": 1.8564, "queue_k_norm": 1.9293, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4256, "sent_len_1": 66.915, "sent_len_max_0": 127.985, "sent_len_max_1": 210.6062, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 98100 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4667, "doc_norm": 1.9285, "encoder_q-embeddings": 10054.3174, "encoder_q-layer.0": 6808.8638, "encoder_q-layer.1": 7259.9106, "encoder_q-layer.10": 12974.458, "encoder_q-layer.11": 29892.3379, "encoder_q-layer.2": 8063.1968, "encoder_q-layer.3": 8052.6099, "encoder_q-layer.4": 8532.6123, "encoder_q-layer.5": 8975.4736, "encoder_q-layer.6": 10197.8252, "encoder_q-layer.7": 11523.2822, "encoder_q-layer.8": 14454.8721, "encoder_q-layer.9": 12736.5537, "epoch": 0.43, "inbatch_neg_score": 1.9255, "inbatch_pos_score": 2.5332, "learning_rate": 2.6789473684210526e-05, "loss": 3.4667, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.8621, "num_token_overlap": 11.6886, "num_token_query": 31.4587, "num_token_union": 65.2107, "num_word_context": 202.5307, "num_word_doc": 49.8931, "num_word_query": 23.3568, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19301.3304, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9258, "query_norm": 1.8579, "queue_k_norm": 1.9297, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4587, "sent_len_1": 66.8621, "sent_len_max_0": 127.9912, "sent_len_max_1": 206.9225, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98200 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.4576, "doc_norm": 1.9294, "encoder_q-embeddings": 9959.624, "encoder_q-layer.0": 6756.1689, "encoder_q-layer.1": 7160.9331, "encoder_q-layer.10": 15573.0654, "encoder_q-layer.11": 31548.791, "encoder_q-layer.2": 8026.0098, "encoder_q-layer.3": 8455.0234, "encoder_q-layer.4": 9153.1875, "encoder_q-layer.5": 9296.8711, "encoder_q-layer.6": 10628.1338, "encoder_q-layer.7": 11921.6699, "encoder_q-layer.8": 14627.7988, "encoder_q-layer.9": 13875.624, "epoch": 0.43, "inbatch_neg_score": 1.9271, "inbatch_pos_score": 2.541, "learning_rate": 2.6763157894736844e-05, "loss": 3.4576, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7448, "num_token_overlap": 11.6729, "num_token_query": 31.3684, "num_token_union": 65.1089, "num_word_context": 202.3467, "num_word_doc": 49.8249, "num_word_query": 23.2969, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19890.0217, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9268, "query_norm": 1.8599, "queue_k_norm": 1.9317, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3684, "sent_len_1": 66.7448, "sent_len_max_0": 127.97, "sent_len_max_1": 207.615, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98300 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.4757, "doc_norm": 1.9339, "encoder_q-embeddings": 10012.5166, "encoder_q-layer.0": 6874.4829, "encoder_q-layer.1": 7153.4873, "encoder_q-layer.10": 12264.7275, "encoder_q-layer.11": 30742.3262, "encoder_q-layer.2": 7952.9434, "encoder_q-layer.3": 8492.9297, "encoder_q-layer.4": 8995.6602, "encoder_q-layer.5": 9576.3398, "encoder_q-layer.6": 10705.6953, "encoder_q-layer.7": 11593.8896, "encoder_q-layer.8": 13445.0957, "encoder_q-layer.9": 12284.8096, "epoch": 0.43, "inbatch_neg_score": 1.9276, "inbatch_pos_score": 2.5195, "learning_rate": 2.673684210526316e-05, "loss": 3.4757, "norm_diff": 0.0856, "norm_loss": 0.0, "num_token_doc": 66.5959, "num_token_overlap": 11.7111, "num_token_query": 31.4533, "num_token_union": 65.0244, "num_word_context": 202.1776, "num_word_doc": 49.6903, "num_word_query": 23.3641, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19438.0988, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9277, "query_norm": 1.8484, "queue_k_norm": 1.9312, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4533, "sent_len_1": 66.5959, "sent_len_max_0": 127.97, "sent_len_max_1": 207.3187, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0491, "stdqueue_q": 0.0, "step": 98400 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4396, "doc_norm": 1.9356, "encoder_q-embeddings": 10241.5166, "encoder_q-layer.0": 7053.1636, "encoder_q-layer.1": 7474.4897, "encoder_q-layer.10": 12666.875, "encoder_q-layer.11": 29597.1816, "encoder_q-layer.2": 8278.4609, "encoder_q-layer.3": 8664.4404, "encoder_q-layer.4": 9034.8555, "encoder_q-layer.5": 8969.8936, "encoder_q-layer.6": 10186.4648, "encoder_q-layer.7": 11135.3955, "encoder_q-layer.8": 13678.8438, "encoder_q-layer.9": 12325.8984, "epoch": 0.43, "inbatch_neg_score": 1.9301, "inbatch_pos_score": 2.5254, "learning_rate": 2.6710526315789474e-05, "loss": 3.4396, "norm_diff": 0.0876, "norm_loss": 0.0, "num_token_doc": 66.7646, "num_token_overlap": 11.685, "num_token_query": 31.3068, "num_token_union": 65.0469, "num_word_context": 202.2602, "num_word_doc": 49.8121, "num_word_query": 23.2507, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19433.7702, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9297, "query_norm": 1.8481, "queue_k_norm": 1.9341, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3068, "sent_len_1": 66.7646, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.8013, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98500 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.4504, "doc_norm": 1.9354, "encoder_q-embeddings": 15518.2832, "encoder_q-layer.0": 10612.9844, "encoder_q-layer.1": 11222.4541, "encoder_q-layer.10": 11890.29, "encoder_q-layer.11": 29454.8594, "encoder_q-layer.2": 13224.208, "encoder_q-layer.3": 12437.8486, "encoder_q-layer.4": 12257.0283, "encoder_q-layer.5": 11416.3691, "encoder_q-layer.6": 11131.4346, "encoder_q-layer.7": 11686.625, "encoder_q-layer.8": 13531.0332, "encoder_q-layer.9": 11627.9395, "epoch": 0.43, "inbatch_neg_score": 1.9337, "inbatch_pos_score": 2.543, "learning_rate": 2.668421052631579e-05, "loss": 3.4504, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 67.0026, "num_token_overlap": 11.7047, "num_token_query": 31.424, "num_token_union": 65.2688, "num_word_context": 202.4062, "num_word_doc": 50.0078, "num_word_query": 23.3506, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21916.4095, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9336, "query_norm": 1.8527, "queue_k_norm": 1.9352, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.424, "sent_len_1": 67.0026, "sent_len_max_0": 127.975, "sent_len_max_1": 208.9, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98600 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.4603, "doc_norm": 1.9355, "encoder_q-embeddings": 9461.0576, "encoder_q-layer.0": 6725.5576, "encoder_q-layer.1": 7182.4346, "encoder_q-layer.10": 12402.9297, "encoder_q-layer.11": 28773.3125, "encoder_q-layer.2": 7998.6929, "encoder_q-layer.3": 8068.9248, "encoder_q-layer.4": 8560.1738, "encoder_q-layer.5": 9061.4414, "encoder_q-layer.6": 9773.5771, "encoder_q-layer.7": 10713.9414, "encoder_q-layer.8": 13056.418, "encoder_q-layer.9": 11809.6006, "epoch": 0.43, "inbatch_neg_score": 1.9331, "inbatch_pos_score": 2.5273, "learning_rate": 2.6657894736842104e-05, "loss": 3.4603, "norm_diff": 0.087, "norm_loss": 0.0, "num_token_doc": 66.9726, "num_token_overlap": 11.6693, "num_token_query": 31.3291, "num_token_union": 65.2344, "num_word_context": 202.5385, "num_word_doc": 49.9703, "num_word_query": 23.2654, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18513.2111, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9336, "query_norm": 1.8484, "queue_k_norm": 1.9374, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3291, "sent_len_1": 66.9726, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.57, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 98700 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.4626, "doc_norm": 1.94, "encoder_q-embeddings": 10427.6514, "encoder_q-layer.0": 7118.749, "encoder_q-layer.1": 7819.3892, "encoder_q-layer.10": 14152.6533, "encoder_q-layer.11": 31454.0293, "encoder_q-layer.2": 8791.9053, "encoder_q-layer.3": 9036.1475, "encoder_q-layer.4": 9551.1318, "encoder_q-layer.5": 9738.6943, "encoder_q-layer.6": 10617.8301, "encoder_q-layer.7": 11291.6182, "encoder_q-layer.8": 14276.4766, "encoder_q-layer.9": 13116.4336, "epoch": 0.43, "inbatch_neg_score": 1.9319, "inbatch_pos_score": 2.5332, "learning_rate": 2.6631578947368426e-05, "loss": 3.4626, "norm_diff": 0.0823, "norm_loss": 0.0, "num_token_doc": 66.6557, "num_token_overlap": 11.6676, "num_token_query": 31.3914, "num_token_union": 65.0741, "num_word_context": 202.1057, "num_word_doc": 49.7402, "num_word_query": 23.3109, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20045.5025, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9316, "query_norm": 1.8577, "queue_k_norm": 1.9382, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3914, "sent_len_1": 66.6557, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.0288, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98800 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.4691, "doc_norm": 1.9395, "encoder_q-embeddings": 9892.6357, "encoder_q-layer.0": 6758.373, "encoder_q-layer.1": 7237.4443, "encoder_q-layer.10": 13417.498, "encoder_q-layer.11": 30111.4375, "encoder_q-layer.2": 8132.3032, "encoder_q-layer.3": 8436.5, "encoder_q-layer.4": 9001.8135, "encoder_q-layer.5": 9594.4268, "encoder_q-layer.6": 10343.915, "encoder_q-layer.7": 11538.5664, "encoder_q-layer.8": 14088.4033, "encoder_q-layer.9": 12626.3359, "epoch": 0.43, "inbatch_neg_score": 1.9339, "inbatch_pos_score": 2.5234, "learning_rate": 2.6605263157894737e-05, "loss": 3.4691, "norm_diff": 0.0842, "norm_loss": 0.0, "num_token_doc": 66.8002, "num_token_overlap": 11.6558, "num_token_query": 31.2964, "num_token_union": 65.1091, "num_word_context": 202.5064, "num_word_doc": 49.8505, "num_word_query": 23.2538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19553.5511, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9346, "query_norm": 1.8553, "queue_k_norm": 1.9388, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2964, "sent_len_1": 66.8002, "sent_len_max_0": 127.985, "sent_len_max_1": 209.9625, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 98900 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4583, "doc_norm": 1.9408, "encoder_q-embeddings": 12724.1465, "encoder_q-layer.0": 8851.2314, "encoder_q-layer.1": 9661.8672, "encoder_q-layer.10": 12110.8906, "encoder_q-layer.11": 29866.7852, "encoder_q-layer.2": 11216.5225, "encoder_q-layer.3": 11414.7949, "encoder_q-layer.4": 11564.8125, "encoder_q-layer.5": 11859.9531, "encoder_q-layer.6": 12410.8672, "encoder_q-layer.7": 14391.0732, "encoder_q-layer.8": 14261.8525, "encoder_q-layer.9": 12117.8486, "epoch": 0.43, "inbatch_neg_score": 1.9328, "inbatch_pos_score": 2.5293, "learning_rate": 2.6578947368421052e-05, "loss": 3.4583, "norm_diff": 0.0898, "norm_loss": 0.0, "num_token_doc": 66.8121, "num_token_overlap": 11.6516, "num_token_query": 31.3241, "num_token_union": 65.1235, "num_word_context": 202.4142, "num_word_doc": 49.8091, "num_word_query": 23.2558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21345.6801, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9326, "query_norm": 1.851, "queue_k_norm": 1.9421, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3241, "sent_len_1": 66.8121, "sent_len_max_0": 127.9925, "sent_len_max_1": 211.56, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99000 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4495, "doc_norm": 1.9431, "encoder_q-embeddings": 11917.3311, "encoder_q-layer.0": 8258.043, "encoder_q-layer.1": 9166.0977, "encoder_q-layer.10": 12628.9805, "encoder_q-layer.11": 31557.0859, "encoder_q-layer.2": 10902.0859, "encoder_q-layer.3": 10497.6992, "encoder_q-layer.4": 10719.1934, "encoder_q-layer.5": 11283.8936, "encoder_q-layer.6": 11474.3477, "encoder_q-layer.7": 12556.3008, "encoder_q-layer.8": 13885.7246, "encoder_q-layer.9": 12463.5684, "epoch": 0.43, "inbatch_neg_score": 1.937, "inbatch_pos_score": 2.5391, "learning_rate": 2.6552631578947367e-05, "loss": 3.4495, "norm_diff": 0.0862, "norm_loss": 0.0, "num_token_doc": 66.8596, "num_token_overlap": 11.7135, "num_token_query": 31.4322, "num_token_union": 65.1675, "num_word_context": 202.2963, "num_word_doc": 49.8834, "num_word_query": 23.3461, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21180.6913, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9375, "query_norm": 1.8568, "queue_k_norm": 1.9415, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4322, "sent_len_1": 66.8596, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.4875, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99100 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.4514, "doc_norm": 1.9436, "encoder_q-embeddings": 15951.8213, "encoder_q-layer.0": 11059.251, "encoder_q-layer.1": 11908.5039, "encoder_q-layer.10": 12950.3223, "encoder_q-layer.11": 30858.6133, "encoder_q-layer.2": 13640.1904, "encoder_q-layer.3": 14636.1494, "encoder_q-layer.4": 15153.9609, "encoder_q-layer.5": 15546.1445, "encoder_q-layer.6": 15189.459, "encoder_q-layer.7": 15558.3975, "encoder_q-layer.8": 15233.0928, "encoder_q-layer.9": 12851.8926, "epoch": 0.43, "inbatch_neg_score": 1.9392, "inbatch_pos_score": 2.5352, "learning_rate": 2.652631578947369e-05, "loss": 3.4514, "norm_diff": 0.0856, "norm_loss": 0.0, "num_token_doc": 66.9713, "num_token_overlap": 11.687, "num_token_query": 31.3871, "num_token_union": 65.2598, "num_word_context": 202.426, "num_word_doc": 49.9533, "num_word_query": 23.3118, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24513.5793, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9395, "query_norm": 1.858, "queue_k_norm": 1.9435, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3871, "sent_len_1": 66.9713, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.5637, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99200 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.457, "doc_norm": 1.9455, "encoder_q-embeddings": 10238.9902, "encoder_q-layer.0": 6953.0278, "encoder_q-layer.1": 7214.8535, "encoder_q-layer.10": 13027.6826, "encoder_q-layer.11": 30460.4727, "encoder_q-layer.2": 7839.7734, "encoder_q-layer.3": 8087.4727, "encoder_q-layer.4": 8505.7207, "encoder_q-layer.5": 8546.4648, "encoder_q-layer.6": 9798.625, "encoder_q-layer.7": 10870.7178, "encoder_q-layer.8": 13215.834, "encoder_q-layer.9": 12271.7256, "epoch": 0.43, "inbatch_neg_score": 1.9422, "inbatch_pos_score": 2.541, "learning_rate": 2.6500000000000004e-05, "loss": 3.457, "norm_diff": 0.0894, "norm_loss": 0.0, "num_token_doc": 66.6997, "num_token_overlap": 11.6972, "num_token_query": 31.4799, "num_token_union": 65.1231, "num_word_context": 202.0771, "num_word_doc": 49.7818, "num_word_query": 23.3792, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19280.8393, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9434, "query_norm": 1.856, "queue_k_norm": 1.9426, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4799, "sent_len_1": 66.6997, "sent_len_max_0": 127.9463, "sent_len_max_1": 209.105, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 99300 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4527, "doc_norm": 1.9409, "encoder_q-embeddings": 10218.7324, "encoder_q-layer.0": 6956.8623, "encoder_q-layer.1": 7430.894, "encoder_q-layer.10": 12314.3613, "encoder_q-layer.11": 29861.6445, "encoder_q-layer.2": 8271.082, "encoder_q-layer.3": 8395.123, "encoder_q-layer.4": 9185.6445, "encoder_q-layer.5": 9260.7891, "encoder_q-layer.6": 10447.4922, "encoder_q-layer.7": 11405.2539, "encoder_q-layer.8": 13502.7705, "encoder_q-layer.9": 12028.4639, "epoch": 0.43, "inbatch_neg_score": 1.9467, "inbatch_pos_score": 2.5586, "learning_rate": 2.647368421052632e-05, "loss": 3.4527, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.7356, "num_token_overlap": 11.6587, "num_token_query": 31.4018, "num_token_union": 65.1186, "num_word_context": 202.415, "num_word_doc": 49.805, "num_word_query": 23.3185, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19322.914, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9453, "query_norm": 1.863, "queue_k_norm": 1.9431, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4018, "sent_len_1": 66.7356, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.53, "stdk": 0.049, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 99400 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.4467, "doc_norm": 1.9419, "encoder_q-embeddings": 10185.9912, "encoder_q-layer.0": 6887.6973, "encoder_q-layer.1": 7269.3862, "encoder_q-layer.10": 12931.9756, "encoder_q-layer.11": 30035.2637, "encoder_q-layer.2": 8093.8589, "encoder_q-layer.3": 8377.5381, "encoder_q-layer.4": 8587.8262, "encoder_q-layer.5": 8584.6133, "encoder_q-layer.6": 9712.8027, "encoder_q-layer.7": 11475.2393, "encoder_q-layer.8": 14096.1387, "encoder_q-layer.9": 12473.0273, "epoch": 0.43, "inbatch_neg_score": 1.9508, "inbatch_pos_score": 2.5469, "learning_rate": 2.644736842105263e-05, "loss": 3.4467, "norm_diff": 0.0884, "norm_loss": 0.0, "num_token_doc": 66.6523, "num_token_overlap": 11.682, "num_token_query": 31.319, "num_token_union": 65.0106, "num_word_context": 202.3033, "num_word_doc": 49.7311, "num_word_query": 23.2568, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19429.4348, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9492, "query_norm": 1.8534, "queue_k_norm": 1.9437, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.319, "sent_len_1": 66.6523, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.005, "stdk": 0.049, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 99500 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.4609, "doc_norm": 1.9434, "encoder_q-embeddings": 19298.7754, "encoder_q-layer.0": 13697.9502, "encoder_q-layer.1": 14287.5645, "encoder_q-layer.10": 28900.334, "encoder_q-layer.11": 63607.543, "encoder_q-layer.2": 16269.5908, "encoder_q-layer.3": 16693.7617, "encoder_q-layer.4": 17251.8789, "encoder_q-layer.5": 17853.6504, "encoder_q-layer.6": 19622.9082, "encoder_q-layer.7": 23122.1055, "encoder_q-layer.8": 28686.8242, "encoder_q-layer.9": 25848.6895, "epoch": 0.43, "inbatch_neg_score": 1.9544, "inbatch_pos_score": 2.5605, "learning_rate": 2.6421052631578945e-05, "loss": 3.4609, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 66.8357, "num_token_overlap": 11.6808, "num_token_query": 31.3382, "num_token_union": 65.1376, "num_word_context": 202.3191, "num_word_doc": 49.8619, "num_word_query": 23.2791, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39668.5378, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.9551, "query_norm": 1.8607, "queue_k_norm": 1.9462, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3382, "sent_len_1": 66.8357, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.9, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99600 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4577, "doc_norm": 1.9488, "encoder_q-embeddings": 20767.8516, "encoder_q-layer.0": 14408.8047, "encoder_q-layer.1": 15188.624, "encoder_q-layer.10": 25234.5762, "encoder_q-layer.11": 60198.2656, "encoder_q-layer.2": 16396.7656, "encoder_q-layer.3": 16524.416, "encoder_q-layer.4": 18058.6016, "encoder_q-layer.5": 18727.6504, "encoder_q-layer.6": 20349.2891, "encoder_q-layer.7": 24850.9062, "encoder_q-layer.8": 28590.7832, "encoder_q-layer.9": 24835.7305, "epoch": 0.43, "inbatch_neg_score": 1.9593, "inbatch_pos_score": 2.5605, "learning_rate": 2.6394736842105267e-05, "loss": 3.4577, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.7467, "num_token_overlap": 11.658, "num_token_query": 31.2623, "num_token_union": 65.051, "num_word_context": 202.2087, "num_word_doc": 49.8061, "num_word_query": 23.2116, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39471.2557, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 1.959, "query_norm": 1.8687, "queue_k_norm": 1.9471, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2623, "sent_len_1": 66.7467, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.4238, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99700 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.4471, "doc_norm": 1.9419, "encoder_q-embeddings": 10205.3838, "encoder_q-layer.0": 7011.4512, "encoder_q-layer.1": 7181.3994, "encoder_q-layer.10": 12867.4648, "encoder_q-layer.11": 29867.125, "encoder_q-layer.2": 7912.3145, "encoder_q-layer.3": 8119.4224, "encoder_q-layer.4": 8932.4922, "encoder_q-layer.5": 8933.5391, "encoder_q-layer.6": 9977.7305, "encoder_q-layer.7": 11761.9199, "encoder_q-layer.8": 13463.7207, "encoder_q-layer.9": 12417.5107, "epoch": 0.43, "inbatch_neg_score": 1.9653, "inbatch_pos_score": 2.5508, "learning_rate": 2.6368421052631582e-05, "loss": 3.4471, "norm_diff": 0.085, "norm_loss": 0.0, "num_token_doc": 66.8264, "num_token_overlap": 11.719, "num_token_query": 31.4392, "num_token_union": 65.1436, "num_word_context": 202.192, "num_word_doc": 49.8329, "num_word_query": 23.3494, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19391.2851, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9648, "query_norm": 1.8569, "queue_k_norm": 1.9479, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4392, "sent_len_1": 66.8264, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.3925, "stdk": 0.0489, "stdq": 0.0425, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 99800 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.4693, "doc_norm": 1.9457, "encoder_q-embeddings": 9699.9326, "encoder_q-layer.0": 6831.0156, "encoder_q-layer.1": 7002.0107, "encoder_q-layer.10": 12609.1074, "encoder_q-layer.11": 28882.3594, "encoder_q-layer.2": 7951.9165, "encoder_q-layer.3": 8223.6865, "encoder_q-layer.4": 8793.3184, "encoder_q-layer.5": 9140.7168, "encoder_q-layer.6": 10021.1084, "encoder_q-layer.7": 11068.4922, "encoder_q-layer.8": 13191.8623, "encoder_q-layer.9": 12262.0762, "epoch": 0.43, "inbatch_neg_score": 1.9686, "inbatch_pos_score": 2.5859, "learning_rate": 2.6342105263157897e-05, "loss": 3.4693, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.7052, "num_token_overlap": 11.6795, "num_token_query": 31.4138, "num_token_union": 65.0982, "num_word_context": 202.118, "num_word_doc": 49.7838, "num_word_query": 23.3307, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18833.7313, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 1.9688, "query_norm": 1.8755, "queue_k_norm": 1.9483, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4138, "sent_len_1": 66.7052, "sent_len_max_0": 127.99, "sent_len_max_1": 207.7775, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 99900 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4397, "doc_norm": 1.9497, "encoder_q-embeddings": 4959.3511, "encoder_q-layer.0": 3616.7417, "encoder_q-layer.1": 3897.3259, "encoder_q-layer.10": 6273.3257, "encoder_q-layer.11": 15140.8594, "encoder_q-layer.2": 4495.3066, "encoder_q-layer.3": 4935.3271, "encoder_q-layer.4": 5008.7827, "encoder_q-layer.5": 5150.9067, "encoder_q-layer.6": 5579.4204, "encoder_q-layer.7": 5904.252, "encoder_q-layer.8": 7002.8027, "encoder_q-layer.9": 6213.5483, "epoch": 0.43, "inbatch_neg_score": 1.9723, "inbatch_pos_score": 2.5781, "learning_rate": 2.6315789473684212e-05, "loss": 3.4397, "norm_diff": 0.0811, "norm_loss": 0.0, "num_token_doc": 66.7786, "num_token_overlap": 11.7171, "num_token_query": 31.451, "num_token_union": 65.1539, "num_word_context": 202.1333, "num_word_doc": 49.8487, "num_word_query": 23.3626, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9957.829, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.9727, "query_norm": 1.8686, "queue_k_norm": 1.9492, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.451, "sent_len_1": 66.7786, "sent_len_max_0": 127.96, "sent_len_max_1": 208.7225, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100000 }, { "dev_runtime": 31.7678, "dev_samples_per_second": 1.007, "dev_steps_per_second": 0.031, "epoch": 0.43, "step": 100000, "test_accuracy": 93.798828125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3444693386554718, "test_doc_norm": 1.931605339050293, "test_inbatch_neg_score": 2.252593517303467, "test_inbatch_pos_score": 3.26454496383667, "test_loss": 0.3444693386554718, "test_loss_align": 1.0754996538162231, "test_loss_unif": -3.4620726108551025, "test_loss_unif_q@queue": -3.4620726108551025, "test_norm_diff": 0.03207101672887802, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.9676735401153564, "test_query_norm": 1.9636763334274292, "test_queue_k_norm": 1.9495453834533691, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044082947075366974, "test_stdq": 0.04386937618255615, "test_stdqueue_k": 0.04925723001360893, "test_stdqueue_q": 0.0 }, { "dev_runtime": 31.7678, "dev_samples_per_second": 1.007, "dev_steps_per_second": 0.031, "epoch": 0.43, "eval_beir-arguana_ndcg@10": 0.38844, "eval_beir-arguana_recall@10": 0.66074, "eval_beir-arguana_recall@100": 0.93385, "eval_beir-arguana_recall@20": 0.79232, "eval_beir-avg_ndcg@10": 0.39649425, "eval_beir-avg_recall@10": 0.4650965833333334, "eval_beir-avg_recall@100": 0.6445751666666666, "eval_beir-avg_recall@20": 0.52575675, "eval_beir-cqadupstack_ndcg@10": 0.2861625, "eval_beir-cqadupstack_recall@10": 0.3857458333333333, "eval_beir-cqadupstack_recall@100": 0.6233916666666667, "eval_beir-cqadupstack_recall@20": 0.4529275000000001, "eval_beir-fiqa_ndcg@10": 0.26042, "eval_beir-fiqa_recall@10": 0.32002, "eval_beir-fiqa_recall@100": 0.60288, "eval_beir-fiqa_recall@20": 0.40158, "eval_beir-nfcorpus_ndcg@10": 0.31588, "eval_beir-nfcorpus_recall@10": 0.15306, "eval_beir-nfcorpus_recall@100": 0.30013, "eval_beir-nfcorpus_recall@20": 0.18707, "eval_beir-nq_ndcg@10": 0.31474, "eval_beir-nq_recall@10": 0.50365, "eval_beir-nq_recall@100": 0.82966, "eval_beir-nq_recall@20": 0.62739, "eval_beir-quora_ndcg@10": 0.78828, "eval_beir-quora_recall@10": 0.89395, "eval_beir-quora_recall@100": 0.97984, "eval_beir-quora_recall@20": 0.93398, "eval_beir-scidocs_ndcg@10": 0.16274, "eval_beir-scidocs_recall@10": 0.17032, "eval_beir-scidocs_recall@100": 0.38188, "eval_beir-scidocs_recall@20": 0.22802, "eval_beir-scifact_ndcg@10": 0.64579, "eval_beir-scifact_recall@10": 0.78133, "eval_beir-scifact_recall@100": 0.91656, "eval_beir-scifact_recall@20": 0.82478, "eval_beir-trec-covid_ndcg@10": 0.62084, "eval_beir-trec-covid_recall@10": 0.654, "eval_beir-trec-covid_recall@100": 0.4618, "eval_beir-trec-covid_recall@20": 0.597, "eval_beir-webis-touche2020_ndcg@10": 0.18165, "eval_beir-webis-touche2020_recall@10": 0.12815, "eval_beir-webis-touche2020_recall@100": 0.41576, "eval_beir-webis-touche2020_recall@20": 0.2125, "eval_senteval-avg_sts": 0.7377236395502316, "eval_senteval-sickr_spearman": 0.726585203583891, "eval_senteval-stsb_spearman": 0.7488620755165721, "step": 100000, "test_accuracy": 93.798828125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3444693386554718, "test_doc_norm": 1.931605339050293, "test_inbatch_neg_score": 2.252593517303467, "test_inbatch_pos_score": 3.26454496383667, "test_loss": 0.3444693386554718, "test_loss_align": 1.0754996538162231, "test_loss_unif": -3.4620726108551025, "test_loss_unif_q@queue": -3.4620726108551025, "test_norm_diff": 0.03207101672887802, "test_norm_loss": 0.0, "test_q@queue_neg_score": 1.9676735401153564, "test_query_norm": 1.9636763334274292, "test_queue_k_norm": 1.9495453834533691, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044082947075366974, "test_stdq": 0.04386937618255615, "test_stdqueue_k": 0.04925723001360893, "test_stdqueue_q": 0.0 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.4481, "doc_norm": 1.9478, "encoder_q-embeddings": 5257.7773, "encoder_q-layer.0": 3573.2087, "encoder_q-layer.1": 3819.657, "encoder_q-layer.10": 6436.9971, "encoder_q-layer.11": 15425.7129, "encoder_q-layer.2": 4356.6543, "encoder_q-layer.3": 4621.5352, "encoder_q-layer.4": 5131.8247, "encoder_q-layer.5": 5128.7021, "encoder_q-layer.6": 5470.3164, "encoder_q-layer.7": 6147.5771, "encoder_q-layer.8": 7753.0215, "encoder_q-layer.9": 6711.7505, "epoch": 0.43, "inbatch_neg_score": 1.9782, "inbatch_pos_score": 2.5742, "learning_rate": 2.6289473684210524e-05, "loss": 3.4481, "norm_diff": 0.0787, "norm_loss": 0.0, "num_token_doc": 66.8936, "num_token_overlap": 11.7083, "num_token_query": 31.3763, "num_token_union": 65.1409, "num_word_context": 202.0507, "num_word_doc": 49.9045, "num_word_query": 23.3018, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10258.4525, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.9775, "query_norm": 1.8691, "queue_k_norm": 1.9509, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3763, "sent_len_1": 66.8936, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.25, "stdk": 0.049, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 100100 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4426, "doc_norm": 1.9512, "encoder_q-embeddings": 4886.251, "encoder_q-layer.0": 3429.1646, "encoder_q-layer.1": 3494.1531, "encoder_q-layer.10": 5872.354, "encoder_q-layer.11": 14699.9111, "encoder_q-layer.2": 4020.3081, "encoder_q-layer.3": 4110.5234, "encoder_q-layer.4": 4361.373, "encoder_q-layer.5": 4486.9385, "encoder_q-layer.6": 5044.0303, "encoder_q-layer.7": 5402.2476, "encoder_q-layer.8": 6539.064, "encoder_q-layer.9": 5986.8091, "epoch": 0.43, "inbatch_neg_score": 1.9773, "inbatch_pos_score": 2.5859, "learning_rate": 2.6263157894736845e-05, "loss": 3.4426, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.6518, "num_token_overlap": 11.6855, "num_token_query": 31.3892, "num_token_union": 65.0607, "num_word_context": 202.1554, "num_word_doc": 49.7442, "num_word_query": 23.3035, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9366.0257, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.9785, "query_norm": 1.879, "queue_k_norm": 1.9511, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3892, "sent_len_1": 66.6518, "sent_len_max_0": 127.94, "sent_len_max_1": 207.6612, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100200 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4527, "doc_norm": 1.9524, "encoder_q-embeddings": 4943.2451, "encoder_q-layer.0": 3384.0132, "encoder_q-layer.1": 3598.6592, "encoder_q-layer.10": 6410.415, "encoder_q-layer.11": 14602.1094, "encoder_q-layer.2": 3953.4053, "encoder_q-layer.3": 4265.5171, "encoder_q-layer.4": 4375.8584, "encoder_q-layer.5": 4432.4844, "encoder_q-layer.6": 4989.3184, "encoder_q-layer.7": 5488.2622, "encoder_q-layer.8": 6798.7642, "encoder_q-layer.9": 6046.5229, "epoch": 0.44, "inbatch_neg_score": 1.9921, "inbatch_pos_score": 2.5879, "learning_rate": 2.623684210526316e-05, "loss": 3.4527, "norm_diff": 0.0647, "norm_loss": 0.0, "num_token_doc": 66.6543, "num_token_overlap": 11.6877, "num_token_query": 31.447, "num_token_union": 65.07, "num_word_context": 202.1362, "num_word_doc": 49.7592, "num_word_query": 23.3608, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9454.1305, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.9922, "query_norm": 1.8877, "queue_k_norm": 1.9524, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.447, "sent_len_1": 66.6543, "sent_len_max_0": 127.9975, "sent_len_max_1": 206.84, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 100300 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.456, "doc_norm": 1.952, "encoder_q-embeddings": 4763.6992, "encoder_q-layer.0": 3420.593, "encoder_q-layer.1": 3759.9739, "encoder_q-layer.10": 6283.9971, "encoder_q-layer.11": 14763.2744, "encoder_q-layer.2": 4148.1475, "encoder_q-layer.3": 4187.0483, "encoder_q-layer.4": 4426.0513, "encoder_q-layer.5": 4403.3589, "encoder_q-layer.6": 4818.5254, "encoder_q-layer.7": 5237.9492, "encoder_q-layer.8": 6568.2217, "encoder_q-layer.9": 6110.6934, "epoch": 0.44, "inbatch_neg_score": 1.9918, "inbatch_pos_score": 2.5742, "learning_rate": 2.6210526315789475e-05, "loss": 3.456, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.844, "num_token_overlap": 11.6695, "num_token_query": 31.3419, "num_token_union": 65.1505, "num_word_context": 202.5033, "num_word_doc": 49.8393, "num_word_query": 23.2626, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9377.1128, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 1.9912, "query_norm": 1.8754, "queue_k_norm": 1.953, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3419, "sent_len_1": 66.844, "sent_len_max_0": 127.99, "sent_len_max_1": 210.3487, "stdk": 0.0491, "stdq": 0.0425, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100400 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.4533, "doc_norm": 1.9539, "encoder_q-embeddings": 5511.7471, "encoder_q-layer.0": 3781.9602, "encoder_q-layer.1": 4002.6511, "encoder_q-layer.10": 6447.7852, "encoder_q-layer.11": 15306.9141, "encoder_q-layer.2": 4597.4165, "encoder_q-layer.3": 4649.7769, "encoder_q-layer.4": 4985.1426, "encoder_q-layer.5": 4847.2627, "encoder_q-layer.6": 5379.5713, "encoder_q-layer.7": 6718.7134, "encoder_q-layer.8": 7355.5923, "encoder_q-layer.9": 6536.9053, "epoch": 0.44, "inbatch_neg_score": 2.0011, "inbatch_pos_score": 2.5879, "learning_rate": 2.618421052631579e-05, "loss": 3.4533, "norm_diff": 0.0594, "norm_loss": 0.0, "num_token_doc": 66.7061, "num_token_overlap": 11.6769, "num_token_query": 31.3803, "num_token_union": 65.0644, "num_word_context": 202.1517, "num_word_doc": 49.7987, "num_word_query": 23.2967, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10064.2601, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0, "query_norm": 1.8945, "queue_k_norm": 1.954, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3803, "sent_len_1": 66.7061, "sent_len_max_0": 127.975, "sent_len_max_1": 207.3587, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100500 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.4538, "doc_norm": 1.9563, "encoder_q-embeddings": 5452.7407, "encoder_q-layer.0": 3732.4902, "encoder_q-layer.1": 3917.1968, "encoder_q-layer.10": 6411.0015, "encoder_q-layer.11": 15700.5312, "encoder_q-layer.2": 4443.9019, "encoder_q-layer.3": 4485.0762, "encoder_q-layer.4": 4677.6665, "encoder_q-layer.5": 4534.7197, "encoder_q-layer.6": 5195.1289, "encoder_q-layer.7": 5669.3613, "encoder_q-layer.8": 6688.3711, "encoder_q-layer.9": 6073.6489, "epoch": 0.44, "inbatch_neg_score": 2.0108, "inbatch_pos_score": 2.5938, "learning_rate": 2.615789473684211e-05, "loss": 3.4538, "norm_diff": 0.0631, "norm_loss": 0.0, "num_token_doc": 66.7544, "num_token_overlap": 11.6549, "num_token_query": 31.3356, "num_token_union": 65.0799, "num_word_context": 202.0723, "num_word_doc": 49.8321, "num_word_query": 23.2692, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9972.0098, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0098, "query_norm": 1.8932, "queue_k_norm": 1.9554, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3356, "sent_len_1": 66.7544, "sent_len_max_0": 127.98, "sent_len_max_1": 208.2475, "stdk": 0.0492, "stdq": 0.0426, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100600 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.4532, "doc_norm": 1.9579, "encoder_q-embeddings": 5019.3804, "encoder_q-layer.0": 3375.9688, "encoder_q-layer.1": 3680.7034, "encoder_q-layer.10": 6223.5098, "encoder_q-layer.11": 15172.2197, "encoder_q-layer.2": 4083.8933, "encoder_q-layer.3": 4138.4424, "encoder_q-layer.4": 4304.9971, "encoder_q-layer.5": 4383.5278, "encoder_q-layer.6": 4809.5781, "encoder_q-layer.7": 5289.5713, "encoder_q-layer.8": 6628.5718, "encoder_q-layer.9": 6189.6284, "epoch": 0.44, "inbatch_neg_score": 2.0222, "inbatch_pos_score": 2.623, "learning_rate": 2.6131578947368424e-05, "loss": 3.4532, "norm_diff": 0.0509, "norm_loss": 0.0, "num_token_doc": 66.7955, "num_token_overlap": 11.6593, "num_token_query": 31.2707, "num_token_union": 65.0871, "num_word_context": 202.2255, "num_word_doc": 49.8707, "num_word_query": 23.2278, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9559.0244, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0215, "query_norm": 1.907, "queue_k_norm": 1.9563, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2707, "sent_len_1": 66.7955, "sent_len_max_0": 127.995, "sent_len_max_1": 207.01, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 100700 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.4311, "doc_norm": 1.9551, "encoder_q-embeddings": 4806.7925, "encoder_q-layer.0": 3484.2271, "encoder_q-layer.1": 3607.0371, "encoder_q-layer.10": 6316.1118, "encoder_q-layer.11": 15058.0332, "encoder_q-layer.2": 3970.8389, "encoder_q-layer.3": 4027.4468, "encoder_q-layer.4": 4273.7812, "encoder_q-layer.5": 4296.9424, "encoder_q-layer.6": 4813.9663, "encoder_q-layer.7": 5409.2886, "encoder_q-layer.8": 6900.1133, "encoder_q-layer.9": 6190.7095, "epoch": 0.44, "inbatch_neg_score": 2.0342, "inbatch_pos_score": 2.625, "learning_rate": 2.610526315789474e-05, "loss": 3.4311, "norm_diff": 0.0379, "norm_loss": 0.0, "num_token_doc": 67.0285, "num_token_overlap": 11.7154, "num_token_query": 31.4551, "num_token_union": 65.3008, "num_word_context": 202.6461, "num_word_doc": 50.0027, "num_word_query": 23.355, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9473.0939, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0332, "query_norm": 1.9171, "queue_k_norm": 1.9589, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4551, "sent_len_1": 67.0285, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.9663, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 100800 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.4439, "doc_norm": 1.9589, "encoder_q-embeddings": 4974.8086, "encoder_q-layer.0": 3486.0542, "encoder_q-layer.1": 3790.478, "encoder_q-layer.10": 6344.792, "encoder_q-layer.11": 14643.4932, "encoder_q-layer.2": 4318.6855, "encoder_q-layer.3": 4403.667, "encoder_q-layer.4": 4562.2881, "encoder_q-layer.5": 4764.5801, "encoder_q-layer.6": 5240.3877, "encoder_q-layer.7": 5679.7861, "encoder_q-layer.8": 7326.5952, "encoder_q-layer.9": 6564.6348, "epoch": 0.44, "inbatch_neg_score": 2.0475, "inbatch_pos_score": 2.6445, "learning_rate": 2.6078947368421053e-05, "loss": 3.4439, "norm_diff": 0.0389, "norm_loss": 0.0, "num_token_doc": 66.7753, "num_token_overlap": 11.6862, "num_token_query": 31.4325, "num_token_union": 65.1617, "num_word_context": 202.1656, "num_word_doc": 49.8122, "num_word_query": 23.3464, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9730.423, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0469, "query_norm": 1.92, "queue_k_norm": 1.9606, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4325, "sent_len_1": 66.7753, "sent_len_max_0": 127.955, "sent_len_max_1": 209.9075, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 100900 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4379, "doc_norm": 1.9608, "encoder_q-embeddings": 4883.0581, "encoder_q-layer.0": 3352.0183, "encoder_q-layer.1": 3593.9478, "encoder_q-layer.10": 7055.9629, "encoder_q-layer.11": 15051.6904, "encoder_q-layer.2": 4001.6506, "encoder_q-layer.3": 4250.1265, "encoder_q-layer.4": 4251.7461, "encoder_q-layer.5": 4462.5552, "encoder_q-layer.6": 4828.8691, "encoder_q-layer.7": 5471.6885, "encoder_q-layer.8": 6927.4727, "encoder_q-layer.9": 6728.8027, "epoch": 0.44, "inbatch_neg_score": 2.0649, "inbatch_pos_score": 2.6758, "learning_rate": 2.605263157894737e-05, "loss": 3.4379, "norm_diff": 0.0318, "norm_loss": 0.0, "num_token_doc": 66.6177, "num_token_overlap": 11.6395, "num_token_query": 31.281, "num_token_union": 65.004, "num_word_context": 201.8091, "num_word_doc": 49.712, "num_word_query": 23.2276, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9538.9715, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0645, "query_norm": 1.9289, "queue_k_norm": 1.9628, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.281, "sent_len_1": 66.6177, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.8825, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101000 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4509, "doc_norm": 1.9654, "encoder_q-embeddings": 4797.8018, "encoder_q-layer.0": 3234.5286, "encoder_q-layer.1": 3462.3113, "encoder_q-layer.10": 6673.2559, "encoder_q-layer.11": 15149.9736, "encoder_q-layer.2": 3831.6519, "encoder_q-layer.3": 4096.981, "encoder_q-layer.4": 4263.0532, "encoder_q-layer.5": 4252.5596, "encoder_q-layer.6": 4814.1685, "encoder_q-layer.7": 5430.7769, "encoder_q-layer.8": 6848.2695, "encoder_q-layer.9": 6276.498, "epoch": 0.44, "inbatch_neg_score": 2.0739, "inbatch_pos_score": 2.6738, "learning_rate": 2.6026315789473687e-05, "loss": 3.4509, "norm_diff": 0.0454, "norm_loss": 0.0, "num_token_doc": 66.9457, "num_token_overlap": 11.7176, "num_token_query": 31.4224, "num_token_union": 65.2086, "num_word_context": 202.5894, "num_word_doc": 49.9662, "num_word_query": 23.3412, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9466.0173, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0723, "query_norm": 1.92, "queue_k_norm": 1.966, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4224, "sent_len_1": 66.9457, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.525, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 101100 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.4339, "doc_norm": 1.9662, "encoder_q-embeddings": 4885.6265, "encoder_q-layer.0": 3251.2581, "encoder_q-layer.1": 3450.1938, "encoder_q-layer.10": 7001.8335, "encoder_q-layer.11": 15655.8438, "encoder_q-layer.2": 3873.3921, "encoder_q-layer.3": 3905.4526, "encoder_q-layer.4": 4324.8516, "encoder_q-layer.5": 4295.3315, "encoder_q-layer.6": 4917.8438, "encoder_q-layer.7": 5619.7075, "encoder_q-layer.8": 6709.6216, "encoder_q-layer.9": 6489.3066, "epoch": 0.44, "inbatch_neg_score": 2.0842, "inbatch_pos_score": 2.6953, "learning_rate": 2.6000000000000002e-05, "loss": 3.4339, "norm_diff": 0.0481, "norm_loss": 0.0, "num_token_doc": 66.711, "num_token_overlap": 11.7022, "num_token_query": 31.4476, "num_token_union": 65.0984, "num_word_context": 202.1631, "num_word_doc": 49.7587, "num_word_query": 23.3555, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9723.0948, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.082, "query_norm": 1.9181, "queue_k_norm": 1.9665, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4476, "sent_len_1": 66.711, "sent_len_max_0": 127.99, "sent_len_max_1": 209.9575, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101200 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4494, "doc_norm": 1.9683, "encoder_q-embeddings": 4727.6299, "encoder_q-layer.0": 3306.7729, "encoder_q-layer.1": 3419.3074, "encoder_q-layer.10": 6536.6006, "encoder_q-layer.11": 14285.623, "encoder_q-layer.2": 3752.4358, "encoder_q-layer.3": 3833.0137, "encoder_q-layer.4": 4014.0037, "encoder_q-layer.5": 4186.2075, "encoder_q-layer.6": 4778.2456, "encoder_q-layer.7": 5529.9546, "encoder_q-layer.8": 6514.7373, "encoder_q-layer.9": 6042.999, "epoch": 0.44, "inbatch_neg_score": 2.0886, "inbatch_pos_score": 2.6992, "learning_rate": 2.5973684210526317e-05, "loss": 3.4494, "norm_diff": 0.0506, "norm_loss": 0.0, "num_token_doc": 66.7939, "num_token_overlap": 11.7137, "num_token_query": 31.4613, "num_token_union": 65.1208, "num_word_context": 202.4829, "num_word_doc": 49.8406, "num_word_query": 23.3623, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9240.04, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0879, "query_norm": 1.9177, "queue_k_norm": 1.9673, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4613, "sent_len_1": 66.7939, "sent_len_max_0": 127.98, "sent_len_max_1": 207.9975, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 101300 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4419, "doc_norm": 1.9695, "encoder_q-embeddings": 5740.6006, "encoder_q-layer.0": 4064.5361, "encoder_q-layer.1": 4550.8501, "encoder_q-layer.10": 6257.6045, "encoder_q-layer.11": 15383.8154, "encoder_q-layer.2": 4899.835, "encoder_q-layer.3": 4876.7817, "encoder_q-layer.4": 5109.4106, "encoder_q-layer.5": 5300.6533, "encoder_q-layer.6": 5297.437, "encoder_q-layer.7": 5891.8623, "encoder_q-layer.8": 6876.1992, "encoder_q-layer.9": 6281.1201, "epoch": 0.44, "inbatch_neg_score": 2.0913, "inbatch_pos_score": 2.6914, "learning_rate": 2.5947368421052632e-05, "loss": 3.4419, "norm_diff": 0.0612, "norm_loss": 0.0, "num_token_doc": 66.8792, "num_token_overlap": 11.7152, "num_token_query": 31.474, "num_token_union": 65.1988, "num_word_context": 202.656, "num_word_doc": 49.9306, "num_word_query": 23.3772, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10379.6011, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0918, "query_norm": 1.9082, "queue_k_norm": 1.9699, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.474, "sent_len_1": 66.8792, "sent_len_max_0": 127.97, "sent_len_max_1": 208.4613, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 101400 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4563, "doc_norm": 1.9711, "encoder_q-embeddings": 5013.541, "encoder_q-layer.0": 3497.8892, "encoder_q-layer.1": 3832.8625, "encoder_q-layer.10": 6577.4746, "encoder_q-layer.11": 17166.1973, "encoder_q-layer.2": 4188.6533, "encoder_q-layer.3": 4344.8809, "encoder_q-layer.4": 4571.8154, "encoder_q-layer.5": 4982.3613, "encoder_q-layer.6": 5248.2686, "encoder_q-layer.7": 5872.2563, "encoder_q-layer.8": 6900.4028, "encoder_q-layer.9": 6350.6821, "epoch": 0.44, "inbatch_neg_score": 2.0932, "inbatch_pos_score": 2.6934, "learning_rate": 2.5921052631578947e-05, "loss": 3.4563, "norm_diff": 0.0617, "norm_loss": 0.0, "num_token_doc": 66.5639, "num_token_overlap": 11.621, "num_token_query": 31.2657, "num_token_union": 64.9977, "num_word_context": 202.0573, "num_word_doc": 49.6968, "num_word_query": 23.2244, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10136.3516, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0938, "query_norm": 1.9095, "queue_k_norm": 1.9718, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2657, "sent_len_1": 66.5639, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.7312, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 101500 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.4535, "doc_norm": 1.973, "encoder_q-embeddings": 5932.2437, "encoder_q-layer.0": 4284.125, "encoder_q-layer.1": 5131.5688, "encoder_q-layer.10": 6670.0088, "encoder_q-layer.11": 14609.875, "encoder_q-layer.2": 6031.2349, "encoder_q-layer.3": 6251.8115, "encoder_q-layer.4": 6248.7827, "encoder_q-layer.5": 7828.1523, "encoder_q-layer.6": 7643.8853, "encoder_q-layer.7": 7225.0845, "encoder_q-layer.8": 6924.5757, "encoder_q-layer.9": 6398.5532, "epoch": 0.44, "inbatch_neg_score": 2.095, "inbatch_pos_score": 2.707, "learning_rate": 2.5894736842105265e-05, "loss": 3.4535, "norm_diff": 0.0737, "norm_loss": 0.0, "num_token_doc": 66.5898, "num_token_overlap": 11.6887, "num_token_query": 31.4426, "num_token_union": 65.0512, "num_word_context": 202.1424, "num_word_doc": 49.7118, "num_word_query": 23.3399, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10912.4734, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0957, "query_norm": 1.8993, "queue_k_norm": 1.9746, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4426, "sent_len_1": 66.5898, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.1813, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101600 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4527, "doc_norm": 1.9738, "encoder_q-embeddings": 4677.3701, "encoder_q-layer.0": 3265.3193, "encoder_q-layer.1": 3457.9626, "encoder_q-layer.10": 6255.2168, "encoder_q-layer.11": 15077.8594, "encoder_q-layer.2": 3878.4773, "encoder_q-layer.3": 4030.9756, "encoder_q-layer.4": 4235.5249, "encoder_q-layer.5": 4394.8647, "encoder_q-layer.6": 5245.9199, "encoder_q-layer.7": 5876.7358, "encoder_q-layer.8": 6670.3955, "encoder_q-layer.9": 6183.96, "epoch": 0.44, "inbatch_neg_score": 2.0949, "inbatch_pos_score": 2.6992, "learning_rate": 2.586842105263158e-05, "loss": 3.4527, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.6256, "num_token_overlap": 11.6383, "num_token_query": 31.3123, "num_token_union": 65.0313, "num_word_context": 202.1111, "num_word_doc": 49.7305, "num_word_query": 23.2621, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9491.3521, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0957, "query_norm": 1.9034, "queue_k_norm": 1.9763, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3123, "sent_len_1": 66.6256, "sent_len_max_0": 127.96, "sent_len_max_1": 209.415, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101700 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.4482, "doc_norm": 1.9787, "encoder_q-embeddings": 6754.769, "encoder_q-layer.0": 4847.4043, "encoder_q-layer.1": 5250.4517, "encoder_q-layer.10": 7137.4282, "encoder_q-layer.11": 15065.3389, "encoder_q-layer.2": 5863.1519, "encoder_q-layer.3": 6131.0669, "encoder_q-layer.4": 6525.5654, "encoder_q-layer.5": 6318.5864, "encoder_q-layer.6": 6305.7573, "encoder_q-layer.7": 7347.0049, "encoder_q-layer.8": 7404.9224, "encoder_q-layer.9": 6657.4883, "epoch": 0.44, "inbatch_neg_score": 2.097, "inbatch_pos_score": 2.7031, "learning_rate": 2.5842105263157895e-05, "loss": 3.4482, "norm_diff": 0.0767, "norm_loss": 0.0, "num_token_doc": 66.8984, "num_token_overlap": 11.6872, "num_token_query": 31.3507, "num_token_union": 65.1937, "num_word_context": 202.3248, "num_word_doc": 49.9097, "num_word_query": 23.2907, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11132.4739, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0977, "query_norm": 1.9021, "queue_k_norm": 1.9771, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3507, "sent_len_1": 66.8984, "sent_len_max_0": 127.955, "sent_len_max_1": 208.285, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101800 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.4542, "doc_norm": 1.9767, "encoder_q-embeddings": 4889.2124, "encoder_q-layer.0": 3327.0962, "encoder_q-layer.1": 3668.6929, "encoder_q-layer.10": 6523.3926, "encoder_q-layer.11": 15749.2158, "encoder_q-layer.2": 3924.0642, "encoder_q-layer.3": 3979.4707, "encoder_q-layer.4": 4224.3735, "encoder_q-layer.5": 4411.0503, "encoder_q-layer.6": 5034.9927, "encoder_q-layer.7": 5548.2559, "encoder_q-layer.8": 6680.2529, "encoder_q-layer.9": 6146.6768, "epoch": 0.44, "inbatch_neg_score": 2.1005, "inbatch_pos_score": 2.6992, "learning_rate": 2.581578947368421e-05, "loss": 3.4542, "norm_diff": 0.0767, "norm_loss": 0.0, "num_token_doc": 66.8191, "num_token_overlap": 11.6985, "num_token_query": 31.3551, "num_token_union": 65.1344, "num_word_context": 202.338, "num_word_doc": 49.8711, "num_word_query": 23.2817, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9838.1791, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.0996, "query_norm": 1.9, "queue_k_norm": 1.9796, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3551, "sent_len_1": 66.8191, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.1238, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 101900 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.4772, "doc_norm": 1.9774, "encoder_q-embeddings": 10269.0449, "encoder_q-layer.0": 7143.418, "encoder_q-layer.1": 7781.3101, "encoder_q-layer.10": 13504.8662, "encoder_q-layer.11": 32350.7773, "encoder_q-layer.2": 8600.041, "encoder_q-layer.3": 8874.5479, "encoder_q-layer.4": 9501.7441, "encoder_q-layer.5": 9711.1465, "encoder_q-layer.6": 10872.6963, "encoder_q-layer.7": 12472.083, "encoder_q-layer.8": 14507.3145, "encoder_q-layer.9": 12718.7744, "epoch": 0.44, "inbatch_neg_score": 2.1015, "inbatch_pos_score": 2.6875, "learning_rate": 2.578947368421053e-05, "loss": 3.4772, "norm_diff": 0.0936, "norm_loss": 0.0, "num_token_doc": 66.7491, "num_token_overlap": 11.6456, "num_token_query": 31.3091, "num_token_union": 65.0885, "num_word_context": 202.2171, "num_word_doc": 49.8041, "num_word_query": 23.2414, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20552.0372, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1016, "query_norm": 1.8838, "queue_k_norm": 1.9796, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3091, "sent_len_1": 66.7491, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.85, "stdk": 0.0491, "stdq": 0.0423, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 102000 }, { "accuracy": 48.5352, "active_queue_size": 16384.0, "cl_loss": 3.4424, "doc_norm": 1.9782, "encoder_q-embeddings": 11189.3281, "encoder_q-layer.0": 7935.3218, "encoder_q-layer.1": 8266.9785, "encoder_q-layer.10": 13540.5879, "encoder_q-layer.11": 30482.3867, "encoder_q-layer.2": 9535.748, "encoder_q-layer.3": 10005.3174, "encoder_q-layer.4": 10976.0869, "encoder_q-layer.5": 10897.0029, "encoder_q-layer.6": 11732.1953, "encoder_q-layer.7": 12618.791, "encoder_q-layer.8": 14319.7109, "encoder_q-layer.9": 12088.1807, "epoch": 0.44, "inbatch_neg_score": 2.1049, "inbatch_pos_score": 2.6992, "learning_rate": 2.5763157894736843e-05, "loss": 3.4424, "norm_diff": 0.0777, "norm_loss": 0.0, "num_token_doc": 66.7565, "num_token_overlap": 11.6961, "num_token_query": 31.4145, "num_token_union": 65.0877, "num_word_context": 202.2166, "num_word_doc": 49.7913, "num_word_query": 23.3116, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20292.427, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1035, "query_norm": 1.9005, "queue_k_norm": 1.982, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4145, "sent_len_1": 66.7565, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.155, "stdk": 0.049, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 102100 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4558, "doc_norm": 1.9844, "encoder_q-embeddings": 10182.2051, "encoder_q-layer.0": 6739.3706, "encoder_q-layer.1": 7220.0474, "encoder_q-layer.10": 13839.7451, "encoder_q-layer.11": 30720.3555, "encoder_q-layer.2": 7928.9067, "encoder_q-layer.3": 8175.9526, "encoder_q-layer.4": 8868.1816, "encoder_q-layer.5": 8829.5205, "encoder_q-layer.6": 9882.1191, "encoder_q-layer.7": 11173.9268, "encoder_q-layer.8": 13311.1934, "encoder_q-layer.9": 12768.8525, "epoch": 0.44, "inbatch_neg_score": 2.1051, "inbatch_pos_score": 2.7051, "learning_rate": 2.5736842105263158e-05, "loss": 3.4558, "norm_diff": 0.0967, "norm_loss": 0.0, "num_token_doc": 66.8708, "num_token_overlap": 11.6759, "num_token_query": 31.3388, "num_token_union": 65.1499, "num_word_context": 202.4966, "num_word_doc": 49.8812, "num_word_query": 23.2643, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19377.7891, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1055, "query_norm": 1.8878, "queue_k_norm": 1.9818, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3388, "sent_len_1": 66.8708, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.7113, "stdk": 0.0493, "stdq": 0.0425, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 102200 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.4423, "doc_norm": 1.9846, "encoder_q-embeddings": 11486.0215, "encoder_q-layer.0": 7841.999, "encoder_q-layer.1": 8284.6035, "encoder_q-layer.10": 13689.8535, "encoder_q-layer.11": 31124.0117, "encoder_q-layer.2": 9506.4619, "encoder_q-layer.3": 9715.6562, "encoder_q-layer.4": 9936.5615, "encoder_q-layer.5": 9962.249, "encoder_q-layer.6": 11173.8105, "encoder_q-layer.7": 12088.0635, "encoder_q-layer.8": 14077.3115, "encoder_q-layer.9": 12562.8418, "epoch": 0.44, "inbatch_neg_score": 2.107, "inbatch_pos_score": 2.6953, "learning_rate": 2.5710526315789473e-05, "loss": 3.4423, "norm_diff": 0.096, "norm_loss": 0.0, "num_token_doc": 66.8323, "num_token_overlap": 11.7026, "num_token_query": 31.4492, "num_token_union": 65.1596, "num_word_context": 202.4233, "num_word_doc": 49.8551, "num_word_query": 23.3591, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20747.7591, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1074, "query_norm": 1.8886, "queue_k_norm": 1.9825, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4492, "sent_len_1": 66.8323, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.0662, "stdk": 0.0493, "stdq": 0.0425, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 102300 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4375, "doc_norm": 1.983, "encoder_q-embeddings": 14783.3242, "encoder_q-layer.0": 10995.5342, "encoder_q-layer.1": 12959.8096, "encoder_q-layer.10": 12833.6982, "encoder_q-layer.11": 30400.2305, "encoder_q-layer.2": 14215.7988, "encoder_q-layer.3": 15099.9238, "encoder_q-layer.4": 16822.5449, "encoder_q-layer.5": 16321.0, "encoder_q-layer.6": 15805.7148, "encoder_q-layer.7": 16419.4102, "encoder_q-layer.8": 16037.4971, "encoder_q-layer.9": 13204.6982, "epoch": 0.44, "inbatch_neg_score": 2.1115, "inbatch_pos_score": 2.7129, "learning_rate": 2.5684210526315788e-05, "loss": 3.4375, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.6893, "num_token_overlap": 11.7125, "num_token_query": 31.4045, "num_token_union": 65.0417, "num_word_context": 202.3552, "num_word_doc": 49.7733, "num_word_query": 23.3096, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24522.7408, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1113, "query_norm": 1.9036, "queue_k_norm": 1.9821, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4045, "sent_len_1": 66.6893, "sent_len_max_0": 127.995, "sent_len_max_1": 208.9775, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 102400 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.4493, "doc_norm": 1.9839, "encoder_q-embeddings": 11142.2559, "encoder_q-layer.0": 7636.9502, "encoder_q-layer.1": 8674.2471, "encoder_q-layer.10": 12909.335, "encoder_q-layer.11": 31825.0391, "encoder_q-layer.2": 9805.9824, "encoder_q-layer.3": 10385.4141, "encoder_q-layer.4": 11194.5605, "encoder_q-layer.5": 11920.6934, "encoder_q-layer.6": 13544.0049, "encoder_q-layer.7": 13366.2236, "encoder_q-layer.8": 14823.2207, "encoder_q-layer.9": 13158.9893, "epoch": 0.44, "inbatch_neg_score": 2.1111, "inbatch_pos_score": 2.7148, "learning_rate": 2.565789473684211e-05, "loss": 3.4493, "norm_diff": 0.0848, "norm_loss": 0.0, "num_token_doc": 66.7751, "num_token_overlap": 11.679, "num_token_query": 31.3414, "num_token_union": 65.1078, "num_word_context": 202.6102, "num_word_doc": 49.8306, "num_word_query": 23.2796, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21324.4118, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1113, "query_norm": 1.8991, "queue_k_norm": 1.9865, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3414, "sent_len_1": 66.7751, "sent_len_max_0": 127.9788, "sent_len_max_1": 206.3825, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 102500 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.4477, "doc_norm": 1.9871, "encoder_q-embeddings": 9827.0703, "encoder_q-layer.0": 6747.394, "encoder_q-layer.1": 7364.8555, "encoder_q-layer.10": 12722.8965, "encoder_q-layer.11": 31618.9922, "encoder_q-layer.2": 8098.937, "encoder_q-layer.3": 8173.877, "encoder_q-layer.4": 8535.8916, "encoder_q-layer.5": 8653.3516, "encoder_q-layer.6": 9369.9502, "encoder_q-layer.7": 11199.0078, "encoder_q-layer.8": 13650.9014, "encoder_q-layer.9": 12336.9053, "epoch": 0.45, "inbatch_neg_score": 2.1149, "inbatch_pos_score": 2.7266, "learning_rate": 2.563157894736842e-05, "loss": 3.4477, "norm_diff": 0.0869, "norm_loss": 0.0, "num_token_doc": 66.6805, "num_token_overlap": 11.6741, "num_token_query": 31.3746, "num_token_union": 65.0813, "num_word_context": 202.3298, "num_word_doc": 49.7663, "num_word_query": 23.2994, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19462.9106, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1172, "query_norm": 1.9002, "queue_k_norm": 1.9856, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3746, "sent_len_1": 66.6805, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.3175, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 102600 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.4249, "doc_norm": 1.9909, "encoder_q-embeddings": 11462.8154, "encoder_q-layer.0": 8109.9111, "encoder_q-layer.1": 8468.749, "encoder_q-layer.10": 13803.0342, "encoder_q-layer.11": 31638.8555, "encoder_q-layer.2": 10112.1631, "encoder_q-layer.3": 10655.0791, "encoder_q-layer.4": 11367.4824, "encoder_q-layer.5": 12020.2979, "encoder_q-layer.6": 13070.1416, "encoder_q-layer.7": 13446.7012, "encoder_q-layer.8": 15541.8281, "encoder_q-layer.9": 12625.7842, "epoch": 0.45, "inbatch_neg_score": 2.1176, "inbatch_pos_score": 2.7285, "learning_rate": 2.5605263157894737e-05, "loss": 3.4249, "norm_diff": 0.0857, "norm_loss": 0.0, "num_token_doc": 66.7611, "num_token_overlap": 11.6753, "num_token_query": 31.4285, "num_token_union": 65.1637, "num_word_context": 202.4854, "num_word_doc": 49.8134, "num_word_query": 23.3436, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21260.0407, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1172, "query_norm": 1.9052, "queue_k_norm": 1.9876, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4285, "sent_len_1": 66.7611, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.995, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 102700 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.4409, "doc_norm": 1.9918, "encoder_q-embeddings": 9588.1016, "encoder_q-layer.0": 6894.6333, "encoder_q-layer.1": 7463.5703, "encoder_q-layer.10": 13108.9414, "encoder_q-layer.11": 29159.5098, "encoder_q-layer.2": 8183.125, "encoder_q-layer.3": 8613.7627, "encoder_q-layer.4": 8842.2441, "encoder_q-layer.5": 9146.2559, "encoder_q-layer.6": 10381.6016, "encoder_q-layer.7": 12035.7812, "encoder_q-layer.8": 14125.3555, "encoder_q-layer.9": 12859.3467, "epoch": 0.45, "inbatch_neg_score": 2.1247, "inbatch_pos_score": 2.7383, "learning_rate": 2.557894736842105e-05, "loss": 3.4409, "norm_diff": 0.0874, "norm_loss": 0.0, "num_token_doc": 66.8742, "num_token_overlap": 11.6965, "num_token_query": 31.3857, "num_token_union": 65.1795, "num_word_context": 202.438, "num_word_doc": 49.9113, "num_word_query": 23.3039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19053.4672, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.125, "query_norm": 1.9044, "queue_k_norm": 1.9871, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3857, "sent_len_1": 66.8742, "sent_len_max_0": 127.9375, "sent_len_max_1": 208.2287, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 102800 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.4481, "doc_norm": 1.983, "encoder_q-embeddings": 10658.7783, "encoder_q-layer.0": 7287.8579, "encoder_q-layer.1": 7973.4126, "encoder_q-layer.10": 12483.0312, "encoder_q-layer.11": 30616.9355, "encoder_q-layer.2": 8636.6855, "encoder_q-layer.3": 9302.1963, "encoder_q-layer.4": 9601.4004, "encoder_q-layer.5": 10116.0439, "encoder_q-layer.6": 10578.375, "encoder_q-layer.7": 11457.9473, "encoder_q-layer.8": 14856.2666, "encoder_q-layer.9": 13021.5098, "epoch": 0.45, "inbatch_neg_score": 2.1301, "inbatch_pos_score": 2.7305, "learning_rate": 2.5552631578947366e-05, "loss": 3.4481, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.7437, "num_token_overlap": 11.6669, "num_token_query": 31.3183, "num_token_union": 65.0571, "num_word_context": 202.1743, "num_word_doc": 49.7757, "num_word_query": 23.2565, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19920.2262, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1309, "query_norm": 1.9075, "queue_k_norm": 1.9897, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3183, "sent_len_1": 66.7437, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.13, "stdk": 0.0489, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 102900 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4406, "doc_norm": 1.9884, "encoder_q-embeddings": 10230.4854, "encoder_q-layer.0": 7218.1338, "encoder_q-layer.1": 7774.9653, "encoder_q-layer.10": 13438.3105, "encoder_q-layer.11": 30926.5762, "encoder_q-layer.2": 8564.6768, "encoder_q-layer.3": 8916.5615, "encoder_q-layer.4": 9732.5078, "encoder_q-layer.5": 9942.8506, "encoder_q-layer.6": 10609.8408, "encoder_q-layer.7": 12107.4785, "encoder_q-layer.8": 14290.0928, "encoder_q-layer.9": 13021.9482, "epoch": 0.45, "inbatch_neg_score": 2.1328, "inbatch_pos_score": 2.7559, "learning_rate": 2.5526315789473688e-05, "loss": 3.4406, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.6968, "num_token_overlap": 11.673, "num_token_query": 31.3585, "num_token_union": 65.0776, "num_word_context": 202.2969, "num_word_doc": 49.8021, "num_word_query": 23.2959, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19957.6452, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1328, "query_norm": 1.918, "queue_k_norm": 1.9898, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3585, "sent_len_1": 66.6968, "sent_len_max_0": 127.96, "sent_len_max_1": 205.7175, "stdk": 0.0491, "stdq": 0.0438, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 103000 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.4339, "doc_norm": 1.9926, "encoder_q-embeddings": 9681.9473, "encoder_q-layer.0": 6581.2202, "encoder_q-layer.1": 7056.9868, "encoder_q-layer.10": 12384.9102, "encoder_q-layer.11": 29085.0801, "encoder_q-layer.2": 8011.0278, "encoder_q-layer.3": 8112.8555, "encoder_q-layer.4": 8433.5645, "encoder_q-layer.5": 8772.4395, "encoder_q-layer.6": 9721.6846, "encoder_q-layer.7": 11005.7969, "encoder_q-layer.8": 13138.5615, "encoder_q-layer.9": 12199.1348, "epoch": 0.45, "inbatch_neg_score": 2.1375, "inbatch_pos_score": 2.7617, "learning_rate": 2.5500000000000003e-05, "loss": 3.4339, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 66.9517, "num_token_overlap": 11.7168, "num_token_query": 31.3922, "num_token_union": 65.199, "num_word_context": 202.207, "num_word_doc": 49.9684, "num_word_query": 23.3184, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18404.1289, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1387, "query_norm": 1.9099, "queue_k_norm": 1.9914, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3922, "sent_len_1": 66.9517, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.8887, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 103100 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.445, "doc_norm": 1.9921, "encoder_q-embeddings": 10762.2363, "encoder_q-layer.0": 7269.4336, "encoder_q-layer.1": 7847.4717, "encoder_q-layer.10": 12326.6553, "encoder_q-layer.11": 28763.3945, "encoder_q-layer.2": 9100.1992, "encoder_q-layer.3": 9450.8066, "encoder_q-layer.4": 10127.3965, "encoder_q-layer.5": 10295.2324, "encoder_q-layer.6": 10987.9678, "encoder_q-layer.7": 11783.5605, "encoder_q-layer.8": 13729.5645, "encoder_q-layer.9": 12091.501, "epoch": 0.45, "inbatch_neg_score": 2.1398, "inbatch_pos_score": 2.7539, "learning_rate": 2.5473684210526315e-05, "loss": 3.445, "norm_diff": 0.0789, "norm_loss": 0.0, "num_token_doc": 66.7574, "num_token_overlap": 11.6681, "num_token_query": 31.2448, "num_token_union": 65.0478, "num_word_context": 202.1995, "num_word_doc": 49.8344, "num_word_query": 23.1959, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19459.4792, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1406, "query_norm": 1.9132, "queue_k_norm": 1.9915, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2448, "sent_len_1": 66.7574, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.6387, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 103200 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4491, "doc_norm": 1.9927, "encoder_q-embeddings": 9632.6016, "encoder_q-layer.0": 6738.877, "encoder_q-layer.1": 7378.854, "encoder_q-layer.10": 12492.1035, "encoder_q-layer.11": 29478.7305, "encoder_q-layer.2": 8104.3525, "encoder_q-layer.3": 8597.0537, "encoder_q-layer.4": 8898.9453, "encoder_q-layer.5": 8728.834, "encoder_q-layer.6": 10199.1875, "encoder_q-layer.7": 11856.1357, "encoder_q-layer.8": 13306.5322, "encoder_q-layer.9": 12239.4414, "epoch": 0.45, "inbatch_neg_score": 2.145, "inbatch_pos_score": 2.752, "learning_rate": 2.544736842105263e-05, "loss": 3.4491, "norm_diff": 0.0873, "norm_loss": 0.0, "num_token_doc": 66.7862, "num_token_overlap": 11.6612, "num_token_query": 31.3349, "num_token_union": 65.1006, "num_word_context": 202.5073, "num_word_doc": 49.8215, "num_word_query": 23.2809, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19269.9875, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1465, "query_norm": 1.9054, "queue_k_norm": 1.9913, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3349, "sent_len_1": 66.7862, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.8862, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 103300 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4315, "doc_norm": 1.9964, "encoder_q-embeddings": 9770.3613, "encoder_q-layer.0": 6783.5195, "encoder_q-layer.1": 7375.9785, "encoder_q-layer.10": 13714.6807, "encoder_q-layer.11": 31319.4277, "encoder_q-layer.2": 8134.6162, "encoder_q-layer.3": 8370.1748, "encoder_q-layer.4": 9160.0225, "encoder_q-layer.5": 9414.8525, "encoder_q-layer.6": 10625.8594, "encoder_q-layer.7": 12125.5293, "encoder_q-layer.8": 14563.6309, "encoder_q-layer.9": 13327.627, "epoch": 0.45, "inbatch_neg_score": 2.1528, "inbatch_pos_score": 2.7656, "learning_rate": 2.542105263157895e-05, "loss": 3.4315, "norm_diff": 0.0855, "norm_loss": 0.0, "num_token_doc": 67.0187, "num_token_overlap": 11.7105, "num_token_query": 31.3795, "num_token_union": 65.2341, "num_word_context": 202.2535, "num_word_doc": 50.0129, "num_word_query": 23.3058, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20207.0208, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1523, "query_norm": 1.9109, "queue_k_norm": 1.9949, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3795, "sent_len_1": 67.0187, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.4988, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 103400 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4517, "doc_norm": 1.9972, "encoder_q-embeddings": 9991.3291, "encoder_q-layer.0": 7017.7959, "encoder_q-layer.1": 7545.0845, "encoder_q-layer.10": 13041.2578, "encoder_q-layer.11": 31673.3711, "encoder_q-layer.2": 8695.2246, "encoder_q-layer.3": 8803.0889, "encoder_q-layer.4": 9149.6553, "encoder_q-layer.5": 9210.1904, "encoder_q-layer.6": 10229.5625, "encoder_q-layer.7": 11789.6035, "encoder_q-layer.8": 14421.5342, "encoder_q-layer.9": 12639.9287, "epoch": 0.45, "inbatch_neg_score": 2.1551, "inbatch_pos_score": 2.7422, "learning_rate": 2.5394736842105266e-05, "loss": 3.4517, "norm_diff": 0.0917, "norm_loss": 0.0, "num_token_doc": 66.8559, "num_token_overlap": 11.6553, "num_token_query": 31.32, "num_token_union": 65.1205, "num_word_context": 202.4359, "num_word_doc": 49.8628, "num_word_query": 23.2583, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20396.7855, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1543, "query_norm": 1.9055, "queue_k_norm": 1.9943, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.32, "sent_len_1": 66.8559, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.2962, "stdk": 0.0493, "stdq": 0.0426, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 103500 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4201, "doc_norm": 1.9954, "encoder_q-embeddings": 10039.4033, "encoder_q-layer.0": 6868.0762, "encoder_q-layer.1": 7627.3916, "encoder_q-layer.10": 13352.5596, "encoder_q-layer.11": 29864.9297, "encoder_q-layer.2": 8591.5928, "encoder_q-layer.3": 8576.5322, "encoder_q-layer.4": 9152.3936, "encoder_q-layer.5": 9383.377, "encoder_q-layer.6": 10171.1094, "encoder_q-layer.7": 12169.791, "encoder_q-layer.8": 13980.502, "encoder_q-layer.9": 12745.1309, "epoch": 0.45, "inbatch_neg_score": 2.1628, "inbatch_pos_score": 2.7773, "learning_rate": 2.536842105263158e-05, "loss": 3.4201, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.6924, "num_token_overlap": 11.7218, "num_token_query": 31.4205, "num_token_union": 65.0626, "num_word_context": 202.0416, "num_word_doc": 49.7824, "num_word_query": 23.3322, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19457.5562, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1621, "query_norm": 1.9199, "queue_k_norm": 1.9968, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4205, "sent_len_1": 66.6924, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.0538, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 103600 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.4551, "doc_norm": 1.9958, "encoder_q-embeddings": 11097.4385, "encoder_q-layer.0": 7493.4536, "encoder_q-layer.1": 7867.6807, "encoder_q-layer.10": 12832.626, "encoder_q-layer.11": 29320.7168, "encoder_q-layer.2": 8940.3926, "encoder_q-layer.3": 9404.8252, "encoder_q-layer.4": 9923.8652, "encoder_q-layer.5": 10765.084, "encoder_q-layer.6": 11708.0537, "encoder_q-layer.7": 13233.8242, "encoder_q-layer.8": 14370.4033, "encoder_q-layer.9": 12939.2998, "epoch": 0.45, "inbatch_neg_score": 2.1615, "inbatch_pos_score": 2.75, "learning_rate": 2.5342105263157896e-05, "loss": 3.4551, "norm_diff": 0.0903, "norm_loss": 0.0, "num_token_doc": 66.9978, "num_token_overlap": 11.6588, "num_token_query": 31.3028, "num_token_union": 65.1996, "num_word_context": 202.4686, "num_word_doc": 49.966, "num_word_query": 23.2522, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20130.8952, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1602, "query_norm": 1.9055, "queue_k_norm": 1.9982, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3028, "sent_len_1": 66.9978, "sent_len_max_0": 128.0, "sent_len_max_1": 209.8862, "stdk": 0.0491, "stdq": 0.0426, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 103700 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4387, "doc_norm": 1.9985, "encoder_q-embeddings": 10613.8438, "encoder_q-layer.0": 7275.4526, "encoder_q-layer.1": 7902.3198, "encoder_q-layer.10": 12613.4521, "encoder_q-layer.11": 30517.5117, "encoder_q-layer.2": 8762.3242, "encoder_q-layer.3": 9052.7861, "encoder_q-layer.4": 9550.9922, "encoder_q-layer.5": 9547.4277, "encoder_q-layer.6": 10496.7314, "encoder_q-layer.7": 11825.5371, "encoder_q-layer.8": 13712.9072, "encoder_q-layer.9": 12532.1016, "epoch": 0.45, "inbatch_neg_score": 2.1659, "inbatch_pos_score": 2.7578, "learning_rate": 2.5315789473684208e-05, "loss": 3.4387, "norm_diff": 0.0904, "norm_loss": 0.0, "num_token_doc": 66.8501, "num_token_overlap": 11.6904, "num_token_query": 31.4129, "num_token_union": 65.1807, "num_word_context": 202.5337, "num_word_doc": 49.8705, "num_word_query": 23.333, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19902.9634, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.166, "query_norm": 1.9081, "queue_k_norm": 1.9974, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4129, "sent_len_1": 66.8501, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.3837, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 103800 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.4488, "doc_norm": 1.998, "encoder_q-embeddings": 10395.127, "encoder_q-layer.0": 6967.9692, "encoder_q-layer.1": 7647.8286, "encoder_q-layer.10": 13599.9912, "encoder_q-layer.11": 30475.043, "encoder_q-layer.2": 8436.9473, "encoder_q-layer.3": 8445.9775, "encoder_q-layer.4": 9219.0088, "encoder_q-layer.5": 9381.2129, "encoder_q-layer.6": 10493.6641, "encoder_q-layer.7": 11667.3789, "encoder_q-layer.8": 13792.3584, "encoder_q-layer.9": 12876.4814, "epoch": 0.45, "inbatch_neg_score": 2.167, "inbatch_pos_score": 2.7578, "learning_rate": 2.528947368421053e-05, "loss": 3.4488, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.5873, "num_token_overlap": 11.6484, "num_token_query": 31.4156, "num_token_union": 65.0369, "num_word_context": 202.2474, "num_word_doc": 49.6654, "num_word_query": 23.3349, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19804.5458, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.166, "query_norm": 1.9187, "queue_k_norm": 1.9981, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4156, "sent_len_1": 66.5873, "sent_len_max_0": 127.99, "sent_len_max_1": 208.9225, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 103900 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.456, "doc_norm": 1.9986, "encoder_q-embeddings": 21105.9512, "encoder_q-layer.0": 14491.8926, "encoder_q-layer.1": 16038.8477, "encoder_q-layer.10": 27222.6797, "encoder_q-layer.11": 58344.3672, "encoder_q-layer.2": 17297.9023, "encoder_q-layer.3": 17474.2324, "encoder_q-layer.4": 18112.6035, "encoder_q-layer.5": 18742.2441, "encoder_q-layer.6": 20506.2715, "encoder_q-layer.7": 21856.4277, "encoder_q-layer.8": 25980.3496, "encoder_q-layer.9": 24561.4102, "epoch": 0.45, "inbatch_neg_score": 2.1697, "inbatch_pos_score": 2.7793, "learning_rate": 2.5263157894736845e-05, "loss": 3.456, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.888, "num_token_overlap": 11.6898, "num_token_query": 31.4082, "num_token_union": 65.2207, "num_word_context": 202.4497, "num_word_doc": 49.916, "num_word_query": 23.3284, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38810.143, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.1699, "query_norm": 1.9172, "queue_k_norm": 2.0005, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4082, "sent_len_1": 66.888, "sent_len_max_0": 127.99, "sent_len_max_1": 206.9013, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104000 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4456, "doc_norm": 2.0021, "encoder_q-embeddings": 9604.5479, "encoder_q-layer.0": 6842.2158, "encoder_q-layer.1": 7438.5078, "encoder_q-layer.10": 13080.3105, "encoder_q-layer.11": 30043.3027, "encoder_q-layer.2": 8408.1279, "encoder_q-layer.3": 8540.8467, "encoder_q-layer.4": 8942.2578, "encoder_q-layer.5": 9334.7607, "encoder_q-layer.6": 9946.4414, "encoder_q-layer.7": 11160.7959, "encoder_q-layer.8": 12733.8789, "encoder_q-layer.9": 12791.0322, "epoch": 0.45, "inbatch_neg_score": 2.1723, "inbatch_pos_score": 2.7676, "learning_rate": 2.523684210526316e-05, "loss": 3.4456, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.7112, "num_token_overlap": 11.6603, "num_token_query": 31.3816, "num_token_union": 65.0971, "num_word_context": 202.2328, "num_word_doc": 49.7709, "num_word_query": 23.3076, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19135.1415, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1719, "query_norm": 1.9172, "queue_k_norm": 2.0008, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3816, "sent_len_1": 66.7112, "sent_len_max_0": 127.9562, "sent_len_max_1": 208.8762, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104100 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.452, "doc_norm": 2.001, "encoder_q-embeddings": 9979.1025, "encoder_q-layer.0": 6800.1318, "encoder_q-layer.1": 7501.4106, "encoder_q-layer.10": 12851.5508, "encoder_q-layer.11": 31263.9219, "encoder_q-layer.2": 8246.6172, "encoder_q-layer.3": 8540.4297, "encoder_q-layer.4": 9011.3125, "encoder_q-layer.5": 9305.0684, "encoder_q-layer.6": 10694.8027, "encoder_q-layer.7": 11929.8828, "encoder_q-layer.8": 14141.623, "encoder_q-layer.9": 12740.9131, "epoch": 0.45, "inbatch_neg_score": 2.1733, "inbatch_pos_score": 2.7637, "learning_rate": 2.5210526315789475e-05, "loss": 3.452, "norm_diff": 0.092, "norm_loss": 0.0, "num_token_doc": 66.8273, "num_token_overlap": 11.7271, "num_token_query": 31.4341, "num_token_union": 65.1417, "num_word_context": 202.4977, "num_word_doc": 49.8742, "num_word_query": 23.3592, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20061.0715, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1719, "query_norm": 1.909, "queue_k_norm": 2.0012, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4341, "sent_len_1": 66.8273, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.5925, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104200 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.4494, "doc_norm": 2.0031, "encoder_q-embeddings": 11178.4229, "encoder_q-layer.0": 8151.9443, "encoder_q-layer.1": 8421.6494, "encoder_q-layer.10": 12981.4541, "encoder_q-layer.11": 32028.0137, "encoder_q-layer.2": 9398.2949, "encoder_q-layer.3": 10228.2021, "encoder_q-layer.4": 10429.2158, "encoder_q-layer.5": 10649.626, "encoder_q-layer.6": 12285.916, "encoder_q-layer.7": 12670.4355, "encoder_q-layer.8": 14058.2773, "encoder_q-layer.9": 12420.9414, "epoch": 0.45, "inbatch_neg_score": 2.176, "inbatch_pos_score": 2.7695, "learning_rate": 2.5184210526315786e-05, "loss": 3.4494, "norm_diff": 0.0905, "norm_loss": 0.0, "num_token_doc": 66.8835, "num_token_overlap": 11.6787, "num_token_query": 31.326, "num_token_union": 65.1499, "num_word_context": 202.1981, "num_word_doc": 49.8824, "num_word_query": 23.2584, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21230.7859, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1758, "query_norm": 1.9125, "queue_k_norm": 2.0035, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.326, "sent_len_1": 66.8835, "sent_len_max_0": 127.985, "sent_len_max_1": 209.0425, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104300 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4411, "doc_norm": 1.9987, "encoder_q-embeddings": 10488.3672, "encoder_q-layer.0": 7296.4111, "encoder_q-layer.1": 8061.8149, "encoder_q-layer.10": 12183.0254, "encoder_q-layer.11": 29189.8398, "encoder_q-layer.2": 9155.1582, "encoder_q-layer.3": 9702.9365, "encoder_q-layer.4": 10723.7344, "encoder_q-layer.5": 10560.1895, "encoder_q-layer.6": 11551.9121, "encoder_q-layer.7": 13405.0898, "encoder_q-layer.8": 14545.043, "encoder_q-layer.9": 12231.5039, "epoch": 0.45, "inbatch_neg_score": 2.1741, "inbatch_pos_score": 2.7812, "learning_rate": 2.5157894736842108e-05, "loss": 3.4411, "norm_diff": 0.0884, "norm_loss": 0.0, "num_token_doc": 66.8321, "num_token_overlap": 11.6775, "num_token_query": 31.3743, "num_token_union": 65.1395, "num_word_context": 202.4955, "num_word_doc": 49.8471, "num_word_query": 23.3092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20111.8304, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1738, "query_norm": 1.9102, "queue_k_norm": 2.0049, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3743, "sent_len_1": 66.8321, "sent_len_max_0": 127.97, "sent_len_max_1": 209.9725, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 104400 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4392, "doc_norm": 2.0011, "encoder_q-embeddings": 11136.7861, "encoder_q-layer.0": 7544.7544, "encoder_q-layer.1": 8286.8203, "encoder_q-layer.10": 13328.9336, "encoder_q-layer.11": 30822.8711, "encoder_q-layer.2": 9339.9951, "encoder_q-layer.3": 9833.2891, "encoder_q-layer.4": 9934.0615, "encoder_q-layer.5": 9863.0674, "encoder_q-layer.6": 10837.3379, "encoder_q-layer.7": 11854.6855, "encoder_q-layer.8": 13541.7754, "encoder_q-layer.9": 12647.2969, "epoch": 0.45, "inbatch_neg_score": 2.1756, "inbatch_pos_score": 2.7754, "learning_rate": 2.5131578947368423e-05, "loss": 3.4392, "norm_diff": 0.0896, "norm_loss": 0.0, "num_token_doc": 66.8382, "num_token_overlap": 11.6332, "num_token_query": 31.314, "num_token_union": 65.1693, "num_word_context": 202.3722, "num_word_doc": 49.9003, "num_word_query": 23.2434, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20219.875, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1758, "query_norm": 1.9115, "queue_k_norm": 2.0026, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.314, "sent_len_1": 66.8382, "sent_len_max_0": 127.9287, "sent_len_max_1": 207.9325, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 104500 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.4558, "doc_norm": 2.0054, "encoder_q-embeddings": 10351.3682, "encoder_q-layer.0": 7064.0947, "encoder_q-layer.1": 7618.8418, "encoder_q-layer.10": 13645.8887, "encoder_q-layer.11": 31058.2949, "encoder_q-layer.2": 8627.3096, "encoder_q-layer.3": 8518.3721, "encoder_q-layer.4": 9001.6094, "encoder_q-layer.5": 9160.2432, "encoder_q-layer.6": 10234.2676, "encoder_q-layer.7": 11395.8799, "encoder_q-layer.8": 13730.5215, "encoder_q-layer.9": 12382.1182, "epoch": 0.45, "inbatch_neg_score": 2.1786, "inbatch_pos_score": 2.7812, "learning_rate": 2.5105263157894738e-05, "loss": 3.4558, "norm_diff": 0.0917, "norm_loss": 0.0, "num_token_doc": 66.7773, "num_token_overlap": 11.6454, "num_token_query": 31.2808, "num_token_union": 65.0935, "num_word_context": 202.3539, "num_word_doc": 49.8411, "num_word_query": 23.2354, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19967.0965, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1777, "query_norm": 1.9137, "queue_k_norm": 2.0036, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2808, "sent_len_1": 66.7773, "sent_len_max_0": 127.9988, "sent_len_max_1": 210.79, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 104600 }, { "accuracy": 48.2422, "active_queue_size": 16384.0, "cl_loss": 3.4463, "doc_norm": 2.0057, "encoder_q-embeddings": 9887.625, "encoder_q-layer.0": 6844.0117, "encoder_q-layer.1": 7311.8608, "encoder_q-layer.10": 13427.5625, "encoder_q-layer.11": 30752.7129, "encoder_q-layer.2": 8191.2124, "encoder_q-layer.3": 8396.2461, "encoder_q-layer.4": 8998.0352, "encoder_q-layer.5": 9305.3213, "encoder_q-layer.6": 10441.8672, "encoder_q-layer.7": 11748.4883, "encoder_q-layer.8": 13843.7461, "encoder_q-layer.9": 13210.7109, "epoch": 0.45, "inbatch_neg_score": 2.1804, "inbatch_pos_score": 2.7852, "learning_rate": 2.5078947368421053e-05, "loss": 3.4463, "norm_diff": 0.0871, "norm_loss": 0.0, "num_token_doc": 66.7246, "num_token_overlap": 11.6517, "num_token_query": 31.3299, "num_token_union": 65.0637, "num_word_context": 202.3401, "num_word_doc": 49.7843, "num_word_query": 23.2653, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19822.2787, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1797, "query_norm": 1.9186, "queue_k_norm": 2.0051, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3299, "sent_len_1": 66.7246, "sent_len_max_0": 127.99, "sent_len_max_1": 208.7962, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4457, "doc_norm": 2.0047, "encoder_q-embeddings": 9704.001, "encoder_q-layer.0": 6551.936, "encoder_q-layer.1": 6806.6699, "encoder_q-layer.10": 14851.3604, "encoder_q-layer.11": 32899.2891, "encoder_q-layer.2": 7666.8208, "encoder_q-layer.3": 7929.457, "encoder_q-layer.4": 8728.6123, "encoder_q-layer.5": 9018.6904, "encoder_q-layer.6": 9738.0752, "encoder_q-layer.7": 11509.7422, "encoder_q-layer.8": 14366.1162, "encoder_q-layer.9": 13496.7607, "epoch": 0.45, "inbatch_neg_score": 2.1844, "inbatch_pos_score": 2.7852, "learning_rate": 2.505263157894737e-05, "loss": 3.4457, "norm_diff": 0.0986, "norm_loss": 0.0, "num_token_doc": 66.9829, "num_token_overlap": 11.7035, "num_token_query": 31.4205, "num_token_union": 65.2496, "num_word_context": 202.51, "num_word_doc": 49.9585, "num_word_query": 23.3477, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19964.8704, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.1836, "query_norm": 1.9061, "queue_k_norm": 2.0057, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4205, "sent_len_1": 66.9829, "sent_len_max_0": 128.0, "sent_len_max_1": 209.8862, "stdk": 0.0492, "stdq": 0.0426, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104800 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.4447, "doc_norm": 2.0092, "encoder_q-embeddings": 4895.1704, "encoder_q-layer.0": 3444.5586, "encoder_q-layer.1": 3619.8521, "encoder_q-layer.10": 6964.1528, "encoder_q-layer.11": 15006.0332, "encoder_q-layer.2": 4084.4663, "encoder_q-layer.3": 4274.8877, "encoder_q-layer.4": 4648.5811, "encoder_q-layer.5": 4877.9883, "encoder_q-layer.6": 5211.7363, "encoder_q-layer.7": 5756.4087, "encoder_q-layer.8": 6737.7603, "encoder_q-layer.9": 6131.4028, "epoch": 0.46, "inbatch_neg_score": 2.1859, "inbatch_pos_score": 2.8066, "learning_rate": 2.5026315789473686e-05, "loss": 3.4447, "norm_diff": 0.087, "norm_loss": 0.0, "num_token_doc": 66.7255, "num_token_overlap": 11.667, "num_token_query": 31.4138, "num_token_union": 65.1121, "num_word_context": 202.0888, "num_word_doc": 49.7756, "num_word_query": 23.3398, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9715.7724, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.1855, "query_norm": 1.9223, "queue_k_norm": 2.0071, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4138, "sent_len_1": 66.7255, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.8313, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 104900 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4407, "doc_norm": 2.0034, "encoder_q-embeddings": 4903.9722, "encoder_q-layer.0": 3357.4346, "encoder_q-layer.1": 3544.7751, "encoder_q-layer.10": 6283.9189, "encoder_q-layer.11": 14671.1152, "encoder_q-layer.2": 3974.7456, "encoder_q-layer.3": 4080.7656, "encoder_q-layer.4": 4378.3379, "encoder_q-layer.5": 4498.1821, "encoder_q-layer.6": 4898.9219, "encoder_q-layer.7": 6229.4644, "encoder_q-layer.8": 7431.5649, "encoder_q-layer.9": 6348.4736, "epoch": 0.46, "inbatch_neg_score": 2.188, "inbatch_pos_score": 2.7969, "learning_rate": 2.5e-05, "loss": 3.4407, "norm_diff": 0.09, "norm_loss": 0.0, "num_token_doc": 66.7498, "num_token_overlap": 11.6838, "num_token_query": 31.3306, "num_token_union": 65.0863, "num_word_context": 202.2933, "num_word_doc": 49.7989, "num_word_query": 23.2717, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9654.2666, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.1875, "query_norm": 1.9134, "queue_k_norm": 2.0062, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3306, "sent_len_1": 66.7498, "sent_len_max_0": 127.98, "sent_len_max_1": 209.895, "stdk": 0.049, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 105000 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4439, "doc_norm": 2.0085, "encoder_q-embeddings": 4961.2656, "encoder_q-layer.0": 3333.1772, "encoder_q-layer.1": 3551.959, "encoder_q-layer.10": 6464.5283, "encoder_q-layer.11": 15005.2188, "encoder_q-layer.2": 3878.2712, "encoder_q-layer.3": 4066.4739, "encoder_q-layer.4": 4293.7812, "encoder_q-layer.5": 4469.7378, "encoder_q-layer.6": 5004.9551, "encoder_q-layer.7": 5686.4634, "encoder_q-layer.8": 6822.3647, "encoder_q-layer.9": 6160.0288, "epoch": 0.46, "inbatch_neg_score": 2.19, "inbatch_pos_score": 2.8027, "learning_rate": 2.4973684210526316e-05, "loss": 3.4439, "norm_diff": 0.0863, "norm_loss": 0.0, "num_token_doc": 66.7789, "num_token_overlap": 11.716, "num_token_query": 31.433, "num_token_union": 65.111, "num_word_context": 202.4498, "num_word_doc": 49.8051, "num_word_query": 23.3349, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9719.8958, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.1914, "query_norm": 1.9222, "queue_k_norm": 2.0078, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.433, "sent_len_1": 66.7789, "sent_len_max_0": 127.995, "sent_len_max_1": 210.2375, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 105100 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.4558, "doc_norm": 2.0102, "encoder_q-embeddings": 5755.9258, "encoder_q-layer.0": 4021.2224, "encoder_q-layer.1": 4481.438, "encoder_q-layer.10": 6118.73, "encoder_q-layer.11": 14876.4912, "encoder_q-layer.2": 5020.104, "encoder_q-layer.3": 5167.7158, "encoder_q-layer.4": 5257.5527, "encoder_q-layer.5": 5334.2148, "encoder_q-layer.6": 6050.1938, "encoder_q-layer.7": 6426.5815, "encoder_q-layer.8": 7013.1802, "encoder_q-layer.9": 6075.4653, "epoch": 0.46, "inbatch_neg_score": 2.194, "inbatch_pos_score": 2.7949, "learning_rate": 2.4947368421052635e-05, "loss": 3.4558, "norm_diff": 0.09, "norm_loss": 0.0, "num_token_doc": 66.8284, "num_token_overlap": 11.7054, "num_token_query": 31.4329, "num_token_union": 65.1534, "num_word_context": 202.3893, "num_word_doc": 49.853, "num_word_query": 23.3495, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10290.9035, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.1953, "query_norm": 1.9202, "queue_k_norm": 2.0079, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4329, "sent_len_1": 66.8284, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.4688, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 105200 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4328, "doc_norm": 2.0126, "encoder_q-embeddings": 5031.5288, "encoder_q-layer.0": 3463.7649, "encoder_q-layer.1": 3740.7002, "encoder_q-layer.10": 7167.7139, "encoder_q-layer.11": 15810.5195, "encoder_q-layer.2": 4125.8467, "encoder_q-layer.3": 4313.9922, "encoder_q-layer.4": 4696.8955, "encoder_q-layer.5": 5001.6499, "encoder_q-layer.6": 5181.562, "encoder_q-layer.7": 5863.8247, "encoder_q-layer.8": 7141.1455, "encoder_q-layer.9": 6467.8418, "epoch": 0.46, "inbatch_neg_score": 2.1972, "inbatch_pos_score": 2.8105, "learning_rate": 2.4921052631578946e-05, "loss": 3.4328, "norm_diff": 0.0811, "norm_loss": 0.0, "num_token_doc": 66.8198, "num_token_overlap": 11.7082, "num_token_query": 31.3896, "num_token_union": 65.1555, "num_word_context": 202.3965, "num_word_doc": 49.8595, "num_word_query": 23.3223, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10028.2639, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.1973, "query_norm": 1.9315, "queue_k_norm": 2.0086, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3896, "sent_len_1": 66.8198, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.6113, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 105300 }, { "accuracy": 48.4863, "active_queue_size": 16384.0, "cl_loss": 3.4471, "doc_norm": 2.0085, "encoder_q-embeddings": 5435.5205, "encoder_q-layer.0": 3828.7061, "encoder_q-layer.1": 4157.8325, "encoder_q-layer.10": 6848.1494, "encoder_q-layer.11": 15057.1182, "encoder_q-layer.2": 4664.7231, "encoder_q-layer.3": 4673.2124, "encoder_q-layer.4": 5217.3877, "encoder_q-layer.5": 5245.8354, "encoder_q-layer.6": 5587.1069, "encoder_q-layer.7": 6158.2935, "encoder_q-layer.8": 7076.0864, "encoder_q-layer.9": 6297.6357, "epoch": 0.46, "inbatch_neg_score": 2.2029, "inbatch_pos_score": 2.7891, "learning_rate": 2.4894736842105264e-05, "loss": 3.4471, "norm_diff": 0.0888, "norm_loss": 0.0, "num_token_doc": 66.6757, "num_token_overlap": 11.6597, "num_token_query": 31.3357, "num_token_union": 65.0538, "num_word_context": 202.2938, "num_word_doc": 49.7519, "num_word_query": 23.2814, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10080.006, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2031, "query_norm": 1.9196, "queue_k_norm": 2.0104, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3357, "sent_len_1": 66.6757, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.7337, "stdk": 0.0491, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 105400 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.4492, "doc_norm": 2.0127, "encoder_q-embeddings": 5175.9858, "encoder_q-layer.0": 3699.6082, "encoder_q-layer.1": 4042.9744, "encoder_q-layer.10": 6145.3833, "encoder_q-layer.11": 14712.8477, "encoder_q-layer.2": 4619.1743, "encoder_q-layer.3": 4849.77, "encoder_q-layer.4": 5160.127, "encoder_q-layer.5": 5435.8354, "encoder_q-layer.6": 5793.4722, "encoder_q-layer.7": 6313.9463, "encoder_q-layer.8": 6597.4951, "encoder_q-layer.9": 5930.0181, "epoch": 0.46, "inbatch_neg_score": 2.2039, "inbatch_pos_score": 2.8086, "learning_rate": 2.486842105263158e-05, "loss": 3.4492, "norm_diff": 0.0906, "norm_loss": 0.0, "num_token_doc": 66.716, "num_token_overlap": 11.6963, "num_token_query": 31.4363, "num_token_union": 65.1236, "num_word_context": 202.3096, "num_word_doc": 49.8133, "num_word_query": 23.3485, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9851.7873, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2051, "query_norm": 1.9222, "queue_k_norm": 2.009, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4363, "sent_len_1": 66.716, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.8475, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 105500 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.4381, "doc_norm": 2.0137, "encoder_q-embeddings": 4980.4751, "encoder_q-layer.0": 3370.729, "encoder_q-layer.1": 3560.665, "encoder_q-layer.10": 6444.1284, "encoder_q-layer.11": 15505.8828, "encoder_q-layer.2": 4038.4734, "encoder_q-layer.3": 4189.1592, "encoder_q-layer.4": 4581.4526, "encoder_q-layer.5": 4589.6211, "encoder_q-layer.6": 5087.6694, "encoder_q-layer.7": 5748.5991, "encoder_q-layer.8": 6897.6748, "encoder_q-layer.9": 6290.4702, "epoch": 0.46, "inbatch_neg_score": 2.2164, "inbatch_pos_score": 2.832, "learning_rate": 2.4842105263157898e-05, "loss": 3.4381, "norm_diff": 0.071, "norm_loss": 0.0, "num_token_doc": 66.7909, "num_token_overlap": 11.6799, "num_token_query": 31.3816, "num_token_union": 65.178, "num_word_context": 202.4001, "num_word_doc": 49.8893, "num_word_query": 23.3147, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9962.3291, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2148, "query_norm": 1.9427, "queue_k_norm": 2.0127, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3816, "sent_len_1": 66.7909, "sent_len_max_0": 127.9688, "sent_len_max_1": 206.2862, "stdk": 0.0493, "stdq": 0.0436, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 105600 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4425, "doc_norm": 2.0145, "encoder_q-embeddings": 5360.73, "encoder_q-layer.0": 3706.2991, "encoder_q-layer.1": 3959.3562, "encoder_q-layer.10": 6662.9648, "encoder_q-layer.11": 14860.1279, "encoder_q-layer.2": 4731.8271, "encoder_q-layer.3": 4859.0859, "encoder_q-layer.4": 5158.6807, "encoder_q-layer.5": 5282.4629, "encoder_q-layer.6": 5810.0312, "encoder_q-layer.7": 6373.4312, "encoder_q-layer.8": 7363.5103, "encoder_q-layer.9": 6357.1499, "epoch": 0.46, "inbatch_neg_score": 2.2223, "inbatch_pos_score": 2.8262, "learning_rate": 2.4815789473684213e-05, "loss": 3.4425, "norm_diff": 0.0741, "norm_loss": 0.0, "num_token_doc": 66.818, "num_token_overlap": 11.6495, "num_token_query": 31.3085, "num_token_union": 65.1486, "num_word_context": 202.1508, "num_word_doc": 49.8415, "num_word_query": 23.2379, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10170.0572, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2227, "query_norm": 1.9404, "queue_k_norm": 2.0129, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3085, "sent_len_1": 66.818, "sent_len_max_0": 127.985, "sent_len_max_1": 209.9675, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 105700 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4425, "doc_norm": 2.0134, "encoder_q-embeddings": 4839.0464, "encoder_q-layer.0": 3297.6709, "encoder_q-layer.1": 3580.2649, "encoder_q-layer.10": 6454.8247, "encoder_q-layer.11": 15107.1162, "encoder_q-layer.2": 4068.9866, "encoder_q-layer.3": 4046.4553, "encoder_q-layer.4": 4276.9429, "encoder_q-layer.5": 4522.8398, "encoder_q-layer.6": 5107.4844, "encoder_q-layer.7": 5742.1953, "encoder_q-layer.8": 6813.2622, "encoder_q-layer.9": 6145.0864, "epoch": 0.46, "inbatch_neg_score": 2.2364, "inbatch_pos_score": 2.8398, "learning_rate": 2.4789473684210528e-05, "loss": 3.4425, "norm_diff": 0.0635, "norm_loss": 0.0, "num_token_doc": 66.6962, "num_token_overlap": 11.6678, "num_token_query": 31.3898, "num_token_union": 65.0802, "num_word_context": 202.2653, "num_word_doc": 49.7446, "num_word_query": 23.3072, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9618.8502, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2363, "query_norm": 1.9499, "queue_k_norm": 2.0126, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3898, "sent_len_1": 66.6962, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.255, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 105800 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4269, "doc_norm": 2.0141, "encoder_q-embeddings": 5199.751, "encoder_q-layer.0": 3815.8425, "encoder_q-layer.1": 4035.9629, "encoder_q-layer.10": 5862.397, "encoder_q-layer.11": 14349.4609, "encoder_q-layer.2": 4399.5713, "encoder_q-layer.3": 4587.8486, "encoder_q-layer.4": 4870.4731, "encoder_q-layer.5": 5119.9585, "encoder_q-layer.6": 5691.3242, "encoder_q-layer.7": 6155.8516, "encoder_q-layer.8": 6846.9819, "encoder_q-layer.9": 6055.9033, "epoch": 0.46, "inbatch_neg_score": 2.2464, "inbatch_pos_score": 2.8633, "learning_rate": 2.4763157894736843e-05, "loss": 3.4269, "norm_diff": 0.0529, "norm_loss": 0.0, "num_token_doc": 66.7687, "num_token_overlap": 11.6963, "num_token_query": 31.4019, "num_token_union": 65.1046, "num_word_context": 202.1256, "num_word_doc": 49.828, "num_word_query": 23.3354, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9726.3921, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2461, "query_norm": 1.9612, "queue_k_norm": 2.0148, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4019, "sent_len_1": 66.7687, "sent_len_max_0": 127.9813, "sent_len_max_1": 211.0575, "stdk": 0.0492, "stdq": 0.0436, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 105900 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.4419, "doc_norm": 2.0167, "encoder_q-embeddings": 4968.8994, "encoder_q-layer.0": 3273.6575, "encoder_q-layer.1": 3513.6106, "encoder_q-layer.10": 7132.2656, "encoder_q-layer.11": 15463.6689, "encoder_q-layer.2": 3819.7896, "encoder_q-layer.3": 4104.3203, "encoder_q-layer.4": 4303.6099, "encoder_q-layer.5": 4367.7192, "encoder_q-layer.6": 5035.9307, "encoder_q-layer.7": 5462.9067, "encoder_q-layer.8": 6790.6309, "encoder_q-layer.9": 6518.0547, "epoch": 0.46, "inbatch_neg_score": 2.2577, "inbatch_pos_score": 2.8398, "learning_rate": 2.4736842105263158e-05, "loss": 3.4419, "norm_diff": 0.0626, "norm_loss": 0.0, "num_token_doc": 66.7946, "num_token_overlap": 11.6715, "num_token_query": 31.4745, "num_token_union": 65.2247, "num_word_context": 202.5287, "num_word_doc": 49.8621, "num_word_query": 23.3819, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9580.8451, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2578, "query_norm": 1.9542, "queue_k_norm": 2.016, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4745, "sent_len_1": 66.7946, "sent_len_max_0": 128.0, "sent_len_max_1": 207.8688, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106000 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.4336, "doc_norm": 2.0189, "encoder_q-embeddings": 4764.7798, "encoder_q-layer.0": 3311.1128, "encoder_q-layer.1": 3505.6858, "encoder_q-layer.10": 6009.1445, "encoder_q-layer.11": 14724.8037, "encoder_q-layer.2": 3940.9702, "encoder_q-layer.3": 4158.7964, "encoder_q-layer.4": 4301.0444, "encoder_q-layer.5": 4478.7144, "encoder_q-layer.6": 5039.0908, "encoder_q-layer.7": 5596.5, "encoder_q-layer.8": 6412.855, "encoder_q-layer.9": 6088.0654, "epoch": 0.46, "inbatch_neg_score": 2.2654, "inbatch_pos_score": 2.8809, "learning_rate": 2.4710526315789476e-05, "loss": 3.4336, "norm_diff": 0.0492, "norm_loss": 0.0, "num_token_doc": 66.9353, "num_token_overlap": 11.6366, "num_token_query": 31.2893, "num_token_union": 65.2133, "num_word_context": 202.5866, "num_word_doc": 49.9628, "num_word_query": 23.234, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9420.1153, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2656, "query_norm": 1.9697, "queue_k_norm": 2.0178, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2893, "sent_len_1": 66.9353, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.05, "stdk": 0.0493, "stdq": 0.0438, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106100 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4307, "doc_norm": 2.0189, "encoder_q-embeddings": 5641.8813, "encoder_q-layer.0": 4067.7712, "encoder_q-layer.1": 4664.0225, "encoder_q-layer.10": 6238.6235, "encoder_q-layer.11": 14539.9043, "encoder_q-layer.2": 5350.082, "encoder_q-layer.3": 5569.5059, "encoder_q-layer.4": 5919.6919, "encoder_q-layer.5": 5988.8281, "encoder_q-layer.6": 6505.1201, "encoder_q-layer.7": 6540.5254, "encoder_q-layer.8": 7131.6621, "encoder_q-layer.9": 6314.0801, "epoch": 0.46, "inbatch_neg_score": 2.2785, "inbatch_pos_score": 2.8828, "learning_rate": 2.468421052631579e-05, "loss": 3.4307, "norm_diff": 0.0606, "norm_loss": 0.0, "num_token_doc": 66.7735, "num_token_overlap": 11.714, "num_token_query": 31.4628, "num_token_union": 65.1419, "num_word_context": 202.3466, "num_word_doc": 49.8039, "num_word_query": 23.3573, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10445.8291, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2793, "query_norm": 1.9583, "queue_k_norm": 2.0195, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4628, "sent_len_1": 66.7735, "sent_len_max_0": 127.9688, "sent_len_max_1": 211.1538, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106200 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4275, "doc_norm": 2.0196, "encoder_q-embeddings": 5118.6953, "encoder_q-layer.0": 3540.1694, "encoder_q-layer.1": 3804.0327, "encoder_q-layer.10": 6265.771, "encoder_q-layer.11": 14947.1953, "encoder_q-layer.2": 4284.6108, "encoder_q-layer.3": 4430.0142, "encoder_q-layer.4": 4718.8589, "encoder_q-layer.5": 4836.6621, "encoder_q-layer.6": 5483.5688, "encoder_q-layer.7": 6187.1387, "encoder_q-layer.8": 7150.1323, "encoder_q-layer.9": 6546.2964, "epoch": 0.46, "inbatch_neg_score": 2.2861, "inbatch_pos_score": 2.8926, "learning_rate": 2.4657894736842106e-05, "loss": 3.4275, "norm_diff": 0.0555, "norm_loss": 0.0, "num_token_doc": 66.7184, "num_token_overlap": 11.6677, "num_token_query": 31.3378, "num_token_union": 65.0555, "num_word_context": 202.3288, "num_word_doc": 49.7947, "num_word_query": 23.2701, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9864.2689, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2852, "query_norm": 1.9642, "queue_k_norm": 2.0197, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3378, "sent_len_1": 66.7184, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.2663, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 106300 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.4361, "doc_norm": 2.022, "encoder_q-embeddings": 5323.9365, "encoder_q-layer.0": 3705.2517, "encoder_q-layer.1": 4031.3298, "encoder_q-layer.10": 6450.1719, "encoder_q-layer.11": 16028.9414, "encoder_q-layer.2": 4550.7725, "encoder_q-layer.3": 4494.0747, "encoder_q-layer.4": 4852.6294, "encoder_q-layer.5": 4927.8687, "encoder_q-layer.6": 5324.1826, "encoder_q-layer.7": 6060.7422, "encoder_q-layer.8": 6783.1943, "encoder_q-layer.9": 6318.6763, "epoch": 0.46, "inbatch_neg_score": 2.2905, "inbatch_pos_score": 2.9043, "learning_rate": 2.463157894736842e-05, "loss": 3.4361, "norm_diff": 0.0506, "norm_loss": 0.0, "num_token_doc": 67.0469, "num_token_overlap": 11.6914, "num_token_query": 31.3031, "num_token_union": 65.2277, "num_word_context": 202.5228, "num_word_doc": 50.0477, "num_word_query": 23.2655, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10050.2864, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2891, "query_norm": 1.9715, "queue_k_norm": 2.0233, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3031, "sent_len_1": 67.0469, "sent_len_max_0": 127.995, "sent_len_max_1": 207.4575, "stdk": 0.0492, "stdq": 0.044, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106400 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.435, "doc_norm": 2.0213, "encoder_q-embeddings": 4763.4502, "encoder_q-layer.0": 3329.9116, "encoder_q-layer.1": 3605.5581, "encoder_q-layer.10": 6765.2983, "encoder_q-layer.11": 15021.9248, "encoder_q-layer.2": 4007.5474, "encoder_q-layer.3": 4058.9932, "encoder_q-layer.4": 4344.9385, "encoder_q-layer.5": 4396.8081, "encoder_q-layer.6": 5010.4917, "encoder_q-layer.7": 5694.5537, "encoder_q-layer.8": 6781.7114, "encoder_q-layer.9": 6156.9092, "epoch": 0.46, "inbatch_neg_score": 2.2942, "inbatch_pos_score": 2.8887, "learning_rate": 2.4605263157894736e-05, "loss": 3.435, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.7889, "num_token_overlap": 11.6854, "num_token_query": 31.4135, "num_token_union": 65.1452, "num_word_context": 202.3277, "num_word_doc": 49.8229, "num_word_query": 23.3184, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9689.0083, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.293, "query_norm": 1.953, "queue_k_norm": 2.0239, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4135, "sent_len_1": 66.7889, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.7025, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106500 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4391, "doc_norm": 2.0254, "encoder_q-embeddings": 5302.9907, "encoder_q-layer.0": 3845.7524, "encoder_q-layer.1": 4358.5015, "encoder_q-layer.10": 6954.6147, "encoder_q-layer.11": 15616.7061, "encoder_q-layer.2": 5004.1616, "encoder_q-layer.3": 5612.3779, "encoder_q-layer.4": 6009.9551, "encoder_q-layer.5": 5297.9272, "encoder_q-layer.6": 5962.9463, "encoder_q-layer.7": 6127.5, "encoder_q-layer.8": 7082.0049, "encoder_q-layer.9": 6505.6084, "epoch": 0.46, "inbatch_neg_score": 2.2956, "inbatch_pos_score": 2.9023, "learning_rate": 2.4578947368421054e-05, "loss": 3.4391, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.761, "num_token_overlap": 11.6613, "num_token_query": 31.3083, "num_token_union": 65.0702, "num_word_context": 202.0957, "num_word_doc": 49.8084, "num_word_query": 23.2562, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10380.9375, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.2949, "query_norm": 1.9458, "queue_k_norm": 2.0259, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3083, "sent_len_1": 66.761, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.1025, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106600 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4193, "doc_norm": 2.0293, "encoder_q-embeddings": 5829.7754, "encoder_q-layer.0": 4050.6465, "encoder_q-layer.1": 4348.4263, "encoder_q-layer.10": 6558.8281, "encoder_q-layer.11": 14620.252, "encoder_q-layer.2": 5007.3394, "encoder_q-layer.3": 5097.9585, "encoder_q-layer.4": 5579.9839, "encoder_q-layer.5": 5950.8086, "encoder_q-layer.6": 6721.0928, "encoder_q-layer.7": 8018.8813, "encoder_q-layer.8": 11861.8145, "encoder_q-layer.9": 9459.209, "epoch": 0.46, "inbatch_neg_score": 2.3012, "inbatch_pos_score": 2.8945, "learning_rate": 2.455263157894737e-05, "loss": 3.4193, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.9168, "num_token_overlap": 11.7011, "num_token_query": 31.4294, "num_token_union": 65.2271, "num_word_context": 202.3235, "num_word_doc": 49.9325, "num_word_query": 23.3539, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11666.1805, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3027, "query_norm": 1.9505, "queue_k_norm": 2.0264, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4294, "sent_len_1": 66.9168, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.1175, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 106700 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.4497, "doc_norm": 2.0258, "encoder_q-embeddings": 6290.189, "encoder_q-layer.0": 4394.9092, "encoder_q-layer.1": 5121.4829, "encoder_q-layer.10": 7081.8286, "encoder_q-layer.11": 15816.3359, "encoder_q-layer.2": 5819.5293, "encoder_q-layer.3": 6231.6074, "encoder_q-layer.4": 6598.4756, "encoder_q-layer.5": 6775.5571, "encoder_q-layer.6": 7203.7397, "encoder_q-layer.7": 7369.1147, "encoder_q-layer.8": 7619.2017, "encoder_q-layer.9": 6484.7983, "epoch": 0.46, "inbatch_neg_score": 2.2997, "inbatch_pos_score": 2.8945, "learning_rate": 2.4526315789473688e-05, "loss": 3.4497, "norm_diff": 0.0708, "norm_loss": 0.0, "num_token_doc": 66.6087, "num_token_overlap": 11.6853, "num_token_query": 31.4431, "num_token_union": 65.0657, "num_word_context": 202.3877, "num_word_doc": 49.6945, "num_word_query": 23.3666, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11269.9224, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3008, "query_norm": 1.955, "queue_k_norm": 2.0277, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4431, "sent_len_1": 66.6087, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.8713, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106800 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.4419, "doc_norm": 2.0318, "encoder_q-embeddings": 9774.3535, "encoder_q-layer.0": 6806.041, "encoder_q-layer.1": 7212.0171, "encoder_q-layer.10": 13225.0537, "encoder_q-layer.11": 31216.9023, "encoder_q-layer.2": 7918.3032, "encoder_q-layer.3": 8216.6484, "encoder_q-layer.4": 9038.7881, "encoder_q-layer.5": 9075.6514, "encoder_q-layer.6": 9855.9678, "encoder_q-layer.7": 11568.9863, "encoder_q-layer.8": 13546.2754, "encoder_q-layer.9": 12563.4131, "epoch": 0.46, "inbatch_neg_score": 2.3042, "inbatch_pos_score": 2.9004, "learning_rate": 2.45e-05, "loss": 3.4419, "norm_diff": 0.0874, "norm_loss": 0.0, "num_token_doc": 66.7496, "num_token_overlap": 11.6711, "num_token_query": 31.3567, "num_token_union": 65.0892, "num_word_context": 202.3436, "num_word_doc": 49.8078, "num_word_query": 23.2916, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20052.9137, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.3047, "query_norm": 1.9444, "queue_k_norm": 2.0303, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3567, "sent_len_1": 66.7496, "sent_len_max_0": 127.98, "sent_len_max_1": 209.1375, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 106900 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.4356, "doc_norm": 2.032, "encoder_q-embeddings": 7583.2002, "encoder_q-layer.0": 5379.9731, "encoder_q-layer.1": 5798.2354, "encoder_q-layer.10": 6604.667, "encoder_q-layer.11": 15635.3652, "encoder_q-layer.2": 6655.9907, "encoder_q-layer.3": 7320.9746, "encoder_q-layer.4": 7487.8799, "encoder_q-layer.5": 7625.0078, "encoder_q-layer.6": 7649.3589, "encoder_q-layer.7": 7405.7285, "encoder_q-layer.8": 7767.8354, "encoder_q-layer.9": 6525.4795, "epoch": 0.46, "inbatch_neg_score": 2.302, "inbatch_pos_score": 2.8965, "learning_rate": 2.4473684210526318e-05, "loss": 3.4356, "norm_diff": 0.0863, "norm_loss": 0.0, "num_token_doc": 66.8812, "num_token_overlap": 11.6762, "num_token_query": 31.3016, "num_token_union": 65.1447, "num_word_context": 202.2275, "num_word_doc": 49.9127, "num_word_query": 23.2353, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12075.9256, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3027, "query_norm": 1.9457, "queue_k_norm": 2.0312, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3016, "sent_len_1": 66.8812, "sent_len_max_0": 127.9638, "sent_len_max_1": 210.2413, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107000 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.4326, "doc_norm": 2.0348, "encoder_q-embeddings": 5400.1914, "encoder_q-layer.0": 3791.4761, "encoder_q-layer.1": 4139.0986, "encoder_q-layer.10": 7288.7378, "encoder_q-layer.11": 16121.2988, "encoder_q-layer.2": 4564.8345, "encoder_q-layer.3": 4670.7002, "encoder_q-layer.4": 5086.6338, "encoder_q-layer.5": 5151.373, "encoder_q-layer.6": 5524.9614, "encoder_q-layer.7": 6258.7188, "encoder_q-layer.8": 7019.9316, "encoder_q-layer.9": 6331.0542, "epoch": 0.46, "inbatch_neg_score": 2.3049, "inbatch_pos_score": 2.916, "learning_rate": 2.4447368421052633e-05, "loss": 3.4326, "norm_diff": 0.0856, "norm_loss": 0.0, "num_token_doc": 66.7492, "num_token_overlap": 11.688, "num_token_query": 31.3251, "num_token_union": 65.0646, "num_word_context": 202.2504, "num_word_doc": 49.8244, "num_word_query": 23.2567, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10329.318, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3047, "query_norm": 1.9492, "queue_k_norm": 2.032, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3251, "sent_len_1": 66.7492, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.5525, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107100 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.4326, "doc_norm": 2.0319, "encoder_q-embeddings": 5850.0444, "encoder_q-layer.0": 4089.3286, "encoder_q-layer.1": 4280.2095, "encoder_q-layer.10": 6362.0181, "encoder_q-layer.11": 15257.1318, "encoder_q-layer.2": 5377.3574, "encoder_q-layer.3": 5389.3896, "encoder_q-layer.4": 5619.7876, "encoder_q-layer.5": 5773.7202, "encoder_q-layer.6": 6429.168, "encoder_q-layer.7": 7614.939, "encoder_q-layer.8": 7710.2974, "encoder_q-layer.9": 6325.5498, "epoch": 0.47, "inbatch_neg_score": 2.3064, "inbatch_pos_score": 2.8965, "learning_rate": 2.4421052631578948e-05, "loss": 3.4326, "norm_diff": 0.0884, "norm_loss": 0.0, "num_token_doc": 66.6919, "num_token_overlap": 11.686, "num_token_query": 31.4243, "num_token_union": 65.095, "num_word_context": 202.4845, "num_word_doc": 49.78, "num_word_query": 23.331, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10812.9344, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3066, "query_norm": 1.9435, "queue_k_norm": 2.034, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4243, "sent_len_1": 66.6919, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.4625, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107200 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.434, "doc_norm": 2.0335, "encoder_q-embeddings": 5025.7095, "encoder_q-layer.0": 3403.4536, "encoder_q-layer.1": 3696.9578, "encoder_q-layer.10": 6701.6597, "encoder_q-layer.11": 15704.8281, "encoder_q-layer.2": 4174.896, "encoder_q-layer.3": 4270.6582, "encoder_q-layer.4": 4656.1851, "encoder_q-layer.5": 4805.0933, "encoder_q-layer.6": 5243.3408, "encoder_q-layer.7": 5685.3501, "encoder_q-layer.8": 7047.1216, "encoder_q-layer.9": 6501.6436, "epoch": 0.47, "inbatch_neg_score": 2.3076, "inbatch_pos_score": 2.9121, "learning_rate": 2.4394736842105266e-05, "loss": 3.434, "norm_diff": 0.0882, "norm_loss": 0.0, "num_token_doc": 66.793, "num_token_overlap": 11.695, "num_token_query": 31.4561, "num_token_union": 65.1376, "num_word_context": 202.5153, "num_word_doc": 49.8104, "num_word_query": 23.3547, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10071.5468, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3086, "query_norm": 1.9454, "queue_k_norm": 2.033, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4561, "sent_len_1": 66.793, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.0575, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0492, "stdqueue_q": 0.0, "step": 107300 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4423, "doc_norm": 2.0356, "encoder_q-embeddings": 6810.3462, "encoder_q-layer.0": 4884.5151, "encoder_q-layer.1": 5083.6777, "encoder_q-layer.10": 6161.3501, "encoder_q-layer.11": 15086.4883, "encoder_q-layer.2": 6012.4854, "encoder_q-layer.3": 6402.0908, "encoder_q-layer.4": 6694.4053, "encoder_q-layer.5": 6695.2222, "encoder_q-layer.6": 7068.8994, "encoder_q-layer.7": 7154.8711, "encoder_q-layer.8": 7547.144, "encoder_q-layer.9": 6327.8843, "epoch": 0.47, "inbatch_neg_score": 2.3065, "inbatch_pos_score": 2.916, "learning_rate": 2.436842105263158e-05, "loss": 3.4423, "norm_diff": 0.0889, "norm_loss": 0.0, "num_token_doc": 66.6529, "num_token_overlap": 11.6681, "num_token_query": 31.3459, "num_token_union": 65.0181, "num_word_context": 202.1632, "num_word_doc": 49.7467, "num_word_query": 23.2828, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11311.5854, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3066, "query_norm": 1.9466, "queue_k_norm": 2.0343, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3459, "sent_len_1": 66.6529, "sent_len_max_0": 127.9788, "sent_len_max_1": 207.905, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107400 }, { "accuracy": 53.3691, "active_queue_size": 16384.0, "cl_loss": 3.4338, "doc_norm": 2.0377, "encoder_q-embeddings": 5184.5781, "encoder_q-layer.0": 3585.8616, "encoder_q-layer.1": 3863.8132, "encoder_q-layer.10": 6355.4717, "encoder_q-layer.11": 15159.084, "encoder_q-layer.2": 4285.355, "encoder_q-layer.3": 4257.6011, "encoder_q-layer.4": 4643.3955, "encoder_q-layer.5": 4597.1973, "encoder_q-layer.6": 4983.5566, "encoder_q-layer.7": 5756.1436, "encoder_q-layer.8": 6412.7871, "encoder_q-layer.9": 6046.0845, "epoch": 0.47, "inbatch_neg_score": 2.3121, "inbatch_pos_score": 2.9316, "learning_rate": 2.4342105263157896e-05, "loss": 3.4338, "norm_diff": 0.081, "norm_loss": 0.0, "num_token_doc": 66.9691, "num_token_overlap": 11.674, "num_token_query": 31.4071, "num_token_union": 65.2388, "num_word_context": 202.5366, "num_word_doc": 49.9452, "num_word_query": 23.3195, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9566.2354, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3125, "query_norm": 1.9567, "queue_k_norm": 2.0373, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4071, "sent_len_1": 66.9691, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.805, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 107500 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4485, "doc_norm": 2.0389, "encoder_q-embeddings": 5576.4521, "encoder_q-layer.0": 3875.3069, "encoder_q-layer.1": 4015.6072, "encoder_q-layer.10": 6415.6724, "encoder_q-layer.11": 15137.1602, "encoder_q-layer.2": 4405.3813, "encoder_q-layer.3": 4527.2944, "encoder_q-layer.4": 4826.0239, "encoder_q-layer.5": 4890.019, "encoder_q-layer.6": 5508.7979, "encoder_q-layer.7": 6186.2583, "encoder_q-layer.8": 7788.6943, "encoder_q-layer.9": 6515.6421, "epoch": 0.47, "inbatch_neg_score": 2.3174, "inbatch_pos_score": 2.9238, "learning_rate": 2.431578947368421e-05, "loss": 3.4485, "norm_diff": 0.0819, "norm_loss": 0.0, "num_token_doc": 66.7334, "num_token_overlap": 11.6782, "num_token_query": 31.4354, "num_token_union": 65.1365, "num_word_context": 202.4649, "num_word_doc": 49.7795, "num_word_query": 23.348, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10307.4938, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3184, "query_norm": 1.957, "queue_k_norm": 2.0371, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4354, "sent_len_1": 66.7334, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.2612, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107600 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.4262, "doc_norm": 2.0358, "encoder_q-embeddings": 5777.1758, "encoder_q-layer.0": 4085.6355, "encoder_q-layer.1": 4505.1992, "encoder_q-layer.10": 6582.5132, "encoder_q-layer.11": 15863.2051, "encoder_q-layer.2": 5088.1655, "encoder_q-layer.3": 5265.6689, "encoder_q-layer.4": 5613.8271, "encoder_q-layer.5": 5979.7354, "encoder_q-layer.6": 6230.5635, "encoder_q-layer.7": 6804.2236, "encoder_q-layer.8": 7475.3296, "encoder_q-layer.9": 6658.0088, "epoch": 0.47, "inbatch_neg_score": 2.3222, "inbatch_pos_score": 2.9355, "learning_rate": 2.428947368421053e-05, "loss": 3.4262, "norm_diff": 0.0803, "norm_loss": 0.0, "num_token_doc": 66.8766, "num_token_overlap": 11.693, "num_token_query": 31.426, "num_token_union": 65.221, "num_word_context": 202.5568, "num_word_doc": 49.9288, "num_word_query": 23.3446, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10844.5936, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3223, "query_norm": 1.9556, "queue_k_norm": 2.0383, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.426, "sent_len_1": 66.8766, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.11, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4359, "doc_norm": 2.04, "encoder_q-embeddings": 5177.749, "encoder_q-layer.0": 3493.6892, "encoder_q-layer.1": 3870.6951, "encoder_q-layer.10": 6599.311, "encoder_q-layer.11": 15081.0508, "encoder_q-layer.2": 4517.7202, "encoder_q-layer.3": 4634.332, "encoder_q-layer.4": 4903.1592, "encoder_q-layer.5": 5047.3838, "encoder_q-layer.6": 5631.8135, "encoder_q-layer.7": 6304.0117, "encoder_q-layer.8": 6988.3457, "encoder_q-layer.9": 6457.4058, "epoch": 0.47, "inbatch_neg_score": 2.3272, "inbatch_pos_score": 2.9375, "learning_rate": 2.4263157894736844e-05, "loss": 3.4359, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.626, "num_token_overlap": 11.6917, "num_token_query": 31.3854, "num_token_union": 64.9882, "num_word_context": 202.1944, "num_word_doc": 49.6982, "num_word_query": 23.3166, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10051.9735, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3281, "query_norm": 1.9643, "queue_k_norm": 2.0387, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3854, "sent_len_1": 66.626, "sent_len_max_0": 127.985, "sent_len_max_1": 209.225, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 107800 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.4444, "doc_norm": 2.0425, "encoder_q-embeddings": 4908.1318, "encoder_q-layer.0": 3347.7812, "encoder_q-layer.1": 3503.4888, "encoder_q-layer.10": 6681.2549, "encoder_q-layer.11": 15399.3076, "encoder_q-layer.2": 3859.9207, "encoder_q-layer.3": 4084.7505, "encoder_q-layer.4": 4425.6162, "encoder_q-layer.5": 4531.3735, "encoder_q-layer.6": 4964.8384, "encoder_q-layer.7": 5669.3145, "encoder_q-layer.8": 6781.8486, "encoder_q-layer.9": 6565.8516, "epoch": 0.47, "inbatch_neg_score": 2.331, "inbatch_pos_score": 2.9277, "learning_rate": 2.423684210526316e-05, "loss": 3.4444, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.6407, "num_token_overlap": 11.6753, "num_token_query": 31.348, "num_token_union": 65.0436, "num_word_context": 202.011, "num_word_doc": 49.6934, "num_word_query": 23.288, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9773.3667, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.332, "query_norm": 1.9576, "queue_k_norm": 2.042, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.348, "sent_len_1": 66.6407, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.7188, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 107900 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.4488, "doc_norm": 2.0404, "encoder_q-embeddings": 15991.8779, "encoder_q-layer.0": 12456.5195, "encoder_q-layer.1": 13949.2109, "encoder_q-layer.10": 6897.2617, "encoder_q-layer.11": 15265.3584, "encoder_q-layer.2": 11028.7266, "encoder_q-layer.3": 10141.6133, "encoder_q-layer.4": 9634.4072, "encoder_q-layer.5": 8416.1279, "encoder_q-layer.6": 7687.6943, "encoder_q-layer.7": 6901.8496, "encoder_q-layer.8": 7179.7642, "encoder_q-layer.9": 6066.7852, "epoch": 0.47, "inbatch_neg_score": 2.3356, "inbatch_pos_score": 2.9277, "learning_rate": 2.4210526315789474e-05, "loss": 3.4488, "norm_diff": 0.0887, "norm_loss": 0.0, "num_token_doc": 66.6531, "num_token_overlap": 11.6318, "num_token_query": 31.2883, "num_token_union": 65.0344, "num_word_context": 202.1979, "num_word_doc": 49.7311, "num_word_query": 23.2394, "postclip_grad_norm": 1.0, "preclip_grad_norm": 16272.3103, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3359, "query_norm": 1.9517, "queue_k_norm": 2.0417, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2883, "sent_len_1": 66.6531, "sent_len_max_0": 127.9675, "sent_len_max_1": 206.62, "stdk": 0.0492, "stdq": 0.0425, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108000 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.4212, "doc_norm": 2.0409, "encoder_q-embeddings": 4728.5269, "encoder_q-layer.0": 3305.4558, "encoder_q-layer.1": 3455.4456, "encoder_q-layer.10": 6545.2104, "encoder_q-layer.11": 14921.5137, "encoder_q-layer.2": 3930.8606, "encoder_q-layer.3": 3951.3967, "encoder_q-layer.4": 4291.9946, "encoder_q-layer.5": 4316.2139, "encoder_q-layer.6": 4781.6582, "encoder_q-layer.7": 5509.1323, "encoder_q-layer.8": 6428.9741, "encoder_q-layer.9": 6095.3496, "epoch": 0.47, "inbatch_neg_score": 2.3407, "inbatch_pos_score": 2.9512, "learning_rate": 2.418421052631579e-05, "loss": 3.4212, "norm_diff": 0.0763, "norm_loss": 0.0, "num_token_doc": 66.8406, "num_token_overlap": 11.6906, "num_token_query": 31.3764, "num_token_union": 65.1734, "num_word_context": 202.2221, "num_word_doc": 49.8986, "num_word_query": 23.3128, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9405.0162, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3398, "query_norm": 1.9646, "queue_k_norm": 2.0429, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3764, "sent_len_1": 66.8406, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.2138, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108100 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.4405, "doc_norm": 2.0421, "encoder_q-embeddings": 5030.0884, "encoder_q-layer.0": 3457.72, "encoder_q-layer.1": 3724.0847, "encoder_q-layer.10": 6595.522, "encoder_q-layer.11": 15571.6045, "encoder_q-layer.2": 4332.8223, "encoder_q-layer.3": 4378.1323, "encoder_q-layer.4": 4582.7256, "encoder_q-layer.5": 4663.8418, "encoder_q-layer.6": 4987.5781, "encoder_q-layer.7": 5858.6895, "encoder_q-layer.8": 6987.3027, "encoder_q-layer.9": 6719.9043, "epoch": 0.47, "inbatch_neg_score": 2.3438, "inbatch_pos_score": 2.9434, "learning_rate": 2.4157894736842107e-05, "loss": 3.4405, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.7295, "num_token_overlap": 11.6554, "num_token_query": 31.3382, "num_token_union": 65.0995, "num_word_context": 202.4384, "num_word_doc": 49.7924, "num_word_query": 23.2696, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9919.694, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3438, "query_norm": 1.9722, "queue_k_norm": 2.042, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3382, "sent_len_1": 66.7295, "sent_len_max_0": 127.985, "sent_len_max_1": 206.1475, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 108200 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4321, "doc_norm": 2.0427, "encoder_q-embeddings": 5073.9746, "encoder_q-layer.0": 3413.2783, "encoder_q-layer.1": 3694.0652, "encoder_q-layer.10": 7319.5239, "encoder_q-layer.11": 15811.2451, "encoder_q-layer.2": 4140.3955, "encoder_q-layer.3": 4331.5151, "encoder_q-layer.4": 4501.4663, "encoder_q-layer.5": 4800.4155, "encoder_q-layer.6": 5240.0688, "encoder_q-layer.7": 6052.2969, "encoder_q-layer.8": 7460.9751, "encoder_q-layer.9": 6934.3159, "epoch": 0.47, "inbatch_neg_score": 2.3474, "inbatch_pos_score": 2.9414, "learning_rate": 2.4131578947368422e-05, "loss": 3.4321, "norm_diff": 0.0815, "norm_loss": 0.0, "num_token_doc": 66.7348, "num_token_overlap": 11.6884, "num_token_query": 31.3949, "num_token_union": 65.0991, "num_word_context": 202.321, "num_word_doc": 49.821, "num_word_query": 23.3361, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10158.5929, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3477, "query_norm": 1.9612, "queue_k_norm": 2.0456, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3949, "sent_len_1": 66.7348, "sent_len_max_0": 128.0, "sent_len_max_1": 207.5275, "stdk": 0.0492, "stdq": 0.0426, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108300 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4105, "doc_norm": 2.0454, "encoder_q-embeddings": 5265.6719, "encoder_q-layer.0": 3638.2275, "encoder_q-layer.1": 3949.9004, "encoder_q-layer.10": 6606.0117, "encoder_q-layer.11": 15400.3809, "encoder_q-layer.2": 4291.5938, "encoder_q-layer.3": 4482.8389, "encoder_q-layer.4": 4646.6108, "encoder_q-layer.5": 4857.2534, "encoder_q-layer.6": 5321.2549, "encoder_q-layer.7": 5796.7554, "encoder_q-layer.8": 6830.5869, "encoder_q-layer.9": 6401.7354, "epoch": 0.47, "inbatch_neg_score": 2.3555, "inbatch_pos_score": 2.9434, "learning_rate": 2.410526315789474e-05, "loss": 3.4105, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.785, "num_token_overlap": 11.7108, "num_token_query": 31.4616, "num_token_union": 65.1408, "num_word_context": 202.0975, "num_word_doc": 49.8259, "num_word_query": 23.362, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9940.9936, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3555, "query_norm": 1.9728, "queue_k_norm": 2.0458, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4616, "sent_len_1": 66.785, "sent_len_max_0": 127.9737, "sent_len_max_1": 207.7237, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108400 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4253, "doc_norm": 2.0475, "encoder_q-embeddings": 4688.751, "encoder_q-layer.0": 3290.1794, "encoder_q-layer.1": 3500.6641, "encoder_q-layer.10": 6430.665, "encoder_q-layer.11": 15126.7324, "encoder_q-layer.2": 3878.0859, "encoder_q-layer.3": 3956.0068, "encoder_q-layer.4": 4355.6699, "encoder_q-layer.5": 4496.0386, "encoder_q-layer.6": 4966.5962, "encoder_q-layer.7": 6013.2305, "encoder_q-layer.8": 7048.5327, "encoder_q-layer.9": 6172.6279, "epoch": 0.47, "inbatch_neg_score": 2.3575, "inbatch_pos_score": 2.9707, "learning_rate": 2.4078947368421052e-05, "loss": 3.4253, "norm_diff": 0.0634, "norm_loss": 0.0, "num_token_doc": 66.5344, "num_token_overlap": 11.656, "num_token_query": 31.4053, "num_token_union": 65.0404, "num_word_context": 201.9437, "num_word_doc": 49.6277, "num_word_query": 23.334, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9645.7469, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3574, "query_norm": 1.9841, "queue_k_norm": 2.0451, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4053, "sent_len_1": 66.5344, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.5263, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 108500 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4348, "doc_norm": 2.0502, "encoder_q-embeddings": 4976.2437, "encoder_q-layer.0": 3347.1511, "encoder_q-layer.1": 3525.0493, "encoder_q-layer.10": 6180.855, "encoder_q-layer.11": 15330.291, "encoder_q-layer.2": 4000.4998, "encoder_q-layer.3": 4179.1699, "encoder_q-layer.4": 4314.771, "encoder_q-layer.5": 4444.1851, "encoder_q-layer.6": 5013.7339, "encoder_q-layer.7": 5820.6821, "encoder_q-layer.8": 7114.0381, "encoder_q-layer.9": 6070.1016, "epoch": 0.47, "inbatch_neg_score": 2.364, "inbatch_pos_score": 2.9688, "learning_rate": 2.4052631578947367e-05, "loss": 3.4348, "norm_diff": 0.0694, "norm_loss": 0.0, "num_token_doc": 66.6841, "num_token_overlap": 11.6779, "num_token_query": 31.4108, "num_token_union": 65.0663, "num_word_context": 201.9908, "num_word_doc": 49.772, "num_word_query": 23.3286, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9822.1291, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3633, "query_norm": 1.9808, "queue_k_norm": 2.0485, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4108, "sent_len_1": 66.6841, "sent_len_max_0": 127.9425, "sent_len_max_1": 209.6687, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108600 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.4376, "doc_norm": 2.0473, "encoder_q-embeddings": 5147.0215, "encoder_q-layer.0": 3471.2109, "encoder_q-layer.1": 3769.45, "encoder_q-layer.10": 6821.7021, "encoder_q-layer.11": 15733.2715, "encoder_q-layer.2": 4134.4146, "encoder_q-layer.3": 4297.3198, "encoder_q-layer.4": 4516.9121, "encoder_q-layer.5": 4847.2437, "encoder_q-layer.6": 5393.3672, "encoder_q-layer.7": 5999.6069, "encoder_q-layer.8": 7268.0664, "encoder_q-layer.9": 6554.4043, "epoch": 0.47, "inbatch_neg_score": 2.3734, "inbatch_pos_score": 2.9648, "learning_rate": 2.4026315789473686e-05, "loss": 3.4376, "norm_diff": 0.0646, "norm_loss": 0.0, "num_token_doc": 66.8864, "num_token_overlap": 11.6597, "num_token_query": 31.3261, "num_token_union": 65.1588, "num_word_context": 202.2801, "num_word_doc": 49.9089, "num_word_query": 23.2538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10096.9085, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.373, "query_norm": 1.9827, "queue_k_norm": 2.049, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3261, "sent_len_1": 66.8864, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.1712, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 108700 }, { "accuracy": 48.0957, "active_queue_size": 16384.0, "cl_loss": 3.4544, "doc_norm": 2.0484, "encoder_q-embeddings": 4814.0288, "encoder_q-layer.0": 3388.2036, "encoder_q-layer.1": 3586.7427, "encoder_q-layer.10": 6654.4131, "encoder_q-layer.11": 15975.6895, "encoder_q-layer.2": 3937.5432, "encoder_q-layer.3": 3988.9741, "encoder_q-layer.4": 4244.6431, "encoder_q-layer.5": 4322.5337, "encoder_q-layer.6": 4712.5879, "encoder_q-layer.7": 5651.4976, "encoder_q-layer.8": 6623.8535, "encoder_q-layer.9": 6343.3169, "epoch": 0.47, "inbatch_neg_score": 2.3879, "inbatch_pos_score": 2.9668, "learning_rate": 2.4e-05, "loss": 3.4544, "norm_diff": 0.0605, "norm_loss": 0.0, "num_token_doc": 66.7785, "num_token_overlap": 11.6969, "num_token_query": 31.3591, "num_token_union": 65.0966, "num_word_context": 202.4368, "num_word_doc": 49.7791, "num_word_query": 23.2753, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9908.2406, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.3867, "query_norm": 1.9879, "queue_k_norm": 2.0489, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3591, "sent_len_1": 66.7785, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.7925, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 108800 }, { "accuracy": 52.832, "active_queue_size": 16384.0, "cl_loss": 3.4153, "doc_norm": 2.0546, "encoder_q-embeddings": 4849.0757, "encoder_q-layer.0": 3341.2598, "encoder_q-layer.1": 3453.2517, "encoder_q-layer.10": 6690.4775, "encoder_q-layer.11": 15660.9746, "encoder_q-layer.2": 3890.2583, "encoder_q-layer.3": 4001.0229, "encoder_q-layer.4": 4358.8594, "encoder_q-layer.5": 4390.4165, "encoder_q-layer.6": 4824.9658, "encoder_q-layer.7": 5507.1753, "encoder_q-layer.8": 6704.7231, "encoder_q-layer.9": 6404.3037, "epoch": 0.47, "inbatch_neg_score": 2.3993, "inbatch_pos_score": 3.0156, "learning_rate": 2.397368421052632e-05, "loss": 3.4153, "norm_diff": 0.0488, "norm_loss": 0.0, "num_token_doc": 66.9796, "num_token_overlap": 11.7474, "num_token_query": 31.4102, "num_token_union": 65.1844, "num_word_context": 202.8318, "num_word_doc": 49.9699, "num_word_query": 23.3258, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9694.2047, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.4004, "query_norm": 2.0058, "queue_k_norm": 2.0508, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4102, "sent_len_1": 66.9796, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.4725, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 108900 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4299, "doc_norm": 2.0521, "encoder_q-embeddings": 9790.3242, "encoder_q-layer.0": 6872.6025, "encoder_q-layer.1": 7487.5586, "encoder_q-layer.10": 12347.8945, "encoder_q-layer.11": 30502.7012, "encoder_q-layer.2": 8207.8174, "encoder_q-layer.3": 8656.9707, "encoder_q-layer.4": 9078.292, "encoder_q-layer.5": 9707.0713, "encoder_q-layer.6": 10266.1895, "encoder_q-layer.7": 11173.9141, "encoder_q-layer.8": 12938.0361, "encoder_q-layer.9": 12358.7314, "epoch": 0.47, "inbatch_neg_score": 2.4113, "inbatch_pos_score": 3.0195, "learning_rate": 2.394736842105263e-05, "loss": 3.4299, "norm_diff": 0.0436, "norm_loss": 0.0, "num_token_doc": 66.7016, "num_token_overlap": 11.7073, "num_token_query": 31.43, "num_token_union": 65.0875, "num_word_context": 202.179, "num_word_doc": 49.7419, "num_word_query": 23.3499, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19632.2701, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4102, "query_norm": 2.0085, "queue_k_norm": 2.0515, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.43, "sent_len_1": 66.7016, "sent_len_max_0": 127.995, "sent_len_max_1": 208.4563, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 109000 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4328, "doc_norm": 2.0524, "encoder_q-embeddings": 9711.8301, "encoder_q-layer.0": 6845.8579, "encoder_q-layer.1": 7379.0059, "encoder_q-layer.10": 12970.7314, "encoder_q-layer.11": 28954.7695, "encoder_q-layer.2": 8106.3823, "encoder_q-layer.3": 8758.709, "encoder_q-layer.4": 9191.1182, "encoder_q-layer.5": 9176.8174, "encoder_q-layer.6": 9811.9023, "encoder_q-layer.7": 11028.9336, "encoder_q-layer.8": 13485.3438, "encoder_q-layer.9": 11995.123, "epoch": 0.47, "inbatch_neg_score": 2.4303, "inbatch_pos_score": 3.041, "learning_rate": 2.392105263157895e-05, "loss": 3.4328, "norm_diff": 0.0342, "norm_loss": 0.0, "num_token_doc": 66.7762, "num_token_overlap": 11.6726, "num_token_query": 31.405, "num_token_union": 65.1458, "num_word_context": 202.5175, "num_word_doc": 49.8453, "num_word_query": 23.3234, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18817.0248, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4297, "query_norm": 2.0182, "queue_k_norm": 2.0544, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.405, "sent_len_1": 66.7762, "sent_len_max_0": 127.995, "sent_len_max_1": 208.4363, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 109100 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4318, "doc_norm": 2.0547, "encoder_q-embeddings": 9886.6973, "encoder_q-layer.0": 7093.7144, "encoder_q-layer.1": 7506.7583, "encoder_q-layer.10": 12743.1279, "encoder_q-layer.11": 29867.4375, "encoder_q-layer.2": 8603.7109, "encoder_q-layer.3": 8708.5195, "encoder_q-layer.4": 9179.0518, "encoder_q-layer.5": 9165.8174, "encoder_q-layer.6": 9805.2344, "encoder_q-layer.7": 11154.8594, "encoder_q-layer.8": 12957.1436, "encoder_q-layer.9": 12552.9395, "epoch": 0.47, "inbatch_neg_score": 2.4425, "inbatch_pos_score": 3.0547, "learning_rate": 2.3894736842105264e-05, "loss": 3.4318, "norm_diff": 0.0324, "norm_loss": 0.0, "num_token_doc": 66.6351, "num_token_overlap": 11.6723, "num_token_query": 31.3681, "num_token_union": 65.0405, "num_word_context": 202.3642, "num_word_doc": 49.7468, "num_word_query": 23.3102, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19279.9773, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4434, "query_norm": 2.0223, "queue_k_norm": 2.0562, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3681, "sent_len_1": 66.6351, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.1188, "stdk": 0.0492, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 109200 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.4272, "doc_norm": 2.0593, "encoder_q-embeddings": 10289.5029, "encoder_q-layer.0": 7132.6987, "encoder_q-layer.1": 7414.8965, "encoder_q-layer.10": 13377.792, "encoder_q-layer.11": 30474.4297, "encoder_q-layer.2": 8184.6191, "encoder_q-layer.3": 8681.1045, "encoder_q-layer.4": 9346.1553, "encoder_q-layer.5": 9706.1504, "encoder_q-layer.6": 10512.0352, "encoder_q-layer.7": 11508.3096, "encoder_q-layer.8": 14418.4795, "encoder_q-layer.9": 12560.042, "epoch": 0.47, "inbatch_neg_score": 2.4607, "inbatch_pos_score": 3.0703, "learning_rate": 2.386842105263158e-05, "loss": 3.4272, "norm_diff": 0.0352, "norm_loss": 0.0, "num_token_doc": 66.8159, "num_token_overlap": 11.6866, "num_token_query": 31.3599, "num_token_union": 65.1021, "num_word_context": 202.2348, "num_word_doc": 49.8285, "num_word_query": 23.3003, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19869.6308, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.459, "query_norm": 2.0241, "queue_k_norm": 2.0579, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3599, "sent_len_1": 66.8159, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.6813, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 109300 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.4314, "doc_norm": 2.0632, "encoder_q-embeddings": 9841.2666, "encoder_q-layer.0": 6758.1704, "encoder_q-layer.1": 7045.1978, "encoder_q-layer.10": 12798.5176, "encoder_q-layer.11": 29204.9609, "encoder_q-layer.2": 7712.4761, "encoder_q-layer.3": 7858.6948, "encoder_q-layer.4": 8481.9082, "encoder_q-layer.5": 8846.6787, "encoder_q-layer.6": 9879.4512, "encoder_q-layer.7": 11112.752, "encoder_q-layer.8": 13224.2363, "encoder_q-layer.9": 12313.3691, "epoch": 0.47, "inbatch_neg_score": 2.4688, "inbatch_pos_score": 3.0898, "learning_rate": 2.3842105263157897e-05, "loss": 3.4314, "norm_diff": 0.0413, "norm_loss": 0.0, "num_token_doc": 66.779, "num_token_overlap": 11.6694, "num_token_query": 31.4095, "num_token_union": 65.1274, "num_word_context": 202.3117, "num_word_doc": 49.8273, "num_word_query": 23.3266, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19065.295, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4688, "query_norm": 2.0219, "queue_k_norm": 2.0604, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4095, "sent_len_1": 66.779, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.0975, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 109400 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.4274, "doc_norm": 2.0666, "encoder_q-embeddings": 9520.334, "encoder_q-layer.0": 6650.4482, "encoder_q-layer.1": 7329.6338, "encoder_q-layer.10": 13929.9277, "encoder_q-layer.11": 31408.9746, "encoder_q-layer.2": 8088.9863, "encoder_q-layer.3": 8262.0293, "encoder_q-layer.4": 9004.2715, "encoder_q-layer.5": 8835.1846, "encoder_q-layer.6": 10008.4131, "encoder_q-layer.7": 11324.0127, "encoder_q-layer.8": 13184.9746, "encoder_q-layer.9": 12990.0586, "epoch": 0.48, "inbatch_neg_score": 2.4733, "inbatch_pos_score": 3.0781, "learning_rate": 2.3815789473684212e-05, "loss": 3.4274, "norm_diff": 0.0474, "norm_loss": 0.0, "num_token_doc": 66.8319, "num_token_overlap": 11.6753, "num_token_query": 31.3295, "num_token_union": 65.0862, "num_word_context": 202.2864, "num_word_doc": 49.8274, "num_word_query": 23.2566, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19313.4101, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4727, "query_norm": 2.0192, "queue_k_norm": 2.0619, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3295, "sent_len_1": 66.8319, "sent_len_max_0": 127.975, "sent_len_max_1": 210.7612, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 109500 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4331, "doc_norm": 2.0661, "encoder_q-embeddings": 10541.9863, "encoder_q-layer.0": 7054.0518, "encoder_q-layer.1": 7580.4619, "encoder_q-layer.10": 13214.1914, "encoder_q-layer.11": 29523.8633, "encoder_q-layer.2": 8520.6221, "encoder_q-layer.3": 8728.3545, "encoder_q-layer.4": 9469.4854, "encoder_q-layer.5": 9773.0029, "encoder_q-layer.6": 10655.6221, "encoder_q-layer.7": 11581.167, "encoder_q-layer.8": 14117.8604, "encoder_q-layer.9": 12232.375, "epoch": 0.48, "inbatch_neg_score": 2.4739, "inbatch_pos_score": 3.084, "learning_rate": 2.3789473684210527e-05, "loss": 3.4331, "norm_diff": 0.0497, "norm_loss": 0.0, "num_token_doc": 66.8949, "num_token_overlap": 11.6802, "num_token_query": 31.2789, "num_token_union": 65.1258, "num_word_context": 202.3769, "num_word_doc": 49.9163, "num_word_query": 23.2235, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19401.8149, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4727, "query_norm": 2.0164, "queue_k_norm": 2.0653, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2789, "sent_len_1": 66.8949, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.6262, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 109600 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4328, "doc_norm": 2.067, "encoder_q-embeddings": 10168.3623, "encoder_q-layer.0": 6834.2793, "encoder_q-layer.1": 7507.6851, "encoder_q-layer.10": 12714.332, "encoder_q-layer.11": 32057.7031, "encoder_q-layer.2": 8520.04, "encoder_q-layer.3": 8639.4854, "encoder_q-layer.4": 9447.9443, "encoder_q-layer.5": 9443.8389, "encoder_q-layer.6": 10498.291, "encoder_q-layer.7": 11741.4727, "encoder_q-layer.8": 13934.3076, "encoder_q-layer.9": 12559.7432, "epoch": 0.48, "inbatch_neg_score": 2.4757, "inbatch_pos_score": 3.0723, "learning_rate": 2.3763157894736842e-05, "loss": 3.4328, "norm_diff": 0.0627, "norm_loss": 0.0, "num_token_doc": 66.7608, "num_token_overlap": 11.6581, "num_token_query": 31.2922, "num_token_union": 65.0834, "num_word_context": 202.1676, "num_word_doc": 49.8229, "num_word_query": 23.2446, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20177.8906, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4746, "query_norm": 2.0043, "queue_k_norm": 2.0658, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2922, "sent_len_1": 66.7608, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.5975, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 109700 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4301, "doc_norm": 2.064, "encoder_q-embeddings": 10197.9307, "encoder_q-layer.0": 7030.5259, "encoder_q-layer.1": 7802.8169, "encoder_q-layer.10": 12772.3848, "encoder_q-layer.11": 29292.373, "encoder_q-layer.2": 8482.0029, "encoder_q-layer.3": 9033.1162, "encoder_q-layer.4": 9529.3213, "encoder_q-layer.5": 10160.6055, "encoder_q-layer.6": 11519.8076, "encoder_q-layer.7": 11704.9248, "encoder_q-layer.8": 13625.416, "encoder_q-layer.9": 12290.9414, "epoch": 0.48, "inbatch_neg_score": 2.4779, "inbatch_pos_score": 3.0762, "learning_rate": 2.373684210526316e-05, "loss": 3.4301, "norm_diff": 0.0649, "norm_loss": 0.0, "num_token_doc": 66.7941, "num_token_overlap": 11.6819, "num_token_query": 31.3674, "num_token_union": 65.0976, "num_word_context": 202.3424, "num_word_doc": 49.8157, "num_word_query": 23.3024, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19665.6325, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4766, "query_norm": 1.999, "queue_k_norm": 2.0686, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3674, "sent_len_1": 66.7941, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.675, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 109800 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.435, "doc_norm": 2.0668, "encoder_q-embeddings": 10859.5381, "encoder_q-layer.0": 7371.5801, "encoder_q-layer.1": 8028.2212, "encoder_q-layer.10": 13338.5547, "encoder_q-layer.11": 32781.1406, "encoder_q-layer.2": 9365.8633, "encoder_q-layer.3": 9726.8428, "encoder_q-layer.4": 10113.4863, "encoder_q-layer.5": 10515.1689, "encoder_q-layer.6": 11440.0791, "encoder_q-layer.7": 12521.0674, "encoder_q-layer.8": 14723.3066, "encoder_q-layer.9": 13009.9785, "epoch": 0.48, "inbatch_neg_score": 2.4788, "inbatch_pos_score": 3.0625, "learning_rate": 2.3710526315789475e-05, "loss": 3.435, "norm_diff": 0.0737, "norm_loss": 0.0, "num_token_doc": 66.7531, "num_token_overlap": 11.6475, "num_token_query": 31.3249, "num_token_union": 65.0554, "num_word_context": 202.1123, "num_word_doc": 49.7848, "num_word_query": 23.2745, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21580.1338, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4785, "query_norm": 1.9932, "queue_k_norm": 2.0687, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3249, "sent_len_1": 66.7531, "sent_len_max_0": 127.9963, "sent_len_max_1": 211.7438, "stdk": 0.0491, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 109900 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.4322, "doc_norm": 2.0732, "encoder_q-embeddings": 9593.2139, "encoder_q-layer.0": 6807.1152, "encoder_q-layer.1": 7271.6621, "encoder_q-layer.10": 13311.2246, "encoder_q-layer.11": 32265.4727, "encoder_q-layer.2": 8062.9829, "encoder_q-layer.3": 8335.8955, "encoder_q-layer.4": 9218.1514, "encoder_q-layer.5": 9045.8281, "encoder_q-layer.6": 10580.2168, "encoder_q-layer.7": 12384.7793, "encoder_q-layer.8": 15262.585, "encoder_q-layer.9": 13420.6309, "epoch": 0.48, "inbatch_neg_score": 2.4785, "inbatch_pos_score": 3.082, "learning_rate": 2.368421052631579e-05, "loss": 3.4322, "norm_diff": 0.0747, "norm_loss": 0.0, "num_token_doc": 66.7948, "num_token_overlap": 11.6887, "num_token_query": 31.3855, "num_token_union": 65.1404, "num_word_context": 202.2149, "num_word_doc": 49.842, "num_word_query": 23.3047, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20134.1328, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4785, "query_norm": 1.9985, "queue_k_norm": 2.0717, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3855, "sent_len_1": 66.7948, "sent_len_max_0": 127.985, "sent_len_max_1": 208.9288, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 110000 }, { "dev_runtime": 28.7174, "dev_samples_per_second": 1.114, "dev_steps_per_second": 0.035, "epoch": 0.48, "step": 110000, "test_accuracy": 94.00634765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3420425057411194, "test_doc_norm": 2.0521957874298096, "test_inbatch_neg_score": 2.7644851207733154, "test_inbatch_pos_score": 3.7816622257232666, "test_loss": 0.3420425057411194, "test_loss_align": 1.0228073596954346, "test_loss_unif": -7.841700553894043, "test_loss_unif_q@queue": -7.841700553894043, "test_norm_diff": 0.0356263667345047, "test_norm_loss": 0.0, "test_q@queue_neg_score": 2.4786529541015625, "test_query_norm": 2.0878219604492188, "test_queue_k_norm": 2.0717804431915283, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04410577565431595, "test_stdq": 0.0439697802066803, "test_stdqueue_k": 0.049378957599401474, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.7174, "dev_samples_per_second": 1.114, "dev_steps_per_second": 0.035, "epoch": 0.48, "eval_beir-arguana_ndcg@10": 0.39232, "eval_beir-arguana_recall@10": 0.66074, "eval_beir-arguana_recall@100": 0.92674, "eval_beir-arguana_recall@20": 0.78663, "eval_beir-avg_ndcg@10": 0.39725274999999993, "eval_beir-avg_recall@10": 0.46700175, "eval_beir-avg_recall@100": 0.6435821666666668, "eval_beir-avg_recall@20": 0.52450575, "eval_beir-cqadupstack_ndcg@10": 0.2840175, "eval_beir-cqadupstack_recall@10": 0.38329750000000007, "eval_beir-cqadupstack_recall@100": 0.6198616666666668, "eval_beir-cqadupstack_recall@20": 0.4506275, "eval_beir-fiqa_ndcg@10": 0.26428, "eval_beir-fiqa_recall@10": 0.33085, "eval_beir-fiqa_recall@100": 0.60649, "eval_beir-fiqa_recall@20": 0.39738, "eval_beir-nfcorpus_ndcg@10": 0.31539, "eval_beir-nfcorpus_recall@10": 0.15548, "eval_beir-nfcorpus_recall@100": 0.2917, "eval_beir-nfcorpus_recall@20": 0.1867, "eval_beir-nq_ndcg@10": 0.32042, "eval_beir-nq_recall@10": 0.50857, "eval_beir-nq_recall@100": 0.83126, "eval_beir-nq_recall@20": 0.62548, "eval_beir-quora_ndcg@10": 0.78265, "eval_beir-quora_recall@10": 0.89118, "eval_beir-quora_recall@100": 0.97847, "eval_beir-quora_recall@20": 0.93034, "eval_beir-scidocs_ndcg@10": 0.15971, "eval_beir-scidocs_recall@10": 0.16972, "eval_beir-scidocs_recall@100": 0.37242, "eval_beir-scidocs_recall@20": 0.22612, "eval_beir-scifact_ndcg@10": 0.66032, "eval_beir-scifact_recall@10": 0.79133, "eval_beir-scifact_recall@100": 0.91989, "eval_beir-scifact_recall@20": 0.82644, "eval_beir-trec-covid_ndcg@10": 0.60827, "eval_beir-trec-covid_recall@10": 0.646, "eval_beir-trec-covid_recall@100": 0.4634, "eval_beir-trec-covid_recall@20": 0.608, "eval_beir-webis-touche2020_ndcg@10": 0.18515, "eval_beir-webis-touche2020_recall@10": 0.13285, "eval_beir-webis-touche2020_recall@100": 0.42559, "eval_beir-webis-touche2020_recall@20": 0.20734, "eval_senteval-avg_sts": 0.7378331437254658, "eval_senteval-sickr_spearman": 0.7251182376837683, "eval_senteval-stsb_spearman": 0.7505480497671634, "step": 110000, "test_accuracy": 94.00634765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3420425057411194, "test_doc_norm": 2.0521957874298096, "test_inbatch_neg_score": 2.7644851207733154, "test_inbatch_pos_score": 3.7816622257232666, "test_loss": 0.3420425057411194, "test_loss_align": 1.0228073596954346, "test_loss_unif": -7.841700553894043, "test_loss_unif_q@queue": -7.841700553894043, "test_norm_diff": 0.0356263667345047, "test_norm_loss": 0.0, "test_q@queue_neg_score": 2.4786529541015625, "test_query_norm": 2.0878219604492188, "test_queue_k_norm": 2.0717804431915283, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04410577565431595, "test_stdq": 0.0439697802066803, "test_stdqueue_k": 0.049378957599401474, "test_stdqueue_q": 0.0 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.418, "doc_norm": 2.0737, "encoder_q-embeddings": 9549.1523, "encoder_q-layer.0": 6440.2559, "encoder_q-layer.1": 6780.7622, "encoder_q-layer.10": 12458.9756, "encoder_q-layer.11": 30263.1191, "encoder_q-layer.2": 7756.6265, "encoder_q-layer.3": 8147.6099, "encoder_q-layer.4": 8715.6533, "encoder_q-layer.5": 8834.6572, "encoder_q-layer.6": 9878.8643, "encoder_q-layer.7": 11999.0996, "encoder_q-layer.8": 14074.4053, "encoder_q-layer.9": 12437.9014, "epoch": 0.48, "inbatch_neg_score": 2.4759, "inbatch_pos_score": 3.0859, "learning_rate": 2.3657894736842105e-05, "loss": 3.418, "norm_diff": 0.0772, "norm_loss": 0.0, "num_token_doc": 66.6738, "num_token_overlap": 11.7214, "num_token_query": 31.5299, "num_token_union": 65.1231, "num_word_context": 202.7617, "num_word_doc": 49.751, "num_word_query": 23.4317, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19495.4814, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4766, "query_norm": 1.9965, "queue_k_norm": 2.0724, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5299, "sent_len_1": 66.6738, "sent_len_max_0": 127.99, "sent_len_max_1": 209.8275, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 110100 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4431, "doc_norm": 2.0743, "encoder_q-embeddings": 9555.3604, "encoder_q-layer.0": 6463.2925, "encoder_q-layer.1": 7175.4375, "encoder_q-layer.10": 12755.4316, "encoder_q-layer.11": 30018.0312, "encoder_q-layer.2": 7891.3472, "encoder_q-layer.3": 8304.8369, "encoder_q-layer.4": 8702.8857, "encoder_q-layer.5": 9152.4648, "encoder_q-layer.6": 10323.5918, "encoder_q-layer.7": 11302.4297, "encoder_q-layer.8": 13057.8369, "encoder_q-layer.9": 12285.4102, "epoch": 0.48, "inbatch_neg_score": 2.4792, "inbatch_pos_score": 3.0859, "learning_rate": 2.363157894736842e-05, "loss": 3.4431, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.7172, "num_token_overlap": 11.6784, "num_token_query": 31.3857, "num_token_union": 65.0779, "num_word_context": 202.2687, "num_word_doc": 49.7653, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19116.2155, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4785, "query_norm": 1.9897, "queue_k_norm": 2.0747, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3857, "sent_len_1": 66.7172, "sent_len_max_0": 127.995, "sent_len_max_1": 209.7138, "stdk": 0.0493, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 110200 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.4281, "doc_norm": 2.0742, "encoder_q-embeddings": 9974.5127, "encoder_q-layer.0": 6812.6523, "encoder_q-layer.1": 7129.0737, "encoder_q-layer.10": 12599.4023, "encoder_q-layer.11": 30910.2539, "encoder_q-layer.2": 8043.2598, "encoder_q-layer.3": 8096.1924, "encoder_q-layer.4": 8716.3672, "encoder_q-layer.5": 9165.8213, "encoder_q-layer.6": 10612.9609, "encoder_q-layer.7": 11180.8984, "encoder_q-layer.8": 13215.4668, "encoder_q-layer.9": 12324.4346, "epoch": 0.48, "inbatch_neg_score": 2.4783, "inbatch_pos_score": 3.0879, "learning_rate": 2.360526315789474e-05, "loss": 3.4281, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.7487, "num_token_overlap": 11.6915, "num_token_query": 31.4444, "num_token_union": 65.1239, "num_word_context": 202.4936, "num_word_doc": 49.841, "num_word_query": 23.3683, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19380.2845, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4785, "query_norm": 1.9966, "queue_k_norm": 2.0749, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4444, "sent_len_1": 66.7487, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.8113, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 110300 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4205, "doc_norm": 2.0749, "encoder_q-embeddings": 9431.8672, "encoder_q-layer.0": 6602.7461, "encoder_q-layer.1": 6978.9902, "encoder_q-layer.10": 12365.0908, "encoder_q-layer.11": 29899.2832, "encoder_q-layer.2": 7670.0933, "encoder_q-layer.3": 7838.25, "encoder_q-layer.4": 8445.6387, "encoder_q-layer.5": 8364.7568, "encoder_q-layer.6": 9335.5645, "encoder_q-layer.7": 10606.5107, "encoder_q-layer.8": 12624.0059, "encoder_q-layer.9": 11988.0557, "epoch": 0.48, "inbatch_neg_score": 2.4803, "inbatch_pos_score": 3.0938, "learning_rate": 2.3578947368421054e-05, "loss": 3.4205, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7226, "num_token_overlap": 11.6582, "num_token_query": 31.3597, "num_token_union": 65.1056, "num_word_context": 202.0499, "num_word_doc": 49.8004, "num_word_query": 23.2937, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18557.381, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4805, "query_norm": 1.9988, "queue_k_norm": 2.0752, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3597, "sent_len_1": 66.7226, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.4875, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 110400 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4126, "doc_norm": 2.0783, "encoder_q-embeddings": 10619.6084, "encoder_q-layer.0": 7476.1812, "encoder_q-layer.1": 7865.5513, "encoder_q-layer.10": 14013.9727, "encoder_q-layer.11": 31574.2422, "encoder_q-layer.2": 8746.6016, "encoder_q-layer.3": 9101.582, "encoder_q-layer.4": 9592.7188, "encoder_q-layer.5": 9563.7275, "encoder_q-layer.6": 10754.6943, "encoder_q-layer.7": 12160.6104, "encoder_q-layer.8": 14512.2539, "encoder_q-layer.9": 13551.835, "epoch": 0.48, "inbatch_neg_score": 2.4816, "inbatch_pos_score": 3.0938, "learning_rate": 2.3552631578947372e-05, "loss": 3.4126, "norm_diff": 0.0711, "norm_loss": 0.0, "num_token_doc": 66.633, "num_token_overlap": 11.687, "num_token_query": 31.3969, "num_token_union": 65.0289, "num_word_context": 202.232, "num_word_doc": 49.7447, "num_word_query": 23.3257, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20557.0156, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4805, "query_norm": 2.0072, "queue_k_norm": 2.0754, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3969, "sent_len_1": 66.633, "sent_len_max_0": 127.985, "sent_len_max_1": 210.1188, "stdk": 0.0494, "stdq": 0.0439, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 110500 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.4216, "doc_norm": 2.0807, "encoder_q-embeddings": 10014.5439, "encoder_q-layer.0": 6929.2549, "encoder_q-layer.1": 7501.3179, "encoder_q-layer.10": 12612.4121, "encoder_q-layer.11": 29319.9238, "encoder_q-layer.2": 8518.332, "encoder_q-layer.3": 8628.6777, "encoder_q-layer.4": 9461.749, "encoder_q-layer.5": 9638.4199, "encoder_q-layer.6": 10120.6855, "encoder_q-layer.7": 11353.6611, "encoder_q-layer.8": 13414.2617, "encoder_q-layer.9": 12235.6211, "epoch": 0.48, "inbatch_neg_score": 2.4826, "inbatch_pos_score": 3.0977, "learning_rate": 2.3526315789473684e-05, "loss": 3.4216, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.9554, "num_token_overlap": 11.6887, "num_token_query": 31.3696, "num_token_union": 65.1949, "num_word_context": 202.273, "num_word_doc": 49.9564, "num_word_query": 23.2869, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19306.2884, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4824, "query_norm": 2.0028, "queue_k_norm": 2.079, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3696, "sent_len_1": 66.9554, "sent_len_max_0": 127.9663, "sent_len_max_1": 211.4087, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 110600 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.4259, "doc_norm": 2.0769, "encoder_q-embeddings": 10381.0039, "encoder_q-layer.0": 7165.481, "encoder_q-layer.1": 7630.0742, "encoder_q-layer.10": 13316.3076, "encoder_q-layer.11": 29659.4102, "encoder_q-layer.2": 8784.5596, "encoder_q-layer.3": 9143.0029, "encoder_q-layer.4": 10006.1094, "encoder_q-layer.5": 10031.6729, "encoder_q-layer.6": 10743.627, "encoder_q-layer.7": 11802.6426, "encoder_q-layer.8": 13519.2227, "encoder_q-layer.9": 12888.498, "epoch": 0.48, "inbatch_neg_score": 2.4846, "inbatch_pos_score": 3.0781, "learning_rate": 2.35e-05, "loss": 3.4259, "norm_diff": 0.0799, "norm_loss": 0.0, "num_token_doc": 66.7686, "num_token_overlap": 11.651, "num_token_query": 31.294, "num_token_union": 65.0779, "num_word_context": 202.2735, "num_word_doc": 49.8226, "num_word_query": 23.2422, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19505.3254, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4844, "query_norm": 1.997, "queue_k_norm": 2.0772, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.294, "sent_len_1": 66.7686, "sent_len_max_0": 127.99, "sent_len_max_1": 210.7138, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 110700 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4271, "doc_norm": 2.0855, "encoder_q-embeddings": 10281.6387, "encoder_q-layer.0": 6977.8901, "encoder_q-layer.1": 7695.0244, "encoder_q-layer.10": 12896.5771, "encoder_q-layer.11": 32116.2305, "encoder_q-layer.2": 8604.7471, "encoder_q-layer.3": 9127.5322, "encoder_q-layer.4": 9576.583, "encoder_q-layer.5": 9662.5742, "encoder_q-layer.6": 10695.8154, "encoder_q-layer.7": 11603.3398, "encoder_q-layer.8": 13850.9707, "encoder_q-layer.9": 13077.1768, "epoch": 0.48, "inbatch_neg_score": 2.4859, "inbatch_pos_score": 3.0918, "learning_rate": 2.3473684210526317e-05, "loss": 3.4271, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.6761, "num_token_overlap": 11.6745, "num_token_query": 31.3312, "num_token_union": 65.0381, "num_word_context": 202.1842, "num_word_doc": 49.7697, "num_word_query": 23.2795, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20280.6463, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4863, "query_norm": 2.0029, "queue_k_norm": 2.0804, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3312, "sent_len_1": 66.6761, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.4737, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 110800 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.4097, "doc_norm": 2.083, "encoder_q-embeddings": 10630.4326, "encoder_q-layer.0": 7340.8281, "encoder_q-layer.1": 7781.9209, "encoder_q-layer.10": 13020.5127, "encoder_q-layer.11": 31973.5312, "encoder_q-layer.2": 8686.8633, "encoder_q-layer.3": 8807.6836, "encoder_q-layer.4": 9600.4795, "encoder_q-layer.5": 9618.8516, "encoder_q-layer.6": 10723.5898, "encoder_q-layer.7": 12353.7988, "encoder_q-layer.8": 13569.4932, "encoder_q-layer.9": 12627.8535, "epoch": 0.48, "inbatch_neg_score": 2.4921, "inbatch_pos_score": 3.1094, "learning_rate": 2.3447368421052632e-05, "loss": 3.4097, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.8908, "num_token_overlap": 11.7122, "num_token_query": 31.386, "num_token_union": 65.1402, "num_word_context": 202.1339, "num_word_doc": 49.8944, "num_word_query": 23.3164, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19908.9593, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.4922, "query_norm": 2.0126, "queue_k_norm": 2.0808, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.386, "sent_len_1": 66.8908, "sent_len_max_0": 127.99, "sent_len_max_1": 210.6925, "stdk": 0.0494, "stdq": 0.0438, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 110900 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.4344, "doc_norm": 2.078, "encoder_q-embeddings": 19967.9297, "encoder_q-layer.0": 13445.248, "encoder_q-layer.1": 14771.4443, "encoder_q-layer.10": 26447.7969, "encoder_q-layer.11": 65742.3125, "encoder_q-layer.2": 15912.9678, "encoder_q-layer.3": 16276.2783, "encoder_q-layer.4": 17408.2715, "encoder_q-layer.5": 17679.4102, "encoder_q-layer.6": 19160.6191, "encoder_q-layer.7": 22556.1348, "encoder_q-layer.8": 28091.75, "encoder_q-layer.9": 25118.7051, "epoch": 0.48, "inbatch_neg_score": 2.4983, "inbatch_pos_score": 3.0977, "learning_rate": 2.342105263157895e-05, "loss": 3.4344, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.8588, "num_token_overlap": 11.6273, "num_token_query": 31.2998, "num_token_union": 65.1644, "num_word_context": 202.3757, "num_word_doc": 49.8928, "num_word_query": 23.2477, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39786.4237, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.498, "query_norm": 2.0022, "queue_k_norm": 2.0816, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2998, "sent_len_1": 66.8588, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.79, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 111000 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.4252, "doc_norm": 2.0862, "encoder_q-embeddings": 20064.6758, "encoder_q-layer.0": 13561.6943, "encoder_q-layer.1": 14204.4434, "encoder_q-layer.10": 26467.7305, "encoder_q-layer.11": 63058.0547, "encoder_q-layer.2": 16326.8135, "encoder_q-layer.3": 16966.1621, "encoder_q-layer.4": 18120.9648, "encoder_q-layer.5": 18443.8027, "encoder_q-layer.6": 20356.1309, "encoder_q-layer.7": 22975.2148, "encoder_q-layer.8": 27268.6172, "encoder_q-layer.9": 24618.0469, "epoch": 0.48, "inbatch_neg_score": 2.5011, "inbatch_pos_score": 3.0957, "learning_rate": 2.3394736842105265e-05, "loss": 3.4252, "norm_diff": 0.0848, "norm_loss": 0.0, "num_token_doc": 66.8435, "num_token_overlap": 11.6639, "num_token_query": 31.3187, "num_token_union": 65.1483, "num_word_context": 202.5453, "num_word_doc": 49.9065, "num_word_query": 23.2624, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39844.8733, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.502, "query_norm": 2.0014, "queue_k_norm": 2.0835, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3187, "sent_len_1": 66.8435, "sent_len_max_0": 127.9437, "sent_len_max_1": 208.7325, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 111100 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4247, "doc_norm": 2.0829, "encoder_q-embeddings": 19529.1328, "encoder_q-layer.0": 13457.5996, "encoder_q-layer.1": 14327.4258, "encoder_q-layer.10": 25801.0254, "encoder_q-layer.11": 63595.0469, "encoder_q-layer.2": 16011.4258, "encoder_q-layer.3": 16700.2949, "encoder_q-layer.4": 18036.9551, "encoder_q-layer.5": 18608.4102, "encoder_q-layer.6": 19912.7988, "encoder_q-layer.7": 22889.2539, "encoder_q-layer.8": 27707.6758, "encoder_q-layer.9": 26296.2129, "epoch": 0.48, "inbatch_neg_score": 2.5076, "inbatch_pos_score": 3.1016, "learning_rate": 2.336842105263158e-05, "loss": 3.4247, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.818, "num_token_overlap": 11.6587, "num_token_query": 31.4032, "num_token_union": 65.2065, "num_word_context": 202.6287, "num_word_doc": 49.875, "num_word_query": 23.3133, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39769.6762, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5078, "query_norm": 2.0089, "queue_k_norm": 2.0834, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4032, "sent_len_1": 66.818, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.8237, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111200 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4229, "doc_norm": 2.0851, "encoder_q-embeddings": 20107.4004, "encoder_q-layer.0": 14253.2236, "encoder_q-layer.1": 14905.4736, "encoder_q-layer.10": 29363.6484, "encoder_q-layer.11": 59738.6172, "encoder_q-layer.2": 17309.2148, "encoder_q-layer.3": 17541.5566, "encoder_q-layer.4": 19487.9609, "encoder_q-layer.5": 19781.3906, "encoder_q-layer.6": 22239.4512, "encoder_q-layer.7": 25583.8457, "encoder_q-layer.8": 28556.9082, "encoder_q-layer.9": 25084.7422, "epoch": 0.48, "inbatch_neg_score": 2.5143, "inbatch_pos_score": 3.1289, "learning_rate": 2.3342105263157895e-05, "loss": 3.4229, "norm_diff": 0.065, "norm_loss": 0.0, "num_token_doc": 66.7641, "num_token_overlap": 11.6839, "num_token_query": 31.4289, "num_token_union": 65.1413, "num_word_context": 202.0146, "num_word_doc": 49.8157, "num_word_query": 23.3363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39492.547, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5137, "query_norm": 2.0201, "queue_k_norm": 2.0832, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4289, "sent_len_1": 66.7641, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.2788, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111300 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.4311, "doc_norm": 2.0864, "encoder_q-embeddings": 20801.8145, "encoder_q-layer.0": 14826.4199, "encoder_q-layer.1": 15887.8496, "encoder_q-layer.10": 30466.8652, "encoder_q-layer.11": 65917.4062, "encoder_q-layer.2": 18338.9043, "encoder_q-layer.3": 19372.8203, "encoder_q-layer.4": 20487.8047, "encoder_q-layer.5": 20736.2871, "encoder_q-layer.6": 22006.4297, "encoder_q-layer.7": 24785.5117, "encoder_q-layer.8": 28543.2012, "encoder_q-layer.9": 26105.4238, "epoch": 0.48, "inbatch_neg_score": 2.5229, "inbatch_pos_score": 3.1172, "learning_rate": 2.331578947368421e-05, "loss": 3.4311, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.6723, "num_token_overlap": 11.6212, "num_token_query": 31.3248, "num_token_union": 65.089, "num_word_context": 202.323, "num_word_doc": 49.758, "num_word_query": 23.2638, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42641.0468, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5234, "query_norm": 2.0166, "queue_k_norm": 2.084, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3248, "sent_len_1": 66.6723, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.7388, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111400 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4284, "doc_norm": 2.0825, "encoder_q-embeddings": 19635.4531, "encoder_q-layer.0": 13573.0654, "encoder_q-layer.1": 14839.2803, "encoder_q-layer.10": 24641.3066, "encoder_q-layer.11": 60938.6367, "encoder_q-layer.2": 16109.6504, "encoder_q-layer.3": 16989.2188, "encoder_q-layer.4": 17834.0566, "encoder_q-layer.5": 17395.5566, "encoder_q-layer.6": 19752.6777, "encoder_q-layer.7": 22630.1562, "encoder_q-layer.8": 26642.5488, "encoder_q-layer.9": 24051.0645, "epoch": 0.48, "inbatch_neg_score": 2.5319, "inbatch_pos_score": 3.1328, "learning_rate": 2.328947368421053e-05, "loss": 3.4284, "norm_diff": 0.0578, "norm_loss": 0.0, "num_token_doc": 66.7645, "num_token_overlap": 11.6565, "num_token_query": 31.4061, "num_token_union": 65.135, "num_word_context": 202.2516, "num_word_doc": 49.7967, "num_word_query": 23.3203, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38381.4125, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5312, "query_norm": 2.0246, "queue_k_norm": 2.0861, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4061, "sent_len_1": 66.7645, "sent_len_max_0": 127.97, "sent_len_max_1": 210.225, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111500 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.4348, "doc_norm": 2.0879, "encoder_q-embeddings": 20653.582, "encoder_q-layer.0": 14573.7197, "encoder_q-layer.1": 15789.4287, "encoder_q-layer.10": 25526.7324, "encoder_q-layer.11": 60676.7695, "encoder_q-layer.2": 17716.8184, "encoder_q-layer.3": 18672.6602, "encoder_q-layer.4": 20165.4805, "encoder_q-layer.5": 20656.3789, "encoder_q-layer.6": 22040.998, "encoder_q-layer.7": 23542.3047, "encoder_q-layer.8": 27438.5391, "encoder_q-layer.9": 25032.3281, "epoch": 0.48, "inbatch_neg_score": 2.5381, "inbatch_pos_score": 3.1504, "learning_rate": 2.3263157894736844e-05, "loss": 3.4348, "norm_diff": 0.0553, "norm_loss": 0.0, "num_token_doc": 66.7685, "num_token_overlap": 11.671, "num_token_query": 31.3288, "num_token_union": 65.0845, "num_word_context": 202.2861, "num_word_doc": 49.8512, "num_word_query": 23.2639, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40100.243, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5371, "query_norm": 2.0326, "queue_k_norm": 2.0868, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3288, "sent_len_1": 66.7685, "sent_len_max_0": 127.975, "sent_len_max_1": 208.7262, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111600 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.4337, "doc_norm": 2.0875, "encoder_q-embeddings": 20104.5332, "encoder_q-layer.0": 13894.7588, "encoder_q-layer.1": 14859.6689, "encoder_q-layer.10": 25437.6719, "encoder_q-layer.11": 60857.1289, "encoder_q-layer.2": 16761.4531, "encoder_q-layer.3": 16925.1387, "encoder_q-layer.4": 18051.6758, "encoder_q-layer.5": 18576.4473, "encoder_q-layer.6": 20815.3086, "encoder_q-layer.7": 23734.1328, "encoder_q-layer.8": 26581.3027, "encoder_q-layer.9": 23873.7383, "epoch": 0.48, "inbatch_neg_score": 2.5434, "inbatch_pos_score": 3.1562, "learning_rate": 2.323684210526316e-05, "loss": 3.4337, "norm_diff": 0.0503, "norm_loss": 0.0, "num_token_doc": 66.8126, "num_token_overlap": 11.6571, "num_token_query": 31.3429, "num_token_union": 65.1207, "num_word_context": 202.654, "num_word_doc": 49.8723, "num_word_query": 23.2904, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38877.1091, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.543, "query_norm": 2.0372, "queue_k_norm": 2.0883, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3429, "sent_len_1": 66.8126, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.7113, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111700 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4228, "doc_norm": 2.0891, "encoder_q-embeddings": 19602.2598, "encoder_q-layer.0": 13133.7051, "encoder_q-layer.1": 14294.5059, "encoder_q-layer.10": 25932.3008, "encoder_q-layer.11": 61418.5625, "encoder_q-layer.2": 15848.3379, "encoder_q-layer.3": 15931.4258, "encoder_q-layer.4": 17821.127, "encoder_q-layer.5": 18187.7734, "encoder_q-layer.6": 19816.6055, "encoder_q-layer.7": 23071.4414, "encoder_q-layer.8": 27216.4199, "encoder_q-layer.9": 24497.4512, "epoch": 0.49, "inbatch_neg_score": 2.5576, "inbatch_pos_score": 3.1484, "learning_rate": 2.3210526315789473e-05, "loss": 3.4228, "norm_diff": 0.0523, "norm_loss": 0.0, "num_token_doc": 66.649, "num_token_overlap": 11.6696, "num_token_query": 31.4535, "num_token_union": 65.0984, "num_word_context": 202.3016, "num_word_doc": 49.73, "num_word_query": 23.3602, "postclip_grad_norm": 1.0, "preclip_grad_norm": 38673.4174, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5566, "query_norm": 2.0368, "queue_k_norm": 2.0891, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4535, "sent_len_1": 66.649, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.4437, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111800 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.422, "doc_norm": 2.0911, "encoder_q-embeddings": 20233.6191, "encoder_q-layer.0": 13686.7119, "encoder_q-layer.1": 14931.3311, "encoder_q-layer.10": 26349.1426, "encoder_q-layer.11": 62165.2656, "encoder_q-layer.2": 16586.7461, "encoder_q-layer.3": 16757.543, "encoder_q-layer.4": 18235.166, "encoder_q-layer.5": 18419.043, "encoder_q-layer.6": 20578.7324, "encoder_q-layer.7": 23611.4902, "encoder_q-layer.8": 28181.8516, "encoder_q-layer.9": 25589.7383, "epoch": 0.49, "inbatch_neg_score": 2.5653, "inbatch_pos_score": 3.168, "learning_rate": 2.3184210526315792e-05, "loss": 3.422, "norm_diff": 0.0443, "norm_loss": 0.0, "num_token_doc": 66.8245, "num_token_overlap": 11.7273, "num_token_query": 31.5078, "num_token_union": 65.2109, "num_word_context": 202.5826, "num_word_doc": 49.8751, "num_word_query": 23.4084, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40111.5885, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 2.5645, "query_norm": 2.0469, "queue_k_norm": 2.0919, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5078, "sent_len_1": 66.8245, "sent_len_max_0": 127.99, "sent_len_max_1": 208.61, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 111900 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.42, "doc_norm": 2.0918, "encoder_q-embeddings": 18681.6582, "encoder_q-layer.0": 12865.29, "encoder_q-layer.1": 13433.2646, "encoder_q-layer.10": 26646.1582, "encoder_q-layer.11": 62492.5703, "encoder_q-layer.2": 14913.4199, "encoder_q-layer.3": 14976.7363, "encoder_q-layer.4": 16138.751, "encoder_q-layer.5": 16501.0781, "encoder_q-layer.6": 19014.7891, "encoder_q-layer.7": 21666.9531, "encoder_q-layer.8": 25955.6758, "encoder_q-layer.9": 26594.3066, "epoch": 0.49, "inbatch_neg_score": 2.5834, "inbatch_pos_score": 3.1973, "learning_rate": 2.3157894736842107e-05, "loss": 3.42, "norm_diff": 0.0376, "norm_loss": 0.0, "num_token_doc": 66.6109, "num_token_overlap": 11.671, "num_token_query": 31.4442, "num_token_union": 65.0943, "num_word_context": 202.0945, "num_word_doc": 49.7192, "num_word_query": 23.3435, "postclip_grad_norm": 1.0, "preclip_grad_norm": 37698.2868, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 2.584, "query_norm": 2.0542, "queue_k_norm": 2.0936, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4442, "sent_len_1": 66.6109, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.445, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 112000 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.4159, "doc_norm": 2.0961, "encoder_q-embeddings": 12121.0205, "encoder_q-layer.0": 8522.6816, "encoder_q-layer.1": 9599.7764, "encoder_q-layer.10": 12939.8545, "encoder_q-layer.11": 30391.7188, "encoder_q-layer.2": 10733.2461, "encoder_q-layer.3": 11228.9805, "encoder_q-layer.4": 11456.2695, "encoder_q-layer.5": 11630.7363, "encoder_q-layer.6": 12432.0186, "encoder_q-layer.7": 12740.6689, "encoder_q-layer.8": 13963.2178, "encoder_q-layer.9": 12620.8574, "epoch": 0.49, "inbatch_neg_score": 2.5929, "inbatch_pos_score": 3.1914, "learning_rate": 2.3131578947368422e-05, "loss": 3.4159, "norm_diff": 0.0407, "norm_loss": 0.0, "num_token_doc": 66.6926, "num_token_overlap": 11.6685, "num_token_query": 31.4513, "num_token_union": 65.1121, "num_word_context": 202.2082, "num_word_doc": 49.7508, "num_word_query": 23.3466, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21455.8754, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.5918, "query_norm": 2.0554, "queue_k_norm": 2.0936, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4513, "sent_len_1": 66.6926, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.2038, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112100 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.4218, "doc_norm": 2.0983, "encoder_q-embeddings": 10699.1846, "encoder_q-layer.0": 7053.311, "encoder_q-layer.1": 7594.7407, "encoder_q-layer.10": 12878.8076, "encoder_q-layer.11": 30824.5449, "encoder_q-layer.2": 8551.4707, "encoder_q-layer.3": 8745.875, "encoder_q-layer.4": 9435.6172, "encoder_q-layer.5": 9549.749, "encoder_q-layer.6": 10098.0195, "encoder_q-layer.7": 11332.5537, "encoder_q-layer.8": 13505.2236, "encoder_q-layer.9": 12349.2998, "epoch": 0.49, "inbatch_neg_score": 2.6064, "inbatch_pos_score": 3.2109, "learning_rate": 2.3105263157894737e-05, "loss": 3.4218, "norm_diff": 0.0366, "norm_loss": 0.0, "num_token_doc": 66.8342, "num_token_overlap": 11.66, "num_token_query": 31.4007, "num_token_union": 65.1839, "num_word_context": 202.3443, "num_word_doc": 49.8526, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19898.6341, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.6055, "query_norm": 2.0617, "queue_k_norm": 2.0966, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4007, "sent_len_1": 66.8342, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.0563, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112200 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.4279, "doc_norm": 2.1029, "encoder_q-embeddings": 10046.2383, "encoder_q-layer.0": 6884.5435, "encoder_q-layer.1": 7527.3096, "encoder_q-layer.10": 12319.6826, "encoder_q-layer.11": 30104.7812, "encoder_q-layer.2": 8023.2021, "encoder_q-layer.3": 8189.3149, "encoder_q-layer.4": 9010.7793, "encoder_q-layer.5": 9253.874, "encoder_q-layer.6": 10165.2695, "encoder_q-layer.7": 11202.7041, "encoder_q-layer.8": 12924.0938, "encoder_q-layer.9": 12231.2754, "epoch": 0.49, "inbatch_neg_score": 2.6228, "inbatch_pos_score": 3.2324, "learning_rate": 2.3078947368421052e-05, "loss": 3.4279, "norm_diff": 0.0381, "norm_loss": 0.0, "num_token_doc": 66.6733, "num_token_overlap": 11.6877, "num_token_query": 31.3704, "num_token_union": 65.0577, "num_word_context": 201.9535, "num_word_doc": 49.7558, "num_word_query": 23.2963, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19410.745, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.623, "query_norm": 2.0648, "queue_k_norm": 2.0978, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3704, "sent_len_1": 66.6733, "sent_len_max_0": 127.9425, "sent_len_max_1": 209.6788, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112300 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4094, "doc_norm": 2.1016, "encoder_q-embeddings": 17599.4062, "encoder_q-layer.0": 12607.1436, "encoder_q-layer.1": 14874.0654, "encoder_q-layer.10": 13041.7734, "encoder_q-layer.11": 30350.0801, "encoder_q-layer.2": 17915.4512, "encoder_q-layer.3": 19889.0684, "encoder_q-layer.4": 21146.5078, "encoder_q-layer.5": 20426.5527, "encoder_q-layer.6": 22669.5664, "encoder_q-layer.7": 22274.7422, "encoder_q-layer.8": 17813.0449, "encoder_q-layer.9": 12755.6934, "epoch": 0.49, "inbatch_neg_score": 2.6342, "inbatch_pos_score": 3.2363, "learning_rate": 2.305263157894737e-05, "loss": 3.4094, "norm_diff": 0.0298, "norm_loss": 0.0, "num_token_doc": 66.9326, "num_token_overlap": 11.7098, "num_token_query": 31.4233, "num_token_union": 65.2174, "num_word_context": 202.351, "num_word_doc": 49.9416, "num_word_query": 23.342, "postclip_grad_norm": 1.0, "preclip_grad_norm": 28990.1716, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 2.6367, "query_norm": 2.0718, "queue_k_norm": 2.1015, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4233, "sent_len_1": 66.9326, "sent_len_max_0": 128.0, "sent_len_max_1": 208.8438, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 112400 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.435, "doc_norm": 2.0998, "encoder_q-embeddings": 9672.7646, "encoder_q-layer.0": 6776.394, "encoder_q-layer.1": 6982.8027, "encoder_q-layer.10": 13655.707, "encoder_q-layer.11": 30611.3398, "encoder_q-layer.2": 7896.8589, "encoder_q-layer.3": 8083.0127, "encoder_q-layer.4": 8586.4453, "encoder_q-layer.5": 8656.2666, "encoder_q-layer.6": 10371.9385, "encoder_q-layer.7": 11589.5361, "encoder_q-layer.8": 13938.4707, "encoder_q-layer.9": 12892.8164, "epoch": 0.49, "inbatch_neg_score": 2.649, "inbatch_pos_score": 3.2383, "learning_rate": 2.3026315789473685e-05, "loss": 3.435, "norm_diff": 0.0352, "norm_loss": 0.0, "num_token_doc": 66.7312, "num_token_overlap": 11.6734, "num_token_query": 31.3521, "num_token_union": 65.0967, "num_word_context": 202.0659, "num_word_doc": 49.8077, "num_word_query": 23.2908, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19349.2261, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.6484, "query_norm": 2.0647, "queue_k_norm": 2.1034, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3521, "sent_len_1": 66.7312, "sent_len_max_0": 127.97, "sent_len_max_1": 208.6337, "stdk": 0.0491, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 112500 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.4225, "doc_norm": 2.1076, "encoder_q-embeddings": 9606.5547, "encoder_q-layer.0": 6724.6606, "encoder_q-layer.1": 7213.3423, "encoder_q-layer.10": 13011.7754, "encoder_q-layer.11": 31506.6191, "encoder_q-layer.2": 8032.5049, "encoder_q-layer.3": 8091.5352, "encoder_q-layer.4": 8670.2178, "encoder_q-layer.5": 8939.5879, "encoder_q-layer.6": 9952.1709, "encoder_q-layer.7": 11021.6465, "encoder_q-layer.8": 13136.0928, "encoder_q-layer.9": 11985.9785, "epoch": 0.49, "inbatch_neg_score": 2.6569, "inbatch_pos_score": 3.2754, "learning_rate": 2.3000000000000003e-05, "loss": 3.4225, "norm_diff": 0.0409, "norm_loss": 0.0, "num_token_doc": 66.8461, "num_token_overlap": 11.6533, "num_token_query": 31.3014, "num_token_union": 65.0994, "num_word_context": 202.3911, "num_word_doc": 49.8682, "num_word_query": 23.233, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19372.651, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.6582, "query_norm": 2.0667, "queue_k_norm": 2.1054, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3014, "sent_len_1": 66.8461, "sent_len_max_0": 127.975, "sent_len_max_1": 209.0437, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112600 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.4143, "doc_norm": 2.1133, "encoder_q-embeddings": 10068.8359, "encoder_q-layer.0": 7041.27, "encoder_q-layer.1": 7537.689, "encoder_q-layer.10": 12436.541, "encoder_q-layer.11": 29298.6875, "encoder_q-layer.2": 8751.9912, "encoder_q-layer.3": 8713.6768, "encoder_q-layer.4": 9153.373, "encoder_q-layer.5": 9354.2988, "encoder_q-layer.6": 10352.5195, "encoder_q-layer.7": 11841.001, "encoder_q-layer.8": 13880.4355, "encoder_q-layer.9": 12104.5234, "epoch": 0.49, "inbatch_neg_score": 2.6661, "inbatch_pos_score": 3.2812, "learning_rate": 2.297368421052632e-05, "loss": 3.4143, "norm_diff": 0.0516, "norm_loss": 0.0, "num_token_doc": 66.7825, "num_token_overlap": 11.6903, "num_token_query": 31.368, "num_token_union": 65.0828, "num_word_context": 202.4825, "num_word_doc": 49.839, "num_word_query": 23.2825, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19186.5215, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.666, "query_norm": 2.0617, "queue_k_norm": 2.1078, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.368, "sent_len_1": 66.7825, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.1125, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112700 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4302, "doc_norm": 2.11, "encoder_q-embeddings": 4581.582, "encoder_q-layer.0": 3295.959, "encoder_q-layer.1": 3422.7332, "encoder_q-layer.10": 6169.3867, "encoder_q-layer.11": 15683.5156, "encoder_q-layer.2": 3886.3362, "encoder_q-layer.3": 4017.3503, "encoder_q-layer.4": 4206.2334, "encoder_q-layer.5": 4252.4658, "encoder_q-layer.6": 5230.3159, "encoder_q-layer.7": 5862.5493, "encoder_q-layer.8": 6780.457, "encoder_q-layer.9": 5975.7642, "epoch": 0.49, "inbatch_neg_score": 2.6707, "inbatch_pos_score": 3.2734, "learning_rate": 2.294736842105263e-05, "loss": 3.4302, "norm_diff": 0.0586, "norm_loss": 0.0, "num_token_doc": 66.7103, "num_token_overlap": 11.6952, "num_token_query": 31.4313, "num_token_union": 65.1151, "num_word_context": 202.5803, "num_word_doc": 49.8072, "num_word_query": 23.3449, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9805.2624, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6699, "query_norm": 2.0514, "queue_k_norm": 2.111, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4313, "sent_len_1": 66.7103, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.295, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 112800 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4293, "doc_norm": 2.1149, "encoder_q-embeddings": 4765.6147, "encoder_q-layer.0": 3286.2532, "encoder_q-layer.1": 3528.1218, "encoder_q-layer.10": 6948.1235, "encoder_q-layer.11": 16144.375, "encoder_q-layer.2": 3980.0774, "encoder_q-layer.3": 4045.4163, "encoder_q-layer.4": 4289.0171, "encoder_q-layer.5": 4323.3369, "encoder_q-layer.6": 4905.5176, "encoder_q-layer.7": 5611.5732, "encoder_q-layer.8": 7539.3525, "encoder_q-layer.9": 7086.8672, "epoch": 0.49, "inbatch_neg_score": 2.6744, "inbatch_pos_score": 3.2793, "learning_rate": 2.292105263157895e-05, "loss": 3.4293, "norm_diff": 0.0645, "norm_loss": 0.0, "num_token_doc": 66.7659, "num_token_overlap": 11.6599, "num_token_query": 31.3057, "num_token_union": 65.116, "num_word_context": 202.2695, "num_word_doc": 49.8325, "num_word_query": 23.2447, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9967.5403, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6758, "query_norm": 2.0504, "queue_k_norm": 2.1121, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3057, "sent_len_1": 66.7659, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.7812, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 112900 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4273, "doc_norm": 2.1153, "encoder_q-embeddings": 4893.8374, "encoder_q-layer.0": 3286.042, "encoder_q-layer.1": 3575.1946, "encoder_q-layer.10": 6780.7446, "encoder_q-layer.11": 15861.1123, "encoder_q-layer.2": 4009.4648, "encoder_q-layer.3": 4148.6626, "encoder_q-layer.4": 4369.27, "encoder_q-layer.5": 4595.8999, "encoder_q-layer.6": 5170.0776, "encoder_q-layer.7": 5887.4766, "encoder_q-layer.8": 7295.0991, "encoder_q-layer.9": 6546.1211, "epoch": 0.49, "inbatch_neg_score": 2.679, "inbatch_pos_score": 3.2793, "learning_rate": 2.2894736842105263e-05, "loss": 3.4273, "norm_diff": 0.0586, "norm_loss": 0.0, "num_token_doc": 66.7245, "num_token_overlap": 11.6077, "num_token_query": 31.2042, "num_token_union": 65.0125, "num_word_context": 202.3028, "num_word_doc": 49.7743, "num_word_query": 23.1566, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10074.817, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6797, "query_norm": 2.0567, "queue_k_norm": 2.1141, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2042, "sent_len_1": 66.7245, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.4563, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 113000 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.4135, "doc_norm": 2.1131, "encoder_q-embeddings": 4955.793, "encoder_q-layer.0": 3445.5449, "encoder_q-layer.1": 3676.449, "encoder_q-layer.10": 7683.1313, "encoder_q-layer.11": 17093.9297, "encoder_q-layer.2": 4108.4878, "encoder_q-layer.3": 4079.2876, "encoder_q-layer.4": 4482.6055, "encoder_q-layer.5": 4649.4482, "encoder_q-layer.6": 5122.4224, "encoder_q-layer.7": 6323.6104, "encoder_q-layer.8": 7243.127, "encoder_q-layer.9": 6913.2207, "epoch": 0.49, "inbatch_neg_score": 2.6811, "inbatch_pos_score": 3.2773, "learning_rate": 2.286842105263158e-05, "loss": 3.4135, "norm_diff": 0.058, "norm_loss": 0.0, "num_token_doc": 66.8755, "num_token_overlap": 11.7048, "num_token_query": 31.4637, "num_token_union": 65.1955, "num_word_context": 202.2358, "num_word_doc": 49.8878, "num_word_query": 23.3843, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10287.0064, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6797, "query_norm": 2.0551, "queue_k_norm": 2.1166, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4637, "sent_len_1": 66.8755, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.9187, "stdk": 0.0491, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 113100 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4234, "doc_norm": 2.1177, "encoder_q-embeddings": 4974.186, "encoder_q-layer.0": 3381.4241, "encoder_q-layer.1": 3610.2493, "encoder_q-layer.10": 6601.5942, "encoder_q-layer.11": 16072.1348, "encoder_q-layer.2": 4080.1614, "encoder_q-layer.3": 4165.6904, "encoder_q-layer.4": 4493.9541, "encoder_q-layer.5": 4656.729, "encoder_q-layer.6": 5302.231, "encoder_q-layer.7": 6083.915, "encoder_q-layer.8": 6756.2026, "encoder_q-layer.9": 6241.894, "epoch": 0.49, "inbatch_neg_score": 2.6805, "inbatch_pos_score": 3.2871, "learning_rate": 2.2842105263157897e-05, "loss": 3.4234, "norm_diff": 0.0671, "norm_loss": 0.0, "num_token_doc": 66.6631, "num_token_overlap": 11.6463, "num_token_query": 31.3681, "num_token_union": 65.0492, "num_word_context": 202.3646, "num_word_doc": 49.7484, "num_word_query": 23.3045, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10011.8826, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6816, "query_norm": 2.0506, "queue_k_norm": 2.1182, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3681, "sent_len_1": 66.6631, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.9212, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 113200 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.4182, "doc_norm": 2.1194, "encoder_q-embeddings": 5004.6064, "encoder_q-layer.0": 3492.6904, "encoder_q-layer.1": 3849.6738, "encoder_q-layer.10": 6205.4751, "encoder_q-layer.11": 15201.3789, "encoder_q-layer.2": 4155.689, "encoder_q-layer.3": 4165.0229, "encoder_q-layer.4": 4454.7876, "encoder_q-layer.5": 4655.8789, "encoder_q-layer.6": 5087.873, "encoder_q-layer.7": 5644.4492, "encoder_q-layer.8": 6843.5874, "encoder_q-layer.9": 6316.3691, "epoch": 0.49, "inbatch_neg_score": 2.692, "inbatch_pos_score": 3.291, "learning_rate": 2.281578947368421e-05, "loss": 3.4182, "norm_diff": 0.0691, "norm_loss": 0.0, "num_token_doc": 66.8213, "num_token_overlap": 11.7077, "num_token_query": 31.4504, "num_token_union": 65.152, "num_word_context": 202.2747, "num_word_doc": 49.8406, "num_word_query": 23.3575, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9833.1448, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6914, "query_norm": 2.0503, "queue_k_norm": 2.1199, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4504, "sent_len_1": 66.8213, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.37, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 113300 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4218, "doc_norm": 2.1216, "encoder_q-embeddings": 4735.7466, "encoder_q-layer.0": 3241.8997, "encoder_q-layer.1": 3432.3931, "encoder_q-layer.10": 6604.8286, "encoder_q-layer.11": 15377.2725, "encoder_q-layer.2": 3895.968, "encoder_q-layer.3": 3979.8499, "encoder_q-layer.4": 4274.8174, "encoder_q-layer.5": 4435.9912, "encoder_q-layer.6": 4979.6729, "encoder_q-layer.7": 5679.002, "encoder_q-layer.8": 6963.9487, "encoder_q-layer.9": 6220.209, "epoch": 0.49, "inbatch_neg_score": 2.6917, "inbatch_pos_score": 3.2871, "learning_rate": 2.2789473684210527e-05, "loss": 3.4218, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.7074, "num_token_overlap": 11.68, "num_token_query": 31.36, "num_token_union": 65.0311, "num_word_context": 202.2297, "num_word_doc": 49.7638, "num_word_query": 23.3008, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9647.8499, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6914, "query_norm": 2.0427, "queue_k_norm": 2.1207, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.36, "sent_len_1": 66.7074, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.7113, "stdk": 0.0493, "stdq": 0.0425, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 113400 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4229, "doc_norm": 2.1257, "encoder_q-embeddings": 5043.5757, "encoder_q-layer.0": 3499.252, "encoder_q-layer.1": 3798.533, "encoder_q-layer.10": 6403.2866, "encoder_q-layer.11": 15170.585, "encoder_q-layer.2": 4254.2505, "encoder_q-layer.3": 4374.5156, "encoder_q-layer.4": 4481.083, "encoder_q-layer.5": 4817.3857, "encoder_q-layer.6": 5106.856, "encoder_q-layer.7": 5883.3833, "encoder_q-layer.8": 6352.3887, "encoder_q-layer.9": 5966.7827, "epoch": 0.49, "inbatch_neg_score": 2.6907, "inbatch_pos_score": 3.3027, "learning_rate": 2.2763157894736845e-05, "loss": 3.4229, "norm_diff": 0.0748, "norm_loss": 0.0, "num_token_doc": 66.7288, "num_token_overlap": 11.6483, "num_token_query": 31.2762, "num_token_union": 65.07, "num_word_context": 201.9703, "num_word_doc": 49.7916, "num_word_query": 23.2252, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9691.9697, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6914, "query_norm": 2.0509, "queue_k_norm": 2.1213, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2762, "sent_len_1": 66.7288, "sent_len_max_0": 127.99, "sent_len_max_1": 207.27, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 113500 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4096, "doc_norm": 2.1258, "encoder_q-embeddings": 5116.8281, "encoder_q-layer.0": 3592.7156, "encoder_q-layer.1": 3758.052, "encoder_q-layer.10": 6953.729, "encoder_q-layer.11": 15277.9609, "encoder_q-layer.2": 4085.4622, "encoder_q-layer.3": 4121.7549, "encoder_q-layer.4": 4255.3936, "encoder_q-layer.5": 4452.6455, "encoder_q-layer.6": 4876.4277, "encoder_q-layer.7": 5510.8242, "encoder_q-layer.8": 6594.7065, "encoder_q-layer.9": 6368.5918, "epoch": 0.49, "inbatch_neg_score": 2.6977, "inbatch_pos_score": 3.3008, "learning_rate": 2.273684210526316e-05, "loss": 3.4096, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.7806, "num_token_overlap": 11.7314, "num_token_query": 31.4779, "num_token_union": 65.1597, "num_word_context": 202.2964, "num_word_doc": 49.8522, "num_word_query": 23.3785, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9880.1309, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.6973, "query_norm": 2.0529, "queue_k_norm": 2.1236, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4779, "sent_len_1": 66.7806, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.0825, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 113600 }, { "accuracy": 47.7051, "active_queue_size": 16384.0, "cl_loss": 3.4064, "doc_norm": 2.1276, "encoder_q-embeddings": 2478.8542, "encoder_q-layer.0": 1700.6613, "encoder_q-layer.1": 1831.6022, "encoder_q-layer.10": 3263.3777, "encoder_q-layer.11": 8004.9951, "encoder_q-layer.2": 2052.7649, "encoder_q-layer.3": 2102.0574, "encoder_q-layer.4": 2213.377, "encoder_q-layer.5": 2320.3889, "encoder_q-layer.6": 2484.106, "encoder_q-layer.7": 2833.0762, "encoder_q-layer.8": 3634.6218, "encoder_q-layer.9": 3256.6941, "epoch": 0.49, "inbatch_neg_score": 2.7016, "inbatch_pos_score": 3.293, "learning_rate": 2.2710526315789475e-05, "loss": 3.4064, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.9743, "num_token_overlap": 11.7119, "num_token_query": 31.3914, "num_token_union": 65.1969, "num_word_context": 202.281, "num_word_doc": 49.9374, "num_word_query": 23.3125, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5066.0688, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7012, "query_norm": 2.0482, "queue_k_norm": 2.1258, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3914, "sent_len_1": 66.9743, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.3288, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 113700 }, { "accuracy": 53.1738, "active_queue_size": 16384.0, "cl_loss": 3.4179, "doc_norm": 2.1268, "encoder_q-embeddings": 2509.0308, "encoder_q-layer.0": 1676.5677, "encoder_q-layer.1": 1804.6854, "encoder_q-layer.10": 3162.3711, "encoder_q-layer.11": 7484.2817, "encoder_q-layer.2": 2060.3181, "encoder_q-layer.3": 2060.3689, "encoder_q-layer.4": 2291.0171, "encoder_q-layer.5": 2358.96, "encoder_q-layer.6": 2620.8018, "encoder_q-layer.7": 2887.9817, "encoder_q-layer.8": 3323.6702, "encoder_q-layer.9": 2973.3516, "epoch": 0.49, "inbatch_neg_score": 2.7045, "inbatch_pos_score": 3.3203, "learning_rate": 2.268421052631579e-05, "loss": 3.4179, "norm_diff": 0.0741, "norm_loss": 0.0, "num_token_doc": 66.6905, "num_token_overlap": 11.6656, "num_token_query": 31.3354, "num_token_union": 65.0216, "num_word_context": 202.1029, "num_word_doc": 49.7399, "num_word_query": 23.2636, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4859.4236, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7051, "query_norm": 2.0527, "queue_k_norm": 2.1274, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3354, "sent_len_1": 66.6905, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.7375, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 113800 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4339, "doc_norm": 2.1292, "encoder_q-embeddings": 2457.5078, "encoder_q-layer.0": 1650.6003, "encoder_q-layer.1": 1757.6956, "encoder_q-layer.10": 3207.0769, "encoder_q-layer.11": 7836.0986, "encoder_q-layer.2": 1982.5603, "encoder_q-layer.3": 2036.2156, "encoder_q-layer.4": 2167.8667, "encoder_q-layer.5": 2191.823, "encoder_q-layer.6": 2512.3113, "encoder_q-layer.7": 2759.0349, "encoder_q-layer.8": 3353.2756, "encoder_q-layer.9": 3031.2334, "epoch": 0.49, "inbatch_neg_score": 2.7045, "inbatch_pos_score": 3.2988, "learning_rate": 2.2657894736842105e-05, "loss": 3.4339, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.6733, "num_token_overlap": 11.6576, "num_token_query": 31.3281, "num_token_union": 65.0631, "num_word_context": 202.1924, "num_word_doc": 49.7678, "num_word_query": 23.2605, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4827.8051, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7031, "query_norm": 2.0496, "queue_k_norm": 2.1282, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3281, "sent_len_1": 66.6733, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.9162, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 113900 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.428, "doc_norm": 2.1303, "encoder_q-embeddings": 2510.3489, "encoder_q-layer.0": 1794.1057, "encoder_q-layer.1": 1923.0101, "encoder_q-layer.10": 3112.6309, "encoder_q-layer.11": 7414.9565, "encoder_q-layer.2": 2188.8511, "encoder_q-layer.3": 2268.5259, "encoder_q-layer.4": 2397.1406, "encoder_q-layer.5": 2389.5486, "encoder_q-layer.6": 2653.5803, "encoder_q-layer.7": 3070.1296, "encoder_q-layer.8": 3270.4905, "encoder_q-layer.9": 2939.4822, "epoch": 0.49, "inbatch_neg_score": 2.7101, "inbatch_pos_score": 3.3086, "learning_rate": 2.2631578947368423e-05, "loss": 3.428, "norm_diff": 0.0848, "norm_loss": 0.0, "num_token_doc": 66.8128, "num_token_overlap": 11.6885, "num_token_query": 31.4833, "num_token_union": 65.2299, "num_word_context": 202.6289, "num_word_doc": 49.888, "num_word_query": 23.4022, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4936.4719, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7109, "query_norm": 2.0454, "queue_k_norm": 2.1289, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4833, "sent_len_1": 66.8128, "sent_len_max_0": 127.99, "sent_len_max_1": 208.215, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114000 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4205, "doc_norm": 2.1307, "encoder_q-embeddings": 4060.1641, "encoder_q-layer.0": 2871.959, "encoder_q-layer.1": 3312.0103, "encoder_q-layer.10": 3246.4683, "encoder_q-layer.11": 7521.9536, "encoder_q-layer.2": 3642.8447, "encoder_q-layer.3": 4116.4194, "encoder_q-layer.4": 4429.6699, "encoder_q-layer.5": 4794.8008, "encoder_q-layer.6": 4727.6709, "encoder_q-layer.7": 4583.4155, "encoder_q-layer.8": 4046.9155, "encoder_q-layer.9": 3090.3445, "epoch": 0.5, "inbatch_neg_score": 2.7123, "inbatch_pos_score": 3.3223, "learning_rate": 2.2605263157894738e-05, "loss": 3.4205, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.9114, "num_token_overlap": 11.696, "num_token_query": 31.3296, "num_token_union": 65.1536, "num_word_context": 202.277, "num_word_doc": 49.9005, "num_word_query": 23.2631, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6415.8936, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7129, "query_norm": 2.0603, "queue_k_norm": 2.1311, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3296, "sent_len_1": 66.9114, "sent_len_max_0": 127.975, "sent_len_max_1": 210.0387, "stdk": 0.0493, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 114100 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.4278, "doc_norm": 2.1316, "encoder_q-embeddings": 2687.238, "encoder_q-layer.0": 1884.4346, "encoder_q-layer.1": 1959.5458, "encoder_q-layer.10": 3716.0859, "encoder_q-layer.11": 7925.1123, "encoder_q-layer.2": 2227.4861, "encoder_q-layer.3": 2326.5378, "encoder_q-layer.4": 2418.8433, "encoder_q-layer.5": 2426.7759, "encoder_q-layer.6": 2617.4648, "encoder_q-layer.7": 2987.885, "encoder_q-layer.8": 3658.4453, "encoder_q-layer.9": 3209.4468, "epoch": 0.5, "inbatch_neg_score": 2.7163, "inbatch_pos_score": 3.3203, "learning_rate": 2.2578947368421057e-05, "loss": 3.4278, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.7184, "num_token_overlap": 11.6709, "num_token_query": 31.3153, "num_token_union": 65.0648, "num_word_context": 202.3058, "num_word_doc": 49.7777, "num_word_query": 23.2539, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5108.4433, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7168, "query_norm": 2.0561, "queue_k_norm": 2.1308, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3153, "sent_len_1": 66.7184, "sent_len_max_0": 127.9638, "sent_len_max_1": 207.87, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114200 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4212, "doc_norm": 2.1352, "encoder_q-embeddings": 2586.2732, "encoder_q-layer.0": 1801.7753, "encoder_q-layer.1": 1921.0271, "encoder_q-layer.10": 3140.249, "encoder_q-layer.11": 7760.5288, "encoder_q-layer.2": 2169.4392, "encoder_q-layer.3": 2243.7329, "encoder_q-layer.4": 2345.3447, "encoder_q-layer.5": 2410.7825, "encoder_q-layer.6": 2605.5654, "encoder_q-layer.7": 2835.249, "encoder_q-layer.8": 3351.0745, "encoder_q-layer.9": 3048.3286, "epoch": 0.5, "inbatch_neg_score": 2.7191, "inbatch_pos_score": 3.3262, "learning_rate": 2.2552631578947368e-05, "loss": 3.4212, "norm_diff": 0.0786, "norm_loss": 0.0, "num_token_doc": 66.8093, "num_token_overlap": 11.668, "num_token_query": 31.38, "num_token_union": 65.1613, "num_word_context": 202.5783, "num_word_doc": 49.8703, "num_word_query": 23.3073, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4960.7333, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7188, "query_norm": 2.0566, "queue_k_norm": 2.1329, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.38, "sent_len_1": 66.8093, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.4437, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 114300 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4391, "doc_norm": 2.1357, "encoder_q-embeddings": 3128.1865, "encoder_q-layer.0": 2294.324, "encoder_q-layer.1": 2509.9902, "encoder_q-layer.10": 3268.0823, "encoder_q-layer.11": 7690.4897, "encoder_q-layer.2": 2597.3833, "encoder_q-layer.3": 2483.7944, "encoder_q-layer.4": 2595.7244, "encoder_q-layer.5": 2443.1636, "encoder_q-layer.6": 2686.2004, "encoder_q-layer.7": 3106.3772, "encoder_q-layer.8": 3614.3865, "encoder_q-layer.9": 3148.5869, "epoch": 0.5, "inbatch_neg_score": 2.7255, "inbatch_pos_score": 3.334, "learning_rate": 2.2526315789473683e-05, "loss": 3.4391, "norm_diff": 0.0819, "norm_loss": 0.0, "num_token_doc": 66.6501, "num_token_overlap": 11.6615, "num_token_query": 31.1774, "num_token_union": 64.9585, "num_word_context": 202.1014, "num_word_doc": 49.7487, "num_word_query": 23.1517, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5288.9147, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7246, "query_norm": 2.0538, "queue_k_norm": 2.133, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.1774, "sent_len_1": 66.6501, "sent_len_max_0": 127.99, "sent_len_max_1": 209.6287, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114400 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4295, "doc_norm": 2.1358, "encoder_q-embeddings": 2392.0879, "encoder_q-layer.0": 1691.7296, "encoder_q-layer.1": 1865.0547, "encoder_q-layer.10": 3450.2502, "encoder_q-layer.11": 8208.4912, "encoder_q-layer.2": 2086.5271, "encoder_q-layer.3": 2123.5269, "encoder_q-layer.4": 2248.8909, "encoder_q-layer.5": 2255.8054, "encoder_q-layer.6": 2468.0881, "encoder_q-layer.7": 2772.7886, "encoder_q-layer.8": 3500.2422, "encoder_q-layer.9": 3289.3933, "epoch": 0.5, "inbatch_neg_score": 2.7282, "inbatch_pos_score": 3.3438, "learning_rate": 2.25e-05, "loss": 3.4295, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.7592, "num_token_overlap": 11.679, "num_token_query": 31.3865, "num_token_union": 65.1573, "num_word_context": 202.1974, "num_word_doc": 49.8259, "num_word_query": 23.3235, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5118.5767, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7285, "query_norm": 2.054, "queue_k_norm": 2.1357, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3865, "sent_len_1": 66.7592, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.5925, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 114500 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.4155, "doc_norm": 2.1331, "encoder_q-embeddings": 2491.001, "encoder_q-layer.0": 1736.5967, "encoder_q-layer.1": 1806.5741, "encoder_q-layer.10": 3125.7764, "encoder_q-layer.11": 7955.3789, "encoder_q-layer.2": 1954.0074, "encoder_q-layer.3": 2030.2955, "encoder_q-layer.4": 2204.4126, "encoder_q-layer.5": 2275.0713, "encoder_q-layer.6": 2495.7981, "encoder_q-layer.7": 2830.7375, "encoder_q-layer.8": 3293.7251, "encoder_q-layer.9": 3008.6624, "epoch": 0.5, "inbatch_neg_score": 2.7382, "inbatch_pos_score": 3.3203, "learning_rate": 2.2473684210526316e-05, "loss": 3.4155, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.9553, "num_token_overlap": 11.7217, "num_token_query": 31.4434, "num_token_union": 65.2602, "num_word_context": 202.4673, "num_word_doc": 49.9822, "num_word_query": 23.3641, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4888.2681, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7383, "query_norm": 2.0552, "queue_k_norm": 2.136, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4434, "sent_len_1": 66.9553, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.6125, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114600 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.4256, "doc_norm": 2.1354, "encoder_q-embeddings": 2822.248, "encoder_q-layer.0": 2272.2568, "encoder_q-layer.1": 2233.5571, "encoder_q-layer.10": 3538.1782, "encoder_q-layer.11": 7305.0957, "encoder_q-layer.2": 2334.7263, "encoder_q-layer.3": 2284.8496, "encoder_q-layer.4": 2236.5281, "encoder_q-layer.5": 2193.0989, "encoder_q-layer.6": 2349.304, "encoder_q-layer.7": 2723.0225, "encoder_q-layer.8": 3195.5293, "encoder_q-layer.9": 3085.8738, "epoch": 0.5, "inbatch_neg_score": 2.7381, "inbatch_pos_score": 3.3633, "learning_rate": 2.2447368421052635e-05, "loss": 3.4256, "norm_diff": 0.0685, "norm_loss": 0.0, "num_token_doc": 66.8052, "num_token_overlap": 11.6878, "num_token_query": 31.4354, "num_token_union": 65.1832, "num_word_context": 202.3745, "num_word_doc": 49.8692, "num_word_query": 23.344, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4894.2363, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7383, "query_norm": 2.0669, "queue_k_norm": 2.1379, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4354, "sent_len_1": 66.8052, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.72, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 114700 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4213, "doc_norm": 2.1364, "encoder_q-embeddings": 2443.7988, "encoder_q-layer.0": 1671.5273, "encoder_q-layer.1": 1789.2651, "encoder_q-layer.10": 3258.0371, "encoder_q-layer.11": 7492.8945, "encoder_q-layer.2": 1951.1069, "encoder_q-layer.3": 2018.7067, "encoder_q-layer.4": 2202.9658, "encoder_q-layer.5": 2227.4233, "encoder_q-layer.6": 2359.2773, "encoder_q-layer.7": 2749.1687, "encoder_q-layer.8": 3427.9099, "encoder_q-layer.9": 3170.2502, "epoch": 0.5, "inbatch_neg_score": 2.7439, "inbatch_pos_score": 3.3535, "learning_rate": 2.242105263157895e-05, "loss": 3.4213, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.7456, "num_token_overlap": 11.6546, "num_token_query": 31.3982, "num_token_union": 65.143, "num_word_context": 202.4547, "num_word_doc": 49.789, "num_word_query": 23.3275, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4776.257, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7441, "query_norm": 2.0619, "queue_k_norm": 2.1378, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3982, "sent_len_1": 66.7456, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.8562, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114800 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.4299, "doc_norm": 2.1415, "encoder_q-embeddings": 2549.4177, "encoder_q-layer.0": 1737.5426, "encoder_q-layer.1": 1804.4525, "encoder_q-layer.10": 3271.4302, "encoder_q-layer.11": 7402.6694, "encoder_q-layer.2": 2038.4944, "encoder_q-layer.3": 2053.29, "encoder_q-layer.4": 2143.2896, "encoder_q-layer.5": 2261.8611, "encoder_q-layer.6": 2589.0996, "encoder_q-layer.7": 2944.5955, "encoder_q-layer.8": 3394.0527, "encoder_q-layer.9": 3133.0283, "epoch": 0.5, "inbatch_neg_score": 2.7474, "inbatch_pos_score": 3.3438, "learning_rate": 2.2394736842105265e-05, "loss": 3.4299, "norm_diff": 0.0863, "norm_loss": 0.0, "num_token_doc": 66.7612, "num_token_overlap": 11.6512, "num_token_query": 31.3097, "num_token_union": 65.0841, "num_word_context": 202.4443, "num_word_doc": 49.7892, "num_word_query": 23.2496, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4847.2574, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7461, "query_norm": 2.0552, "queue_k_norm": 2.1389, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3097, "sent_len_1": 66.7612, "sent_len_max_0": 127.985, "sent_len_max_1": 210.3462, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 114900 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4159, "doc_norm": 2.1441, "encoder_q-embeddings": 2513.7019, "encoder_q-layer.0": 1730.3173, "encoder_q-layer.1": 1784.3765, "encoder_q-layer.10": 3010.9299, "encoder_q-layer.11": 7675.313, "encoder_q-layer.2": 1978.5632, "encoder_q-layer.3": 2015.5157, "encoder_q-layer.4": 2171.4104, "encoder_q-layer.5": 2182.8733, "encoder_q-layer.6": 2490.1729, "encoder_q-layer.7": 2742.4351, "encoder_q-layer.8": 3266.9565, "encoder_q-layer.9": 3059.7207, "epoch": 0.5, "inbatch_neg_score": 2.7517, "inbatch_pos_score": 3.3613, "learning_rate": 2.236842105263158e-05, "loss": 3.4159, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.9445, "num_token_overlap": 11.6641, "num_token_query": 31.3134, "num_token_union": 65.1975, "num_word_context": 202.5595, "num_word_doc": 49.9484, "num_word_query": 23.2468, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4881.1085, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.752, "query_norm": 2.0665, "queue_k_norm": 2.1413, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3134, "sent_len_1": 66.9445, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.7038, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 115000 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.4422, "doc_norm": 2.1403, "encoder_q-embeddings": 2541.6912, "encoder_q-layer.0": 1724.3837, "encoder_q-layer.1": 1828.4619, "encoder_q-layer.10": 3286.1438, "encoder_q-layer.11": 7784.7559, "encoder_q-layer.2": 2035.1235, "encoder_q-layer.3": 2097.4385, "encoder_q-layer.4": 2210.7959, "encoder_q-layer.5": 2263.5059, "encoder_q-layer.6": 2609.9526, "encoder_q-layer.7": 2885.3281, "encoder_q-layer.8": 3465.8181, "encoder_q-layer.9": 3214.3459, "epoch": 0.5, "inbatch_neg_score": 2.753, "inbatch_pos_score": 3.3457, "learning_rate": 2.2342105263157895e-05, "loss": 3.4422, "norm_diff": 0.0833, "norm_loss": 0.0, "num_token_doc": 66.5129, "num_token_overlap": 11.6084, "num_token_query": 31.3511, "num_token_union": 64.9944, "num_word_context": 202.2818, "num_word_doc": 49.6401, "num_word_query": 23.2764, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4945.5547, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7539, "query_norm": 2.057, "queue_k_norm": 2.1407, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3511, "sent_len_1": 66.5129, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.195, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115100 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.4081, "doc_norm": 2.1415, "encoder_q-embeddings": 2741.6426, "encoder_q-layer.0": 1973.8651, "encoder_q-layer.1": 2137.9814, "encoder_q-layer.10": 3272.0913, "encoder_q-layer.11": 7467.5894, "encoder_q-layer.2": 2428.5562, "encoder_q-layer.3": 2491.3701, "encoder_q-layer.4": 2695.0034, "encoder_q-layer.5": 2834.2686, "encoder_q-layer.6": 3059.7856, "encoder_q-layer.7": 3144.5457, "encoder_q-layer.8": 3596.2307, "encoder_q-layer.9": 3093.6179, "epoch": 0.5, "inbatch_neg_score": 2.7571, "inbatch_pos_score": 3.3691, "learning_rate": 2.2315789473684213e-05, "loss": 3.4081, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.8676, "num_token_overlap": 11.71, "num_token_query": 31.4928, "num_token_union": 65.2552, "num_word_context": 202.5179, "num_word_doc": 49.8955, "num_word_query": 23.3956, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5143.337, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7578, "query_norm": 2.0613, "queue_k_norm": 2.1413, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4928, "sent_len_1": 66.8676, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.79, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115200 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4087, "doc_norm": 2.142, "encoder_q-embeddings": 2366.4673, "encoder_q-layer.0": 1616.6335, "encoder_q-layer.1": 1751.0562, "encoder_q-layer.10": 2919.8054, "encoder_q-layer.11": 7329.23, "encoder_q-layer.2": 1947.7787, "encoder_q-layer.3": 2015.8093, "encoder_q-layer.4": 2098.637, "encoder_q-layer.5": 2172.7063, "encoder_q-layer.6": 2410.3264, "encoder_q-layer.7": 2875.8123, "encoder_q-layer.8": 3226.0017, "encoder_q-layer.9": 2877.5024, "epoch": 0.5, "inbatch_neg_score": 2.7602, "inbatch_pos_score": 3.3691, "learning_rate": 2.2289473684210528e-05, "loss": 3.4087, "norm_diff": 0.0783, "norm_loss": 0.0, "num_token_doc": 66.7288, "num_token_overlap": 11.7128, "num_token_query": 31.4341, "num_token_union": 65.0971, "num_word_context": 202.4602, "num_word_doc": 49.8059, "num_word_query": 23.3534, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4628.9927, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7598, "query_norm": 2.0637, "queue_k_norm": 2.1429, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4341, "sent_len_1": 66.7288, "sent_len_max_0": 127.9875, "sent_len_max_1": 206.915, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115300 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4077, "doc_norm": 2.1447, "encoder_q-embeddings": 2452.4194, "encoder_q-layer.0": 1748.7606, "encoder_q-layer.1": 1904.9729, "encoder_q-layer.10": 3502.3545, "encoder_q-layer.11": 7895.5454, "encoder_q-layer.2": 2124.8425, "encoder_q-layer.3": 2065.7273, "encoder_q-layer.4": 2308.0593, "encoder_q-layer.5": 2383.7278, "encoder_q-layer.6": 2585.2871, "encoder_q-layer.7": 3247.4199, "encoder_q-layer.8": 3934.0425, "encoder_q-layer.9": 3430.4219, "epoch": 0.5, "inbatch_neg_score": 2.7642, "inbatch_pos_score": 3.3711, "learning_rate": 2.2263157894736843e-05, "loss": 3.4077, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.8153, "num_token_overlap": 11.6703, "num_token_query": 31.3516, "num_token_union": 65.1468, "num_word_context": 202.4078, "num_word_doc": 49.8603, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5082.8772, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7656, "query_norm": 2.0668, "queue_k_norm": 2.1445, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3516, "sent_len_1": 66.8153, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.15, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 115400 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.4114, "doc_norm": 2.1464, "encoder_q-embeddings": 2826.3635, "encoder_q-layer.0": 1969.0277, "encoder_q-layer.1": 2122.0127, "encoder_q-layer.10": 3370.5891, "encoder_q-layer.11": 8199.6074, "encoder_q-layer.2": 2497.042, "encoder_q-layer.3": 2545.5264, "encoder_q-layer.4": 2650.55, "encoder_q-layer.5": 2542.0681, "encoder_q-layer.6": 2625.3628, "encoder_q-layer.7": 2867.145, "encoder_q-layer.8": 3538.7268, "encoder_q-layer.9": 3241.2222, "epoch": 0.5, "inbatch_neg_score": 2.7666, "inbatch_pos_score": 3.3711, "learning_rate": 2.2236842105263158e-05, "loss": 3.4114, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.7039, "num_token_overlap": 11.6671, "num_token_query": 31.2856, "num_token_union": 65.0242, "num_word_context": 201.9031, "num_word_doc": 49.7861, "num_word_query": 23.2266, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5351.7136, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7676, "query_norm": 2.0715, "queue_k_norm": 2.1442, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2856, "sent_len_1": 66.7039, "sent_len_max_0": 127.9688, "sent_len_max_1": 206.9275, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115500 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.4248, "doc_norm": 2.1466, "encoder_q-embeddings": 2365.6921, "encoder_q-layer.0": 1701.97, "encoder_q-layer.1": 1805.3589, "encoder_q-layer.10": 3378.7595, "encoder_q-layer.11": 7527.4258, "encoder_q-layer.2": 1956.5985, "encoder_q-layer.3": 1984.2422, "encoder_q-layer.4": 2101.6277, "encoder_q-layer.5": 2162.2661, "encoder_q-layer.6": 2360.5984, "encoder_q-layer.7": 2667.9312, "encoder_q-layer.8": 3243.8232, "encoder_q-layer.9": 3042.5991, "epoch": 0.5, "inbatch_neg_score": 2.7707, "inbatch_pos_score": 3.3789, "learning_rate": 2.2210526315789476e-05, "loss": 3.4248, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.7527, "num_token_overlap": 11.6574, "num_token_query": 31.3697, "num_token_union": 65.1314, "num_word_context": 202.193, "num_word_doc": 49.8193, "num_word_query": 23.3107, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4785.5001, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 2.7715, "query_norm": 2.0673, "queue_k_norm": 2.1452, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3697, "sent_len_1": 66.7527, "sent_len_max_0": 127.975, "sent_len_max_1": 207.67, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115600 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.4057, "doc_norm": 2.145, "encoder_q-embeddings": 4876.686, "encoder_q-layer.0": 3446.6204, "encoder_q-layer.1": 3581.9702, "encoder_q-layer.10": 6601.3428, "encoder_q-layer.11": 15798.4824, "encoder_q-layer.2": 3959.7544, "encoder_q-layer.3": 4041.1848, "encoder_q-layer.4": 4314.0767, "encoder_q-layer.5": 4380.8179, "encoder_q-layer.6": 5074.9077, "encoder_q-layer.7": 5883.0298, "encoder_q-layer.8": 7224.4971, "encoder_q-layer.9": 6499.1455, "epoch": 0.5, "inbatch_neg_score": 2.7726, "inbatch_pos_score": 3.3867, "learning_rate": 2.218421052631579e-05, "loss": 3.4057, "norm_diff": 0.0762, "norm_loss": 0.0, "num_token_doc": 66.9731, "num_token_overlap": 11.7297, "num_token_query": 31.3932, "num_token_union": 65.2073, "num_word_context": 202.5055, "num_word_doc": 49.9728, "num_word_query": 23.304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10032.9193, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7734, "query_norm": 2.0688, "queue_k_norm": 2.1467, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3932, "sent_len_1": 66.9731, "sent_len_max_0": 127.99, "sent_len_max_1": 208.0412, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115700 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.4208, "doc_norm": 2.1507, "encoder_q-embeddings": 4967.9209, "encoder_q-layer.0": 3547.9426, "encoder_q-layer.1": 3923.5383, "encoder_q-layer.10": 6381.6787, "encoder_q-layer.11": 14774.8359, "encoder_q-layer.2": 4402.4028, "encoder_q-layer.3": 4353.3999, "encoder_q-layer.4": 4671.6133, "encoder_q-layer.5": 4606.5342, "encoder_q-layer.6": 5119.5112, "encoder_q-layer.7": 6107.0005, "encoder_q-layer.8": 7246.4878, "encoder_q-layer.9": 6255.5918, "epoch": 0.5, "inbatch_neg_score": 2.7766, "inbatch_pos_score": 3.3809, "learning_rate": 2.2157894736842106e-05, "loss": 3.4208, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.7157, "num_token_overlap": 11.7012, "num_token_query": 31.3775, "num_token_union": 65.0517, "num_word_context": 202.0813, "num_word_doc": 49.8017, "num_word_query": 23.2996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9871.436, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7754, "query_norm": 2.0677, "queue_k_norm": 2.1486, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3775, "sent_len_1": 66.7157, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6825, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 115800 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4253, "doc_norm": 2.1502, "encoder_q-embeddings": 4635.2554, "encoder_q-layer.0": 3306.1338, "encoder_q-layer.1": 3438.9194, "encoder_q-layer.10": 6621.0259, "encoder_q-layer.11": 15453.2344, "encoder_q-layer.2": 3973.1414, "encoder_q-layer.3": 4013.2612, "encoder_q-layer.4": 4241.9067, "encoder_q-layer.5": 4465.1538, "encoder_q-layer.6": 4769.3984, "encoder_q-layer.7": 5576.7939, "encoder_q-layer.8": 6625.3818, "encoder_q-layer.9": 6281.7314, "epoch": 0.5, "inbatch_neg_score": 2.7747, "inbatch_pos_score": 3.3809, "learning_rate": 2.213157894736842e-05, "loss": 3.4253, "norm_diff": 0.0853, "norm_loss": 0.0, "num_token_doc": 66.8544, "num_token_overlap": 11.6633, "num_token_query": 31.3533, "num_token_union": 65.1691, "num_word_context": 202.3707, "num_word_doc": 49.8716, "num_word_query": 23.2855, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9743.5314, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7754, "query_norm": 2.0649, "queue_k_norm": 2.1481, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3533, "sent_len_1": 66.8544, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.4162, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 115900 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4237, "doc_norm": 2.1511, "encoder_q-embeddings": 5406.7153, "encoder_q-layer.0": 3835.417, "encoder_q-layer.1": 4216.4829, "encoder_q-layer.10": 6343.7617, "encoder_q-layer.11": 15155.7207, "encoder_q-layer.2": 4709.2114, "encoder_q-layer.3": 4731.9307, "encoder_q-layer.4": 5153.7295, "encoder_q-layer.5": 5072.1621, "encoder_q-layer.6": 5374.2188, "encoder_q-layer.7": 5599.6802, "encoder_q-layer.8": 6797.0186, "encoder_q-layer.9": 6164.2788, "epoch": 0.5, "inbatch_neg_score": 2.7801, "inbatch_pos_score": 3.3809, "learning_rate": 2.2105263157894736e-05, "loss": 3.4237, "norm_diff": 0.0887, "norm_loss": 0.0, "num_token_doc": 66.6769, "num_token_overlap": 11.6661, "num_token_query": 31.3594, "num_token_union": 65.0568, "num_word_context": 202.2237, "num_word_doc": 49.761, "num_word_query": 23.2965, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10094.7378, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7793, "query_norm": 2.0623, "queue_k_norm": 2.1496, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3594, "sent_len_1": 66.6769, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.3338, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116000 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4403, "doc_norm": 2.1521, "encoder_q-embeddings": 5101.7837, "encoder_q-layer.0": 3495.637, "encoder_q-layer.1": 3594.8628, "encoder_q-layer.10": 6251.3257, "encoder_q-layer.11": 15279.6162, "encoder_q-layer.2": 4016.0535, "encoder_q-layer.3": 4134.6611, "encoder_q-layer.4": 4391.4927, "encoder_q-layer.5": 4638.5391, "encoder_q-layer.6": 4935.1255, "encoder_q-layer.7": 5518.3677, "encoder_q-layer.8": 6788.0093, "encoder_q-layer.9": 6142.856, "epoch": 0.5, "inbatch_neg_score": 2.7792, "inbatch_pos_score": 3.3691, "learning_rate": 2.2078947368421055e-05, "loss": 3.4403, "norm_diff": 0.0845, "norm_loss": 0.0, "num_token_doc": 66.8044, "num_token_overlap": 11.618, "num_token_query": 31.2721, "num_token_union": 65.1045, "num_word_context": 202.347, "num_word_doc": 49.8357, "num_word_query": 23.2092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9817.6341, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7793, "query_norm": 2.0676, "queue_k_norm": 2.1503, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2721, "sent_len_1": 66.8044, "sent_len_max_0": 127.99, "sent_len_max_1": 210.5238, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116100 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.4094, "doc_norm": 2.1506, "encoder_q-embeddings": 4994.5884, "encoder_q-layer.0": 3552.6892, "encoder_q-layer.1": 3771.6533, "encoder_q-layer.10": 6781.7822, "encoder_q-layer.11": 15609.8232, "encoder_q-layer.2": 4141.6895, "encoder_q-layer.3": 4374.8496, "encoder_q-layer.4": 4538.334, "encoder_q-layer.5": 4770.7427, "encoder_q-layer.6": 5119.335, "encoder_q-layer.7": 5718.7363, "encoder_q-layer.8": 7207.0039, "encoder_q-layer.9": 6732.5527, "epoch": 0.5, "inbatch_neg_score": 2.7786, "inbatch_pos_score": 3.375, "learning_rate": 2.205263157894737e-05, "loss": 3.4094, "norm_diff": 0.0823, "norm_loss": 0.0, "num_token_doc": 66.6781, "num_token_overlap": 11.6625, "num_token_query": 31.3643, "num_token_union": 65.1192, "num_word_context": 202.3996, "num_word_doc": 49.7972, "num_word_query": 23.3132, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9959.4477, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7793, "query_norm": 2.0682, "queue_k_norm": 2.1514, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3643, "sent_len_1": 66.6781, "sent_len_max_0": 127.9737, "sent_len_max_1": 206.4725, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116200 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4152, "doc_norm": 2.1546, "encoder_q-embeddings": 4780.0078, "encoder_q-layer.0": 3317.356, "encoder_q-layer.1": 3458.3159, "encoder_q-layer.10": 6558.5298, "encoder_q-layer.11": 15083.3887, "encoder_q-layer.2": 3850.9355, "encoder_q-layer.3": 3968.2651, "encoder_q-layer.4": 4150.1602, "encoder_q-layer.5": 4313.3818, "encoder_q-layer.6": 4880.0586, "encoder_q-layer.7": 6089.4897, "encoder_q-layer.8": 7163.9302, "encoder_q-layer.9": 6376.5601, "epoch": 0.5, "inbatch_neg_score": 2.7832, "inbatch_pos_score": 3.3984, "learning_rate": 2.2026315789473688e-05, "loss": 3.4152, "norm_diff": 0.0852, "norm_loss": 0.0, "num_token_doc": 66.6336, "num_token_overlap": 11.6726, "num_token_query": 31.3028, "num_token_union": 64.9979, "num_word_context": 202.0859, "num_word_doc": 49.7216, "num_word_query": 23.2409, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9685.404, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7832, "query_norm": 2.0693, "queue_k_norm": 2.151, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3028, "sent_len_1": 66.6336, "sent_len_max_0": 127.97, "sent_len_max_1": 208.1075, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 116300 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.4246, "doc_norm": 2.1526, "encoder_q-embeddings": 4806.9712, "encoder_q-layer.0": 3256.8491, "encoder_q-layer.1": 3427.7395, "encoder_q-layer.10": 6141.7744, "encoder_q-layer.11": 14666.75, "encoder_q-layer.2": 3771.8347, "encoder_q-layer.3": 3842.8506, "encoder_q-layer.4": 4213.0923, "encoder_q-layer.5": 4244.8218, "encoder_q-layer.6": 4913.082, "encoder_q-layer.7": 5490.6118, "encoder_q-layer.8": 6397.2607, "encoder_q-layer.9": 5770.7295, "epoch": 0.51, "inbatch_neg_score": 2.7862, "inbatch_pos_score": 3.3965, "learning_rate": 2.2000000000000003e-05, "loss": 3.4246, "norm_diff": 0.0834, "norm_loss": 0.0, "num_token_doc": 66.7823, "num_token_overlap": 11.6659, "num_token_query": 31.42, "num_token_union": 65.1661, "num_word_context": 202.3403, "num_word_doc": 49.8237, "num_word_query": 23.3304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9258.5566, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7871, "query_norm": 2.0692, "queue_k_norm": 2.152, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.42, "sent_len_1": 66.7823, "sent_len_max_0": 127.9875, "sent_len_max_1": 206.1875, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 116400 }, { "accuracy": 48.7793, "active_queue_size": 16384.0, "cl_loss": 3.4194, "doc_norm": 2.15, "encoder_q-embeddings": 5148.0928, "encoder_q-layer.0": 3485.575, "encoder_q-layer.1": 3926.1208, "encoder_q-layer.10": 6819.0791, "encoder_q-layer.11": 16240.1611, "encoder_q-layer.2": 4314.7515, "encoder_q-layer.3": 4318.5146, "encoder_q-layer.4": 4653.7549, "encoder_q-layer.5": 4740.5249, "encoder_q-layer.6": 5146.2583, "encoder_q-layer.7": 5972.3369, "encoder_q-layer.8": 7267.9863, "encoder_q-layer.9": 6415.2246, "epoch": 0.51, "inbatch_neg_score": 2.7892, "inbatch_pos_score": 3.377, "learning_rate": 2.1973684210526314e-05, "loss": 3.4194, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.6995, "num_token_overlap": 11.6722, "num_token_query": 31.3711, "num_token_union": 65.0642, "num_word_context": 202.3894, "num_word_doc": 49.7533, "num_word_query": 23.2944, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10276.7684, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7891, "query_norm": 2.0706, "queue_k_norm": 2.1536, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3711, "sent_len_1": 66.6995, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.545, "stdk": 0.0491, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116500 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.411, "doc_norm": 2.1521, "encoder_q-embeddings": 4784.2744, "encoder_q-layer.0": 3256.5022, "encoder_q-layer.1": 3513.0859, "encoder_q-layer.10": 6594.3701, "encoder_q-layer.11": 15236.2188, "encoder_q-layer.2": 4007.1272, "encoder_q-layer.3": 4076.8887, "encoder_q-layer.4": 4275.9902, "encoder_q-layer.5": 4546.2646, "encoder_q-layer.6": 5033.0503, "encoder_q-layer.7": 5600.9619, "encoder_q-layer.8": 6794.0815, "encoder_q-layer.9": 6140.4907, "epoch": 0.51, "inbatch_neg_score": 2.7938, "inbatch_pos_score": 3.3965, "learning_rate": 2.1947368421052633e-05, "loss": 3.411, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.9444, "num_token_overlap": 11.6736, "num_token_query": 31.4046, "num_token_union": 65.2178, "num_word_context": 202.3572, "num_word_doc": 49.9319, "num_word_query": 23.3142, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9675.736, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.793, "query_norm": 2.0746, "queue_k_norm": 2.1544, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4046, "sent_len_1": 66.9444, "sent_len_max_0": 127.96, "sent_len_max_1": 208.875, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116600 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.3994, "doc_norm": 2.1503, "encoder_q-embeddings": 5102.7153, "encoder_q-layer.0": 3494.2891, "encoder_q-layer.1": 3744.0664, "encoder_q-layer.10": 6760.894, "encoder_q-layer.11": 16259.1357, "encoder_q-layer.2": 4110.2539, "encoder_q-layer.3": 4368.8159, "encoder_q-layer.4": 4566.8999, "encoder_q-layer.5": 4825.1377, "encoder_q-layer.6": 5478.5503, "encoder_q-layer.7": 5976.8521, "encoder_q-layer.8": 7175.7642, "encoder_q-layer.9": 6659.9546, "epoch": 0.51, "inbatch_neg_score": 2.7988, "inbatch_pos_score": 3.3945, "learning_rate": 2.1921052631578948e-05, "loss": 3.3994, "norm_diff": 0.0608, "norm_loss": 0.0, "num_token_doc": 66.8085, "num_token_overlap": 11.7217, "num_token_query": 31.4561, "num_token_union": 65.1431, "num_word_context": 202.1876, "num_word_doc": 49.8738, "num_word_query": 23.3572, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10267.747, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.7988, "query_norm": 2.0895, "queue_k_norm": 2.1548, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4561, "sent_len_1": 66.8085, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.43, "stdk": 0.0491, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116700 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.4043, "doc_norm": 2.1521, "encoder_q-embeddings": 4994.4551, "encoder_q-layer.0": 3471.2522, "encoder_q-layer.1": 3698.521, "encoder_q-layer.10": 6385.2324, "encoder_q-layer.11": 15472.0, "encoder_q-layer.2": 4299.1318, "encoder_q-layer.3": 4290.8506, "encoder_q-layer.4": 4555.687, "encoder_q-layer.5": 4711.896, "encoder_q-layer.6": 5290.8115, "encoder_q-layer.7": 6010.1123, "encoder_q-layer.8": 6894.2036, "encoder_q-layer.9": 6359.1572, "epoch": 0.51, "inbatch_neg_score": 2.807, "inbatch_pos_score": 3.4102, "learning_rate": 2.1894736842105266e-05, "loss": 3.4043, "norm_diff": 0.0661, "norm_loss": 0.0, "num_token_doc": 66.7676, "num_token_overlap": 11.6484, "num_token_query": 31.3172, "num_token_union": 65.1088, "num_word_context": 202.2576, "num_word_doc": 49.8124, "num_word_query": 23.2718, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10047.7048, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8066, "query_norm": 2.0861, "queue_k_norm": 2.1555, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3172, "sent_len_1": 66.7676, "sent_len_max_0": 127.985, "sent_len_max_1": 209.53, "stdk": 0.0491, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116800 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.4223, "doc_norm": 2.1574, "encoder_q-embeddings": 5278.5601, "encoder_q-layer.0": 3731.4023, "encoder_q-layer.1": 4008.0498, "encoder_q-layer.10": 6812.6738, "encoder_q-layer.11": 15542.4482, "encoder_q-layer.2": 4622.4907, "encoder_q-layer.3": 4791.3164, "encoder_q-layer.4": 5097.5464, "encoder_q-layer.5": 5467.623, "encoder_q-layer.6": 5767.4424, "encoder_q-layer.7": 5921.123, "encoder_q-layer.8": 6871.9243, "encoder_q-layer.9": 6239.7988, "epoch": 0.51, "inbatch_neg_score": 2.8184, "inbatch_pos_score": 3.416, "learning_rate": 2.186842105263158e-05, "loss": 3.4223, "norm_diff": 0.0672, "norm_loss": 0.0, "num_token_doc": 66.8386, "num_token_overlap": 11.6758, "num_token_query": 31.2846, "num_token_union": 65.1308, "num_word_context": 202.3398, "num_word_doc": 49.8866, "num_word_query": 23.2455, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10142.5938, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8184, "query_norm": 2.0902, "queue_k_norm": 2.1561, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2846, "sent_len_1": 66.8386, "sent_len_max_0": 127.99, "sent_len_max_1": 208.3187, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 116900 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.4241, "doc_norm": 2.1588, "encoder_q-embeddings": 4912.8643, "encoder_q-layer.0": 3430.8662, "encoder_q-layer.1": 3664.48, "encoder_q-layer.10": 6655.9194, "encoder_q-layer.11": 15684.3369, "encoder_q-layer.2": 3995.6265, "encoder_q-layer.3": 4039.4768, "encoder_q-layer.4": 4445.1216, "encoder_q-layer.5": 4538.6431, "encoder_q-layer.6": 5226.4351, "encoder_q-layer.7": 5817.6616, "encoder_q-layer.8": 6930.2437, "encoder_q-layer.9": 6315.6348, "epoch": 0.51, "inbatch_neg_score": 2.8312, "inbatch_pos_score": 3.4375, "learning_rate": 2.1842105263157896e-05, "loss": 3.4241, "norm_diff": 0.0591, "norm_loss": 0.0, "num_token_doc": 66.8151, "num_token_overlap": 11.6652, "num_token_query": 31.3733, "num_token_union": 65.1296, "num_word_context": 202.4873, "num_word_doc": 49.8239, "num_word_query": 23.3049, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9979.2559, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.832, "query_norm": 2.0997, "queue_k_norm": 2.1563, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3733, "sent_len_1": 66.8151, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.4062, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 117000 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.4163, "doc_norm": 2.1611, "encoder_q-embeddings": 4791.1475, "encoder_q-layer.0": 3281.7471, "encoder_q-layer.1": 3484.5696, "encoder_q-layer.10": 6480.8438, "encoder_q-layer.11": 14844.7852, "encoder_q-layer.2": 3940.8218, "encoder_q-layer.3": 3997.6208, "encoder_q-layer.4": 4198.1279, "encoder_q-layer.5": 4322.4653, "encoder_q-layer.6": 4809.1177, "encoder_q-layer.7": 5533.8369, "encoder_q-layer.8": 6530.9419, "encoder_q-layer.9": 6131.5547, "epoch": 0.51, "inbatch_neg_score": 2.8412, "inbatch_pos_score": 3.4453, "learning_rate": 2.181578947368421e-05, "loss": 3.4163, "norm_diff": 0.0594, "norm_loss": 0.0, "num_token_doc": 66.6542, "num_token_overlap": 11.6814, "num_token_query": 31.303, "num_token_union": 64.991, "num_word_context": 201.9491, "num_word_doc": 49.7307, "num_word_query": 23.2315, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9528.6262, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8418, "query_norm": 2.1017, "queue_k_norm": 2.1589, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.303, "sent_len_1": 66.6542, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.68, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117100 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.4286, "doc_norm": 2.1583, "encoder_q-embeddings": 7913.2632, "encoder_q-layer.0": 5934.897, "encoder_q-layer.1": 7114.2515, "encoder_q-layer.10": 6195.5962, "encoder_q-layer.11": 15170.1963, "encoder_q-layer.2": 8945.3477, "encoder_q-layer.3": 9851.7744, "encoder_q-layer.4": 10390.9717, "encoder_q-layer.5": 11153.8154, "encoder_q-layer.6": 10835.2871, "encoder_q-layer.7": 9713.252, "encoder_q-layer.8": 8120.9546, "encoder_q-layer.9": 6287.7383, "epoch": 0.51, "inbatch_neg_score": 2.852, "inbatch_pos_score": 3.4434, "learning_rate": 2.1789473684210526e-05, "loss": 3.4286, "norm_diff": 0.0531, "norm_loss": 0.0, "num_token_doc": 66.7733, "num_token_overlap": 11.6362, "num_token_query": 31.2912, "num_token_union": 65.0802, "num_word_context": 202.2768, "num_word_doc": 49.8322, "num_word_query": 23.2422, "postclip_grad_norm": 1.0, "preclip_grad_norm": 14077.2843, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8516, "query_norm": 2.1052, "queue_k_norm": 2.1597, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2912, "sent_len_1": 66.7733, "sent_len_max_0": 127.9525, "sent_len_max_1": 208.8638, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117200 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.427, "doc_norm": 2.1608, "encoder_q-embeddings": 5049.5332, "encoder_q-layer.0": 3535.2029, "encoder_q-layer.1": 3645.9568, "encoder_q-layer.10": 6680.4888, "encoder_q-layer.11": 15499.9092, "encoder_q-layer.2": 4105.7437, "encoder_q-layer.3": 4362.7041, "encoder_q-layer.4": 4611.1235, "encoder_q-layer.5": 4878.0415, "encoder_q-layer.6": 5263.9849, "encoder_q-layer.7": 5963.3184, "encoder_q-layer.8": 7068.147, "encoder_q-layer.9": 6428.9399, "epoch": 0.51, "inbatch_neg_score": 2.8614, "inbatch_pos_score": 3.4648, "learning_rate": 2.1763157894736844e-05, "loss": 3.427, "norm_diff": 0.0563, "norm_loss": 0.0, "num_token_doc": 66.8121, "num_token_overlap": 11.6733, "num_token_query": 31.416, "num_token_union": 65.1532, "num_word_context": 202.2925, "num_word_doc": 49.8487, "num_word_query": 23.3333, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10134.8477, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8613, "query_norm": 2.1045, "queue_k_norm": 2.1598, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.416, "sent_len_1": 66.8121, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.175, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 117300 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4131, "doc_norm": 2.1615, "encoder_q-embeddings": 4768.6475, "encoder_q-layer.0": 3343.7261, "encoder_q-layer.1": 3568.0942, "encoder_q-layer.10": 6938.2461, "encoder_q-layer.11": 16149.5762, "encoder_q-layer.2": 3963.0122, "encoder_q-layer.3": 3997.9883, "encoder_q-layer.4": 4306.542, "encoder_q-layer.5": 4418.8545, "encoder_q-layer.6": 5103.311, "encoder_q-layer.7": 6046.5107, "encoder_q-layer.8": 7197.6963, "encoder_q-layer.9": 6465.896, "epoch": 0.51, "inbatch_neg_score": 2.8703, "inbatch_pos_score": 3.4707, "learning_rate": 2.173684210526316e-05, "loss": 3.4131, "norm_diff": 0.0607, "norm_loss": 0.0, "num_token_doc": 66.8076, "num_token_overlap": 11.6969, "num_token_query": 31.4095, "num_token_union": 65.1663, "num_word_context": 202.3397, "num_word_doc": 49.832, "num_word_query": 23.3205, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10138.299, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8711, "query_norm": 2.1008, "queue_k_norm": 2.1632, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4095, "sent_len_1": 66.8076, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.4725, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117400 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4097, "doc_norm": 2.1627, "encoder_q-embeddings": 4673.7202, "encoder_q-layer.0": 3386.4641, "encoder_q-layer.1": 3577.9846, "encoder_q-layer.10": 6468.498, "encoder_q-layer.11": 15446.9854, "encoder_q-layer.2": 4016.1492, "encoder_q-layer.3": 4023.2766, "encoder_q-layer.4": 4377.0278, "encoder_q-layer.5": 4450.6382, "encoder_q-layer.6": 5039.8662, "encoder_q-layer.7": 6012.9189, "encoder_q-layer.8": 6998.3511, "encoder_q-layer.9": 6195.2075, "epoch": 0.51, "inbatch_neg_score": 2.8788, "inbatch_pos_score": 3.4883, "learning_rate": 2.1710526315789474e-05, "loss": 3.4097, "norm_diff": 0.0518, "norm_loss": 0.0, "num_token_doc": 66.7227, "num_token_overlap": 11.6678, "num_token_query": 31.3494, "num_token_union": 65.065, "num_word_context": 202.2993, "num_word_doc": 49.7923, "num_word_query": 23.2839, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9718.4063, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8789, "query_norm": 2.1108, "queue_k_norm": 2.1643, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3494, "sent_len_1": 66.7227, "sent_len_max_0": 127.99, "sent_len_max_1": 207.7537, "stdk": 0.0492, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117500 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.413, "doc_norm": 2.1652, "encoder_q-embeddings": 4596.7598, "encoder_q-layer.0": 3172.1357, "encoder_q-layer.1": 3346.6099, "encoder_q-layer.10": 6700.1943, "encoder_q-layer.11": 14749.877, "encoder_q-layer.2": 3690.0615, "encoder_q-layer.3": 3835.3015, "encoder_q-layer.4": 4094.8267, "encoder_q-layer.5": 4137.7441, "encoder_q-layer.6": 4577.4077, "encoder_q-layer.7": 5588.9263, "encoder_q-layer.8": 6509.2202, "encoder_q-layer.9": 5970.0981, "epoch": 0.51, "inbatch_neg_score": 2.8815, "inbatch_pos_score": 3.4961, "learning_rate": 2.168421052631579e-05, "loss": 3.413, "norm_diff": 0.0642, "norm_loss": 0.0, "num_token_doc": 66.9116, "num_token_overlap": 11.6779, "num_token_query": 31.3285, "num_token_union": 65.1753, "num_word_context": 202.2924, "num_word_doc": 49.8926, "num_word_query": 23.2736, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9293.3594, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.8809, "query_norm": 2.101, "queue_k_norm": 2.1647, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3285, "sent_len_1": 66.9116, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.35, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 117600 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.436, "doc_norm": 2.1706, "encoder_q-embeddings": 12306.1123, "encoder_q-layer.0": 8768.377, "encoder_q-layer.1": 9850.8027, "encoder_q-layer.10": 12954.1074, "encoder_q-layer.11": 30249.9336, "encoder_q-layer.2": 11491.8379, "encoder_q-layer.3": 12345.9697, "encoder_q-layer.4": 14746.3945, "encoder_q-layer.5": 14836.1074, "encoder_q-layer.6": 14865.4902, "encoder_q-layer.7": 16380.5293, "encoder_q-layer.8": 15191.1406, "encoder_q-layer.9": 12121.2871, "epoch": 0.51, "inbatch_neg_score": 2.8924, "inbatch_pos_score": 3.5039, "learning_rate": 2.1657894736842108e-05, "loss": 3.436, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.7491, "num_token_overlap": 11.6294, "num_token_query": 31.2557, "num_token_union": 65.0713, "num_word_context": 202.1488, "num_word_doc": 49.8024, "num_word_query": 23.2039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22643.7884, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.8926, "query_norm": 2.1036, "queue_k_norm": 2.167, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2557, "sent_len_1": 66.7491, "sent_len_max_0": 127.9475, "sent_len_max_1": 208.91, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 117700 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.421, "doc_norm": 2.1702, "encoder_q-embeddings": 12104.3301, "encoder_q-layer.0": 8673.792, "encoder_q-layer.1": 9954.2129, "encoder_q-layer.10": 12811.6699, "encoder_q-layer.11": 29637.6113, "encoder_q-layer.2": 11637.1221, "encoder_q-layer.3": 12896.4473, "encoder_q-layer.4": 14837.9199, "encoder_q-layer.5": 15018.916, "encoder_q-layer.6": 9643.4414, "encoder_q-layer.7": 10842.4473, "encoder_q-layer.8": 13363.8057, "encoder_q-layer.9": 12129.5928, "epoch": 0.51, "inbatch_neg_score": 2.8942, "inbatch_pos_score": 3.4922, "learning_rate": 2.1631578947368423e-05, "loss": 3.421, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.7026, "num_token_overlap": 11.6534, "num_token_query": 31.3377, "num_token_union": 65.0718, "num_word_context": 202.0471, "num_word_doc": 49.7483, "num_word_query": 23.2755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21633.9696, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.8945, "query_norm": 2.0989, "queue_k_norm": 2.1695, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3377, "sent_len_1": 66.7026, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.6375, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3986, "doc_norm": 2.1716, "encoder_q-embeddings": 10155.9668, "encoder_q-layer.0": 6935.9209, "encoder_q-layer.1": 7286.9209, "encoder_q-layer.10": 13213.6025, "encoder_q-layer.11": 30298.5957, "encoder_q-layer.2": 8145.3701, "encoder_q-layer.3": 8337.832, "encoder_q-layer.4": 9120.7715, "encoder_q-layer.5": 9372.3271, "encoder_q-layer.6": 9956.7266, "encoder_q-layer.7": 11542.6738, "encoder_q-layer.8": 13524.8721, "encoder_q-layer.9": 12126.3867, "epoch": 0.51, "inbatch_neg_score": 2.8965, "inbatch_pos_score": 3.4902, "learning_rate": 2.1605263157894738e-05, "loss": 3.3986, "norm_diff": 0.0805, "norm_loss": 0.0, "num_token_doc": 66.7934, "num_token_overlap": 11.6959, "num_token_query": 31.4465, "num_token_union": 65.1283, "num_word_context": 202.1951, "num_word_doc": 49.8296, "num_word_query": 23.3487, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19436.0703, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.8965, "query_norm": 2.0911, "queue_k_norm": 2.1703, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4465, "sent_len_1": 66.7934, "sent_len_max_0": 127.975, "sent_len_max_1": 209.9137, "stdk": 0.0494, "stdq": 0.0426, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 117900 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.4182, "doc_norm": 2.1742, "encoder_q-embeddings": 10547.2861, "encoder_q-layer.0": 7010.8037, "encoder_q-layer.1": 7329.2607, "encoder_q-layer.10": 13579.4004, "encoder_q-layer.11": 30469.832, "encoder_q-layer.2": 8316.377, "encoder_q-layer.3": 8513.1934, "encoder_q-layer.4": 9124.4414, "encoder_q-layer.5": 9267.3301, "encoder_q-layer.6": 9977.4062, "encoder_q-layer.7": 11084.2188, "encoder_q-layer.8": 12874.8252, "encoder_q-layer.9": 11886.3105, "epoch": 0.51, "inbatch_neg_score": 2.9022, "inbatch_pos_score": 3.5098, "learning_rate": 2.1578947368421053e-05, "loss": 3.4182, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.8613, "num_token_overlap": 11.6671, "num_token_query": 31.3163, "num_token_union": 65.1613, "num_word_context": 202.0738, "num_word_doc": 49.8742, "num_word_query": 23.2613, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19598.2205, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9023, "query_norm": 2.1027, "queue_k_norm": 2.1719, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3163, "sent_len_1": 66.8613, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.6425, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118000 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.4177, "doc_norm": 2.1727, "encoder_q-embeddings": 10908.8838, "encoder_q-layer.0": 7518.5098, "encoder_q-layer.1": 8283.3232, "encoder_q-layer.10": 12468.6719, "encoder_q-layer.11": 30967.4863, "encoder_q-layer.2": 8964.457, "encoder_q-layer.3": 9791.0156, "encoder_q-layer.4": 9757.2412, "encoder_q-layer.5": 10332.7832, "encoder_q-layer.6": 11169.7578, "encoder_q-layer.7": 11967.1953, "encoder_q-layer.8": 14230.8203, "encoder_q-layer.9": 12814.335, "epoch": 0.51, "inbatch_neg_score": 2.9054, "inbatch_pos_score": 3.5176, "learning_rate": 2.1552631578947367e-05, "loss": 3.4177, "norm_diff": 0.0669, "norm_loss": 0.0, "num_token_doc": 66.732, "num_token_overlap": 11.6588, "num_token_query": 31.3741, "num_token_union": 65.0861, "num_word_context": 202.2327, "num_word_doc": 49.8041, "num_word_query": 23.3129, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20704.3655, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9062, "query_norm": 2.1058, "queue_k_norm": 2.174, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3741, "sent_len_1": 66.732, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.96, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118100 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4109, "doc_norm": 2.175, "encoder_q-embeddings": 10299.2129, "encoder_q-layer.0": 7169.7759, "encoder_q-layer.1": 7778.6963, "encoder_q-layer.10": 14721.8779, "encoder_q-layer.11": 33034.8984, "encoder_q-layer.2": 8626.5732, "encoder_q-layer.3": 8963.6006, "encoder_q-layer.4": 9614.6807, "encoder_q-layer.5": 10098.6904, "encoder_q-layer.6": 11132.0361, "encoder_q-layer.7": 13041.0723, "encoder_q-layer.8": 14855.5234, "encoder_q-layer.9": 13966.3203, "epoch": 0.51, "inbatch_neg_score": 2.907, "inbatch_pos_score": 3.5039, "learning_rate": 2.1526315789473686e-05, "loss": 3.4109, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.7825, "num_token_overlap": 11.6702, "num_token_query": 31.3534, "num_token_union": 65.1486, "num_word_context": 202.4362, "num_word_doc": 49.8746, "num_word_query": 23.2951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20893.5449, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9062, "query_norm": 2.1033, "queue_k_norm": 2.1739, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3534, "sent_len_1": 66.7825, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.9575, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 118200 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.4195, "doc_norm": 2.1744, "encoder_q-embeddings": 9990.9053, "encoder_q-layer.0": 6981.8057, "encoder_q-layer.1": 7579.8486, "encoder_q-layer.10": 12438.5078, "encoder_q-layer.11": 31193.6641, "encoder_q-layer.2": 8696.5039, "encoder_q-layer.3": 9018.9951, "encoder_q-layer.4": 9719.834, "encoder_q-layer.5": 9635.4844, "encoder_q-layer.6": 10756.3096, "encoder_q-layer.7": 11910.0674, "encoder_q-layer.8": 13580.584, "encoder_q-layer.9": 12160.2129, "epoch": 0.51, "inbatch_neg_score": 2.9101, "inbatch_pos_score": 3.5215, "learning_rate": 2.15e-05, "loss": 3.4195, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.7439, "num_token_overlap": 11.666, "num_token_query": 31.424, "num_token_union": 65.1364, "num_word_context": 202.2163, "num_word_doc": 49.8043, "num_word_query": 23.3361, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19946.3413, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9102, "query_norm": 2.1045, "queue_k_norm": 2.1759, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.424, "sent_len_1": 66.7439, "sent_len_max_0": 127.98, "sent_len_max_1": 208.435, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118300 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4223, "doc_norm": 2.1746, "encoder_q-embeddings": 4873.707, "encoder_q-layer.0": 3347.4473, "encoder_q-layer.1": 3534.8108, "encoder_q-layer.10": 6308.2378, "encoder_q-layer.11": 14507.6045, "encoder_q-layer.2": 3886.8313, "encoder_q-layer.3": 4009.9924, "encoder_q-layer.4": 4370.9312, "encoder_q-layer.5": 4410.7012, "encoder_q-layer.6": 4899.6289, "encoder_q-layer.7": 5615.7007, "encoder_q-layer.8": 6581.4507, "encoder_q-layer.9": 5862.3486, "epoch": 0.51, "inbatch_neg_score": 2.9135, "inbatch_pos_score": 3.5195, "learning_rate": 2.147368421052632e-05, "loss": 3.4223, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.9198, "num_token_overlap": 11.6461, "num_token_query": 31.3018, "num_token_union": 65.1726, "num_word_context": 202.3755, "num_word_doc": 49.9083, "num_word_query": 23.2588, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9487.2597, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9121, "query_norm": 2.1026, "queue_k_norm": 2.1765, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3018, "sent_len_1": 66.9198, "sent_len_max_0": 127.97, "sent_len_max_1": 211.5987, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118400 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.4089, "doc_norm": 2.1796, "encoder_q-embeddings": 5583.8506, "encoder_q-layer.0": 3788.5034, "encoder_q-layer.1": 4082.3608, "encoder_q-layer.10": 6178.4023, "encoder_q-layer.11": 16092.6562, "encoder_q-layer.2": 4580.6929, "encoder_q-layer.3": 4766.9556, "encoder_q-layer.4": 5323.0679, "encoder_q-layer.5": 5568.3359, "encoder_q-layer.6": 5958.8789, "encoder_q-layer.7": 6601.521, "encoder_q-layer.8": 7122.106, "encoder_q-layer.9": 6122.5781, "epoch": 0.51, "inbatch_neg_score": 2.9158, "inbatch_pos_score": 3.5254, "learning_rate": 2.1447368421052634e-05, "loss": 3.4089, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.7376, "num_token_overlap": 11.6661, "num_token_query": 31.3042, "num_token_union": 65.1038, "num_word_context": 202.2945, "num_word_doc": 49.8108, "num_word_query": 23.2423, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10634.5564, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.916, "query_norm": 2.1001, "queue_k_norm": 2.178, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3042, "sent_len_1": 66.7376, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.3462, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118500 }, { "accuracy": 48.0469, "active_queue_size": 16384.0, "cl_loss": 3.4251, "doc_norm": 2.1778, "encoder_q-embeddings": 4756.8257, "encoder_q-layer.0": 3353.4355, "encoder_q-layer.1": 3483.0164, "encoder_q-layer.10": 7744.9512, "encoder_q-layer.11": 17196.7168, "encoder_q-layer.2": 4035.9163, "encoder_q-layer.3": 4274.7881, "encoder_q-layer.4": 4531.123, "encoder_q-layer.5": 4620.7915, "encoder_q-layer.6": 5308.1304, "encoder_q-layer.7": 5986.2461, "encoder_q-layer.8": 7297.2632, "encoder_q-layer.9": 6761.4185, "epoch": 0.51, "inbatch_neg_score": 2.9181, "inbatch_pos_score": 3.5078, "learning_rate": 2.1421052631578946e-05, "loss": 3.4251, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 67.0186, "num_token_overlap": 11.7035, "num_token_query": 31.3251, "num_token_union": 65.2239, "num_word_context": 202.6404, "num_word_doc": 50.0122, "num_word_query": 23.2756, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10330.4867, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.918, "query_norm": 2.1024, "queue_k_norm": 2.1796, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3251, "sent_len_1": 67.0186, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.8137, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118600 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.4207, "doc_norm": 2.1805, "encoder_q-embeddings": 5012.5601, "encoder_q-layer.0": 3455.0596, "encoder_q-layer.1": 3736.0212, "encoder_q-layer.10": 6317.6865, "encoder_q-layer.11": 14700.4287, "encoder_q-layer.2": 4171.8384, "encoder_q-layer.3": 4466.0664, "encoder_q-layer.4": 4732.48, "encoder_q-layer.5": 4734.9658, "encoder_q-layer.6": 5184.7783, "encoder_q-layer.7": 5849.6528, "encoder_q-layer.8": 6606.1333, "encoder_q-layer.9": 6028.8799, "epoch": 0.52, "inbatch_neg_score": 2.92, "inbatch_pos_score": 3.5273, "learning_rate": 2.1394736842105264e-05, "loss": 3.4207, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.6795, "num_token_overlap": 11.6306, "num_token_query": 31.2313, "num_token_union": 65.0018, "num_word_context": 202.3383, "num_word_doc": 49.7679, "num_word_query": 23.1901, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9642.2503, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9199, "query_norm": 2.1037, "queue_k_norm": 2.1807, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2313, "sent_len_1": 66.6795, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.2188, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118700 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.4189, "doc_norm": 2.1826, "encoder_q-embeddings": 4963.5542, "encoder_q-layer.0": 3600.085, "encoder_q-layer.1": 3901.3296, "encoder_q-layer.10": 6493.8765, "encoder_q-layer.11": 15219.9258, "encoder_q-layer.2": 4408.4932, "encoder_q-layer.3": 4601.8501, "encoder_q-layer.4": 4902.4546, "encoder_q-layer.5": 5036.1479, "encoder_q-layer.6": 5468.8447, "encoder_q-layer.7": 5779.3911, "encoder_q-layer.8": 6679.4185, "encoder_q-layer.9": 6253.4307, "epoch": 0.52, "inbatch_neg_score": 2.9245, "inbatch_pos_score": 3.5371, "learning_rate": 2.136842105263158e-05, "loss": 3.4189, "norm_diff": 0.0828, "norm_loss": 0.0, "num_token_doc": 66.8146, "num_token_overlap": 11.6665, "num_token_query": 31.3595, "num_token_union": 65.1376, "num_word_context": 202.2718, "num_word_doc": 49.8502, "num_word_query": 23.302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9980.8313, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9238, "query_norm": 2.0998, "queue_k_norm": 2.181, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3595, "sent_len_1": 66.8146, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.6425, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118800 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.4074, "doc_norm": 2.1812, "encoder_q-embeddings": 5023.7544, "encoder_q-layer.0": 3444.1094, "encoder_q-layer.1": 3685.9216, "encoder_q-layer.10": 6221.1797, "encoder_q-layer.11": 15428.5859, "encoder_q-layer.2": 3986.8372, "encoder_q-layer.3": 4133.7627, "encoder_q-layer.4": 4420.1504, "encoder_q-layer.5": 4532.0439, "encoder_q-layer.6": 5076.957, "encoder_q-layer.7": 5827.0967, "encoder_q-layer.8": 6880.1553, "encoder_q-layer.9": 6237.8857, "epoch": 0.52, "inbatch_neg_score": 2.9252, "inbatch_pos_score": 3.5312, "learning_rate": 2.1342105263157897e-05, "loss": 3.4074, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.6015, "num_token_overlap": 11.642, "num_token_query": 31.326, "num_token_union": 64.9809, "num_word_context": 202.1656, "num_word_doc": 49.6816, "num_word_query": 23.2516, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9891.2486, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9258, "query_norm": 2.1039, "queue_k_norm": 2.1819, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.326, "sent_len_1": 66.6015, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.1488, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 118900 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.4247, "doc_norm": 2.1837, "encoder_q-embeddings": 5159.957, "encoder_q-layer.0": 3638.2957, "encoder_q-layer.1": 4000.1047, "encoder_q-layer.10": 6568.2769, "encoder_q-layer.11": 15506.1934, "encoder_q-layer.2": 4467.7324, "encoder_q-layer.3": 4968.8345, "encoder_q-layer.4": 5508.1836, "encoder_q-layer.5": 5556.6401, "encoder_q-layer.6": 5676.2246, "encoder_q-layer.7": 6868.2334, "encoder_q-layer.8": 7398.3564, "encoder_q-layer.9": 6686.8379, "epoch": 0.52, "inbatch_neg_score": 2.9298, "inbatch_pos_score": 3.5469, "learning_rate": 2.1315789473684212e-05, "loss": 3.4247, "norm_diff": 0.0797, "norm_loss": 0.0, "num_token_doc": 66.6512, "num_token_overlap": 11.6617, "num_token_query": 31.3634, "num_token_union": 65.0335, "num_word_context": 202.0906, "num_word_doc": 49.6745, "num_word_query": 23.2991, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10473.7812, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9297, "query_norm": 2.1041, "queue_k_norm": 2.1845, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3634, "sent_len_1": 66.6512, "sent_len_max_0": 127.9775, "sent_len_max_1": 212.1287, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 119000 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4164, "doc_norm": 2.1823, "encoder_q-embeddings": 4873.3262, "encoder_q-layer.0": 3367.2368, "encoder_q-layer.1": 3602.3113, "encoder_q-layer.10": 7003.2085, "encoder_q-layer.11": 15365.0078, "encoder_q-layer.2": 4143.1597, "encoder_q-layer.3": 4194.1108, "encoder_q-layer.4": 4627.5859, "encoder_q-layer.5": 4786.0195, "encoder_q-layer.6": 5208.6958, "encoder_q-layer.7": 5793.7324, "encoder_q-layer.8": 6810.2451, "encoder_q-layer.9": 5833.1899, "epoch": 0.52, "inbatch_neg_score": 2.9322, "inbatch_pos_score": 3.5254, "learning_rate": 2.1289473684210527e-05, "loss": 3.4164, "norm_diff": 0.084, "norm_loss": 0.0, "num_token_doc": 66.5315, "num_token_overlap": 11.6121, "num_token_query": 31.251, "num_token_union": 64.9333, "num_word_context": 201.8042, "num_word_doc": 49.6544, "num_word_query": 23.1971, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9946.0492, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9316, "query_norm": 2.0983, "queue_k_norm": 2.1834, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.251, "sent_len_1": 66.5315, "sent_len_max_0": 127.9975, "sent_len_max_1": 206.8638, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119100 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.4202, "doc_norm": 2.1897, "encoder_q-embeddings": 8725.2529, "encoder_q-layer.0": 5921.2847, "encoder_q-layer.1": 6561.6841, "encoder_q-layer.10": 7000.7803, "encoder_q-layer.11": 15433.4199, "encoder_q-layer.2": 7845.0879, "encoder_q-layer.3": 7744.7769, "encoder_q-layer.4": 8528.6992, "encoder_q-layer.5": 8763.3311, "encoder_q-layer.6": 10176.0811, "encoder_q-layer.7": 9945.1025, "encoder_q-layer.8": 8437.2148, "encoder_q-layer.9": 6616.3208, "epoch": 0.52, "inbatch_neg_score": 2.9353, "inbatch_pos_score": 3.5469, "learning_rate": 2.1263157894736842e-05, "loss": 3.4202, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.9767, "num_token_overlap": 11.6924, "num_token_query": 31.386, "num_token_union": 65.2093, "num_word_context": 202.4866, "num_word_doc": 49.9618, "num_word_query": 23.31, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13254.5638, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9336, "query_norm": 2.1117, "queue_k_norm": 2.1861, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.386, "sent_len_1": 66.9767, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.6788, "stdk": 0.0496, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 119200 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.4223, "doc_norm": 2.1871, "encoder_q-embeddings": 5690.0415, "encoder_q-layer.0": 4194.2207, "encoder_q-layer.1": 4638.7334, "encoder_q-layer.10": 6341.7407, "encoder_q-layer.11": 15205.0234, "encoder_q-layer.2": 5210.5034, "encoder_q-layer.3": 5474.9863, "encoder_q-layer.4": 4967.437, "encoder_q-layer.5": 4902.7217, "encoder_q-layer.6": 5314.6069, "encoder_q-layer.7": 5921.7524, "encoder_q-layer.8": 7023.6318, "encoder_q-layer.9": 6281.5127, "epoch": 0.52, "inbatch_neg_score": 2.9368, "inbatch_pos_score": 3.5449, "learning_rate": 2.1236842105263157e-05, "loss": 3.4223, "norm_diff": 0.082, "norm_loss": 0.0, "num_token_doc": 66.5532, "num_token_overlap": 11.6695, "num_token_query": 31.4044, "num_token_union": 64.9969, "num_word_context": 202.2271, "num_word_doc": 49.6806, "num_word_query": 23.3228, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10353.2394, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9375, "query_norm": 2.1051, "queue_k_norm": 2.1862, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4044, "sent_len_1": 66.5532, "sent_len_max_0": 127.98, "sent_len_max_1": 208.9462, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119300 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.423, "doc_norm": 2.1854, "encoder_q-embeddings": 5274.7002, "encoder_q-layer.0": 3650.9856, "encoder_q-layer.1": 3829.8022, "encoder_q-layer.10": 6075.8945, "encoder_q-layer.11": 15498.457, "encoder_q-layer.2": 4173.4399, "encoder_q-layer.3": 4171.917, "encoder_q-layer.4": 4646.9106, "encoder_q-layer.5": 4640.708, "encoder_q-layer.6": 4949.979, "encoder_q-layer.7": 6196.5728, "encoder_q-layer.8": 6652.1914, "encoder_q-layer.9": 6018.5708, "epoch": 0.52, "inbatch_neg_score": 2.9403, "inbatch_pos_score": 3.5371, "learning_rate": 2.1210526315789476e-05, "loss": 3.423, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.7127, "num_token_overlap": 11.664, "num_token_query": 31.3099, "num_token_union": 65.0394, "num_word_context": 202.0589, "num_word_doc": 49.7879, "num_word_query": 23.25, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10040.6378, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9414, "query_norm": 2.1008, "queue_k_norm": 2.1868, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3099, "sent_len_1": 66.7127, "sent_len_max_0": 127.98, "sent_len_max_1": 209.3938, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119400 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.4171, "doc_norm": 2.1923, "encoder_q-embeddings": 4954.9429, "encoder_q-layer.0": 3389.4153, "encoder_q-layer.1": 3693.3633, "encoder_q-layer.10": 6315.4307, "encoder_q-layer.11": 15762.4805, "encoder_q-layer.2": 4043.0171, "encoder_q-layer.3": 4161.8442, "encoder_q-layer.4": 4533.6123, "encoder_q-layer.5": 4547.7778, "encoder_q-layer.6": 5156.2036, "encoder_q-layer.7": 5841.7368, "encoder_q-layer.8": 6876.4312, "encoder_q-layer.9": 6163.5244, "epoch": 0.52, "inbatch_neg_score": 2.9464, "inbatch_pos_score": 3.5469, "learning_rate": 2.118421052631579e-05, "loss": 3.4171, "norm_diff": 0.0849, "norm_loss": 0.0, "num_token_doc": 66.7463, "num_token_overlap": 11.6705, "num_token_query": 31.3894, "num_token_union": 65.108, "num_word_context": 202.2526, "num_word_doc": 49.8158, "num_word_query": 23.3198, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9987.5457, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9473, "query_norm": 2.1074, "queue_k_norm": 2.1875, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3894, "sent_len_1": 66.7463, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.0563, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119500 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.4104, "doc_norm": 2.1881, "encoder_q-embeddings": 5485.6406, "encoder_q-layer.0": 3899.3418, "encoder_q-layer.1": 4047.2637, "encoder_q-layer.10": 7064.0947, "encoder_q-layer.11": 15784.417, "encoder_q-layer.2": 4636.2651, "encoder_q-layer.3": 4961.6826, "encoder_q-layer.4": 5418.4805, "encoder_q-layer.5": 5813.7974, "encoder_q-layer.6": 6180.9756, "encoder_q-layer.7": 7164.4678, "encoder_q-layer.8": 7232.4272, "encoder_q-layer.9": 6610.9014, "epoch": 0.52, "inbatch_neg_score": 2.9479, "inbatch_pos_score": 3.5664, "learning_rate": 2.1157894736842106e-05, "loss": 3.4104, "norm_diff": 0.0684, "norm_loss": 0.0, "num_token_doc": 66.8055, "num_token_overlap": 11.6913, "num_token_query": 31.436, "num_token_union": 65.189, "num_word_context": 202.3862, "num_word_doc": 49.8699, "num_word_query": 23.3597, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10499.5278, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9473, "query_norm": 2.1197, "queue_k_norm": 2.1875, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.436, "sent_len_1": 66.8055, "sent_len_max_0": 127.9562, "sent_len_max_1": 209.5175, "stdk": 0.0493, "stdq": 0.0438, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 119600 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.406, "doc_norm": 2.1858, "encoder_q-embeddings": 4662.5601, "encoder_q-layer.0": 3407.1655, "encoder_q-layer.1": 3458.8752, "encoder_q-layer.10": 6697.6089, "encoder_q-layer.11": 14921.9893, "encoder_q-layer.2": 3839.2124, "encoder_q-layer.3": 3923.8521, "encoder_q-layer.4": 4243.0493, "encoder_q-layer.5": 4351.6206, "encoder_q-layer.6": 4657.3647, "encoder_q-layer.7": 5205.1826, "encoder_q-layer.8": 6344.0073, "encoder_q-layer.9": 5864.4897, "epoch": 0.52, "inbatch_neg_score": 2.9512, "inbatch_pos_score": 3.5645, "learning_rate": 2.113157894736842e-05, "loss": 3.406, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.8916, "num_token_overlap": 11.6213, "num_token_query": 31.2118, "num_token_union": 65.1395, "num_word_context": 202.5673, "num_word_doc": 49.9123, "num_word_query": 23.1558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9413.6548, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9512, "query_norm": 2.1125, "queue_k_norm": 2.1889, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2118, "sent_len_1": 66.8916, "sent_len_max_0": 127.9788, "sent_len_max_1": 207.0813, "stdk": 0.0491, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119700 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4029, "doc_norm": 2.19, "encoder_q-embeddings": 5185.6982, "encoder_q-layer.0": 3664.0759, "encoder_q-layer.1": 4108.7554, "encoder_q-layer.10": 6307.335, "encoder_q-layer.11": 15134.4775, "encoder_q-layer.2": 4609.25, "encoder_q-layer.3": 4717.2212, "encoder_q-layer.4": 5158.3979, "encoder_q-layer.5": 5182.4189, "encoder_q-layer.6": 5844.3696, "encoder_q-layer.7": 6073.7568, "encoder_q-layer.8": 6854.0117, "encoder_q-layer.9": 6173.1235, "epoch": 0.52, "inbatch_neg_score": 2.9531, "inbatch_pos_score": 3.5586, "learning_rate": 2.110526315789474e-05, "loss": 3.4029, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7348, "num_token_overlap": 11.6523, "num_token_query": 31.3264, "num_token_union": 65.0872, "num_word_context": 202.2041, "num_word_doc": 49.7952, "num_word_query": 23.2634, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10189.1767, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9531, "query_norm": 2.1104, "queue_k_norm": 2.1892, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3264, "sent_len_1": 66.7348, "sent_len_max_0": 127.97, "sent_len_max_1": 209.95, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119800 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.412, "doc_norm": 2.1925, "encoder_q-embeddings": 4711.0425, "encoder_q-layer.0": 3335.6565, "encoder_q-layer.1": 3579.1562, "encoder_q-layer.10": 6564.7754, "encoder_q-layer.11": 15260.9375, "encoder_q-layer.2": 3945.4133, "encoder_q-layer.3": 4130.5571, "encoder_q-layer.4": 4284.3149, "encoder_q-layer.5": 4360.4663, "encoder_q-layer.6": 4893.0972, "encoder_q-layer.7": 5408.728, "encoder_q-layer.8": 6776.5845, "encoder_q-layer.9": 6117.4019, "epoch": 0.52, "inbatch_neg_score": 2.9588, "inbatch_pos_score": 3.5645, "learning_rate": 2.1078947368421054e-05, "loss": 3.412, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.9405, "num_token_overlap": 11.6925, "num_token_query": 31.3573, "num_token_union": 65.1858, "num_word_context": 202.5657, "num_word_doc": 49.9576, "num_word_query": 23.3061, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9617.6252, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.957, "query_norm": 2.1141, "queue_k_norm": 2.1907, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3573, "sent_len_1": 66.9405, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.9837, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 119900 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4053, "doc_norm": 2.1906, "encoder_q-embeddings": 4828.6128, "encoder_q-layer.0": 3428.675, "encoder_q-layer.1": 3737.1218, "encoder_q-layer.10": 6855.4463, "encoder_q-layer.11": 16196.7412, "encoder_q-layer.2": 4193.0356, "encoder_q-layer.3": 4220.6885, "encoder_q-layer.4": 4435.8696, "encoder_q-layer.5": 4624.4897, "encoder_q-layer.6": 5138.7871, "encoder_q-layer.7": 5684.481, "encoder_q-layer.8": 6844.395, "encoder_q-layer.9": 6521.2036, "epoch": 0.52, "inbatch_neg_score": 2.9617, "inbatch_pos_score": 3.5664, "learning_rate": 2.105263157894737e-05, "loss": 3.4053, "norm_diff": 0.0734, "norm_loss": 0.0, "num_token_doc": 66.8423, "num_token_overlap": 11.6817, "num_token_query": 31.3342, "num_token_union": 65.0932, "num_word_context": 202.3705, "num_word_doc": 49.8655, "num_word_query": 23.2786, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10002.5835, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9609, "query_norm": 2.1172, "queue_k_norm": 2.1905, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3342, "sent_len_1": 66.8423, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.21, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 120000 }, { "dev_runtime": 28.9407, "dev_samples_per_second": 1.106, "dev_steps_per_second": 0.035, "epoch": 0.52, "step": 120000, "test_accuracy": 94.00634765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3361322581768036, "test_doc_norm": 2.172987222671509, "test_inbatch_neg_score": 3.245227336883545, "test_inbatch_pos_score": 4.270808219909668, "test_loss": 0.3361322581768036, "test_loss_align": 1.0378576517105103, "test_loss_unif": -12.944015502929688, "test_loss_unif_q@queue": -12.944015502929688, "test_norm_diff": 0.027903735637664795, "test_norm_loss": 0.0, "test_q@queue_neg_score": 2.964024543762207, "test_query_norm": 2.2008910179138184, "test_queue_k_norm": 2.1906793117523193, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04413545876741409, "test_stdq": 0.044152792543172836, "test_stdqueue_k": 0.049348633736371994, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.9407, "dev_samples_per_second": 1.106, "dev_steps_per_second": 0.035, "epoch": 0.52, "eval_beir-arguana_ndcg@10": 0.39856, "eval_beir-arguana_recall@10": 0.67639, "eval_beir-arguana_recall@100": 0.94666, "eval_beir-arguana_recall@20": 0.80583, "eval_beir-avg_ndcg@10": 0.39836658333333336, "eval_beir-avg_recall@10": 0.46876125, "eval_beir-avg_recall@100": 0.650735, "eval_beir-avg_recall@20": 0.5324384166666667, "eval_beir-cqadupstack_ndcg@10": 0.2903958333333333, "eval_beir-cqadupstack_recall@10": 0.3878825000000001, "eval_beir-cqadupstack_recall@100": 0.62447, "eval_beir-cqadupstack_recall@20": 0.45635416666666667, "eval_beir-fiqa_ndcg@10": 0.264, "eval_beir-fiqa_recall@10": 0.33605, "eval_beir-fiqa_recall@100": 0.61013, "eval_beir-fiqa_recall@20": 0.40682, "eval_beir-nfcorpus_ndcg@10": 0.31345, "eval_beir-nfcorpus_recall@10": 0.15265, "eval_beir-nfcorpus_recall@100": 0.29863, "eval_beir-nfcorpus_recall@20": 0.19286, "eval_beir-nq_ndcg@10": 0.33015, "eval_beir-nq_recall@10": 0.52549, "eval_beir-nq_recall@100": 0.84171, "eval_beir-nq_recall@20": 0.6389, "eval_beir-quora_ndcg@10": 0.79033, "eval_beir-quora_recall@10": 0.89659, "eval_beir-quora_recall@100": 0.98065, "eval_beir-quora_recall@20": 0.93528, "eval_beir-scidocs_ndcg@10": 0.15991, "eval_beir-scidocs_recall@10": 0.16712, "eval_beir-scidocs_recall@100": 0.38467, "eval_beir-scidocs_recall@20": 0.23315, "eval_beir-scifact_ndcg@10": 0.66006, "eval_beir-scifact_recall@10": 0.789, "eval_beir-scifact_recall@100": 0.92489, "eval_beir-scifact_recall@20": 0.83244, "eval_beir-trec-covid_ndcg@10": 0.58925, "eval_beir-trec-covid_recall@10": 0.622, "eval_beir-trec-covid_recall@100": 0.4582, "eval_beir-trec-covid_recall@20": 0.605, "eval_beir-webis-touche2020_ndcg@10": 0.18756, "eval_beir-webis-touche2020_recall@10": 0.13444, "eval_beir-webis-touche2020_recall@100": 0.43734, "eval_beir-webis-touche2020_recall@20": 0.21775, "eval_senteval-avg_sts": 0.7348425475923509, "eval_senteval-sickr_spearman": 0.7321203077136191, "eval_senteval-stsb_spearman": 0.7375647874710827, "step": 120000, "test_accuracy": 94.00634765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3361322581768036, "test_doc_norm": 2.172987222671509, "test_inbatch_neg_score": 3.245227336883545, "test_inbatch_pos_score": 4.270808219909668, "test_loss": 0.3361322581768036, "test_loss_align": 1.0378576517105103, "test_loss_unif": -12.944015502929688, "test_loss_unif_q@queue": -12.944015502929688, "test_norm_diff": 0.027903735637664795, "test_norm_loss": 0.0, "test_q@queue_neg_score": 2.964024543762207, "test_query_norm": 2.2008910179138184, "test_queue_k_norm": 2.1906793117523193, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04413545876741409, "test_stdq": 0.044152792543172836, "test_stdqueue_k": 0.049348633736371994, "test_stdqueue_q": 0.0 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.4111, "doc_norm": 2.194, "encoder_q-embeddings": 5350.2876, "encoder_q-layer.0": 3733.3853, "encoder_q-layer.1": 4029.3606, "encoder_q-layer.10": 7370.624, "encoder_q-layer.11": 15192.498, "encoder_q-layer.2": 4640.9487, "encoder_q-layer.3": 4833.0596, "encoder_q-layer.4": 5386.1899, "encoder_q-layer.5": 5475.6045, "encoder_q-layer.6": 6049.9956, "encoder_q-layer.7": 6855.248, "encoder_q-layer.8": 7895.4922, "encoder_q-layer.9": 6656.1396, "epoch": 0.52, "inbatch_neg_score": 2.9655, "inbatch_pos_score": 3.5801, "learning_rate": 2.1026315789473687e-05, "loss": 3.4111, "norm_diff": 0.0835, "norm_loss": 0.0, "num_token_doc": 66.6592, "num_token_overlap": 11.724, "num_token_query": 31.5291, "num_token_union": 65.1117, "num_word_context": 202.1575, "num_word_doc": 49.7366, "num_word_query": 23.4276, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10397.0808, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9648, "query_norm": 2.1105, "queue_k_norm": 2.1926, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5291, "sent_len_1": 66.6592, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.5337, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120100 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.408, "doc_norm": 2.1934, "encoder_q-embeddings": 5467.0542, "encoder_q-layer.0": 3967.6738, "encoder_q-layer.1": 4368.73, "encoder_q-layer.10": 6529.9927, "encoder_q-layer.11": 14827.4551, "encoder_q-layer.2": 4771.6323, "encoder_q-layer.3": 4840.2773, "encoder_q-layer.4": 4984.1919, "encoder_q-layer.5": 5153.627, "encoder_q-layer.6": 5520.3628, "encoder_q-layer.7": 5901.6611, "encoder_q-layer.8": 6685.2715, "encoder_q-layer.9": 6040.3223, "epoch": 0.52, "inbatch_neg_score": 2.9739, "inbatch_pos_score": 3.5859, "learning_rate": 2.1e-05, "loss": 3.408, "norm_diff": 0.0753, "norm_loss": 0.0, "num_token_doc": 66.869, "num_token_overlap": 11.7011, "num_token_query": 31.4215, "num_token_union": 65.173, "num_word_context": 202.3763, "num_word_doc": 49.8925, "num_word_query": 23.3491, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10055.3526, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9746, "query_norm": 2.1181, "queue_k_norm": 2.1935, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4215, "sent_len_1": 66.869, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.8338, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120200 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.4104, "doc_norm": 2.1982, "encoder_q-embeddings": 5478.3486, "encoder_q-layer.0": 3794.1506, "encoder_q-layer.1": 4222.231, "encoder_q-layer.10": 7061.2944, "encoder_q-layer.11": 16699.4395, "encoder_q-layer.2": 4757.7524, "encoder_q-layer.3": 5256.3042, "encoder_q-layer.4": 5488.1343, "encoder_q-layer.5": 5878.7788, "encoder_q-layer.6": 6256.1963, "encoder_q-layer.7": 6258.8657, "encoder_q-layer.8": 7554.2783, "encoder_q-layer.9": 6553.2402, "epoch": 0.52, "inbatch_neg_score": 2.9783, "inbatch_pos_score": 3.5723, "learning_rate": 2.0973684210526317e-05, "loss": 3.4104, "norm_diff": 0.0815, "norm_loss": 0.0, "num_token_doc": 66.6808, "num_token_overlap": 11.6879, "num_token_query": 31.3871, "num_token_union": 65.062, "num_word_context": 202.3916, "num_word_doc": 49.7394, "num_word_query": 23.3046, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11062.8326, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 2.9785, "query_norm": 2.1167, "queue_k_norm": 2.1936, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3871, "sent_len_1": 66.6808, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.2425, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120300 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.394, "doc_norm": 2.1938, "encoder_q-embeddings": 11023.0547, "encoder_q-layer.0": 7741.1514, "encoder_q-layer.1": 8273.8447, "encoder_q-layer.10": 13602.4688, "encoder_q-layer.11": 31348.7344, "encoder_q-layer.2": 9801.2988, "encoder_q-layer.3": 10258.2051, "encoder_q-layer.4": 10563.3076, "encoder_q-layer.5": 11089.7812, "encoder_q-layer.6": 11516.9531, "encoder_q-layer.7": 13181.4277, "encoder_q-layer.8": 15117.0967, "encoder_q-layer.9": 12856.2568, "epoch": 0.52, "inbatch_neg_score": 2.9805, "inbatch_pos_score": 3.5938, "learning_rate": 2.0947368421052632e-05, "loss": 3.394, "norm_diff": 0.0635, "norm_loss": 0.0, "num_token_doc": 66.7825, "num_token_overlap": 11.7001, "num_token_query": 31.4457, "num_token_union": 65.1282, "num_word_context": 202.152, "num_word_doc": 49.8385, "num_word_query": 23.3628, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20941.7253, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9805, "query_norm": 2.1303, "queue_k_norm": 2.1949, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4457, "sent_len_1": 66.7825, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.68, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120400 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4183, "doc_norm": 2.1971, "encoder_q-embeddings": 9929.4385, "encoder_q-layer.0": 6956.9658, "encoder_q-layer.1": 7209.2827, "encoder_q-layer.10": 12797.624, "encoder_q-layer.11": 30413.4531, "encoder_q-layer.2": 8008.668, "encoder_q-layer.3": 8338.8945, "encoder_q-layer.4": 9074.1123, "encoder_q-layer.5": 9796.7373, "encoder_q-layer.6": 10488.3545, "encoder_q-layer.7": 11790.2861, "encoder_q-layer.8": 14384.4434, "encoder_q-layer.9": 12785.4688, "epoch": 0.52, "inbatch_neg_score": 2.9819, "inbatch_pos_score": 3.6035, "learning_rate": 2.092105263157895e-05, "loss": 3.4183, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.7777, "num_token_overlap": 11.6178, "num_token_query": 31.2239, "num_token_union": 65.1096, "num_word_context": 202.315, "num_word_doc": 49.8412, "num_word_query": 23.1672, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19726.3545, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9824, "query_norm": 2.1312, "queue_k_norm": 2.1951, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2239, "sent_len_1": 66.7777, "sent_len_max_0": 127.9638, "sent_len_max_1": 206.4112, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120500 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.4123, "doc_norm": 2.1959, "encoder_q-embeddings": 9742.2461, "encoder_q-layer.0": 6803.939, "encoder_q-layer.1": 7331.7739, "encoder_q-layer.10": 12445.416, "encoder_q-layer.11": 30455.2832, "encoder_q-layer.2": 8346.7061, "encoder_q-layer.3": 8413.3398, "encoder_q-layer.4": 9276.832, "encoder_q-layer.5": 8987.9141, "encoder_q-layer.6": 9806.6328, "encoder_q-layer.7": 11040.3467, "encoder_q-layer.8": 13040.4092, "encoder_q-layer.9": 12004.5225, "epoch": 0.52, "inbatch_neg_score": 2.9843, "inbatch_pos_score": 3.5977, "learning_rate": 2.0894736842105266e-05, "loss": 3.4123, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.7969, "num_token_overlap": 11.6542, "num_token_query": 31.315, "num_token_union": 65.1195, "num_word_context": 202.041, "num_word_doc": 49.8191, "num_word_query": 23.2552, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19147.0361, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9844, "query_norm": 2.1304, "queue_k_norm": 2.1972, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.315, "sent_len_1": 66.7969, "sent_len_max_0": 127.9062, "sent_len_max_1": 209.3475, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120600 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4113, "doc_norm": 2.197, "encoder_q-embeddings": 10379.0, "encoder_q-layer.0": 7406.8975, "encoder_q-layer.1": 8005.3525, "encoder_q-layer.10": 12317.6758, "encoder_q-layer.11": 29837.6934, "encoder_q-layer.2": 8896.9824, "encoder_q-layer.3": 9143.6318, "encoder_q-layer.4": 9748.7666, "encoder_q-layer.5": 10249.2285, "encoder_q-layer.6": 11227.79, "encoder_q-layer.7": 11990.4824, "encoder_q-layer.8": 13383.4219, "encoder_q-layer.9": 12312.335, "epoch": 0.52, "inbatch_neg_score": 2.9871, "inbatch_pos_score": 3.5879, "learning_rate": 2.0868421052631577e-05, "loss": 3.4113, "norm_diff": 0.0693, "norm_loss": 0.0, "num_token_doc": 66.6906, "num_token_overlap": 11.681, "num_token_query": 31.3736, "num_token_union": 65.0999, "num_word_context": 202.4238, "num_word_doc": 49.7843, "num_word_query": 23.3005, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19776.6002, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9863, "query_norm": 2.1277, "queue_k_norm": 2.1957, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3736, "sent_len_1": 66.6906, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.6337, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 120700 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.4008, "doc_norm": 2.1965, "encoder_q-embeddings": 9489.0049, "encoder_q-layer.0": 6552.9917, "encoder_q-layer.1": 6949.1899, "encoder_q-layer.10": 13077.1992, "encoder_q-layer.11": 31056.6094, "encoder_q-layer.2": 7805.2241, "encoder_q-layer.3": 7816.5684, "encoder_q-layer.4": 8563.8193, "encoder_q-layer.5": 8753.2549, "encoder_q-layer.6": 10114.7217, "encoder_q-layer.7": 11681.7236, "encoder_q-layer.8": 13267.2207, "encoder_q-layer.9": 12853.4023, "epoch": 0.52, "inbatch_neg_score": 2.9919, "inbatch_pos_score": 3.5938, "learning_rate": 2.0842105263157895e-05, "loss": 3.4008, "norm_diff": 0.0688, "norm_loss": 0.0, "num_token_doc": 66.7957, "num_token_overlap": 11.7168, "num_token_query": 31.4811, "num_token_union": 65.1749, "num_word_context": 202.5028, "num_word_doc": 49.821, "num_word_query": 23.3863, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19386.6676, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 2.9902, "query_norm": 2.1277, "queue_k_norm": 2.1989, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4811, "sent_len_1": 66.7957, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.9512, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120800 }, { "accuracy": 53.0762, "active_queue_size": 16384.0, "cl_loss": 3.4014, "doc_norm": 2.2036, "encoder_q-embeddings": 9891.4707, "encoder_q-layer.0": 6863.0327, "encoder_q-layer.1": 7387.5522, "encoder_q-layer.10": 12389.6592, "encoder_q-layer.11": 29998.0117, "encoder_q-layer.2": 8347.1221, "encoder_q-layer.3": 8945.6338, "encoder_q-layer.4": 9349.541, "encoder_q-layer.5": 9586.7051, "encoder_q-layer.6": 10524.96, "encoder_q-layer.7": 11858.0928, "encoder_q-layer.8": 13799.6035, "encoder_q-layer.9": 11979.3809, "epoch": 0.52, "inbatch_neg_score": 3.0017, "inbatch_pos_score": 3.6387, "learning_rate": 2.081578947368421e-05, "loss": 3.4014, "norm_diff": 0.063, "norm_loss": 0.0, "num_token_doc": 66.8353, "num_token_overlap": 11.6806, "num_token_query": 31.4397, "num_token_union": 65.2136, "num_word_context": 202.7802, "num_word_doc": 49.896, "num_word_query": 23.3515, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19558.7655, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.002, "query_norm": 2.1405, "queue_k_norm": 2.1996, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4397, "sent_len_1": 66.8353, "sent_len_max_0": 127.985, "sent_len_max_1": 207.4875, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 120900 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4073, "doc_norm": 2.1967, "encoder_q-embeddings": 9929.1787, "encoder_q-layer.0": 6735.1162, "encoder_q-layer.1": 7136.5161, "encoder_q-layer.10": 14632.4609, "encoder_q-layer.11": 31425.9805, "encoder_q-layer.2": 7985.0464, "encoder_q-layer.3": 8292.7051, "encoder_q-layer.4": 8775.8848, "encoder_q-layer.5": 9319.1777, "encoder_q-layer.6": 10302.3926, "encoder_q-layer.7": 11527.6865, "encoder_q-layer.8": 13402.5801, "encoder_q-layer.9": 12382.1113, "epoch": 0.53, "inbatch_neg_score": 3.0109, "inbatch_pos_score": 3.6094, "learning_rate": 2.078947368421053e-05, "loss": 3.4073, "norm_diff": 0.0581, "norm_loss": 0.0, "num_token_doc": 66.9453, "num_token_overlap": 11.6528, "num_token_query": 31.2364, "num_token_union": 65.1097, "num_word_context": 202.2225, "num_word_doc": 49.9238, "num_word_query": 23.193, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19727.2788, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0117, "query_norm": 2.1386, "queue_k_norm": 2.2008, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2364, "sent_len_1": 66.9453, "sent_len_max_0": 127.9813, "sent_len_max_1": 211.71, "stdk": 0.0492, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121000 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.4031, "doc_norm": 2.2026, "encoder_q-embeddings": 9965.8418, "encoder_q-layer.0": 6933.6963, "encoder_q-layer.1": 7330.8252, "encoder_q-layer.10": 11781.6436, "encoder_q-layer.11": 30928.9785, "encoder_q-layer.2": 8258.1602, "encoder_q-layer.3": 8530.6006, "encoder_q-layer.4": 9337.4668, "encoder_q-layer.5": 9681.2725, "encoder_q-layer.6": 10430.668, "encoder_q-layer.7": 11456.5098, "encoder_q-layer.8": 12714.8291, "encoder_q-layer.9": 11679.6875, "epoch": 0.53, "inbatch_neg_score": 3.0168, "inbatch_pos_score": 3.6328, "learning_rate": 2.0763157894736844e-05, "loss": 3.4031, "norm_diff": 0.0582, "norm_loss": 0.0, "num_token_doc": 66.8091, "num_token_overlap": 11.6938, "num_token_query": 31.4281, "num_token_union": 65.1322, "num_word_context": 202.2865, "num_word_doc": 49.8415, "num_word_query": 23.3535, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19539.0065, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0176, "query_norm": 2.1444, "queue_k_norm": 2.2025, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4281, "sent_len_1": 66.8091, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.395, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 121100 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.4092, "doc_norm": 2.2021, "encoder_q-embeddings": 10854.9014, "encoder_q-layer.0": 7574.311, "encoder_q-layer.1": 8213.9521, "encoder_q-layer.10": 12851.9619, "encoder_q-layer.11": 30462.2793, "encoder_q-layer.2": 9412.9609, "encoder_q-layer.3": 9946.5918, "encoder_q-layer.4": 11189.5293, "encoder_q-layer.5": 11223.875, "encoder_q-layer.6": 11883.6309, "encoder_q-layer.7": 13513.1318, "encoder_q-layer.8": 14172.0674, "encoder_q-layer.9": 12896.6621, "epoch": 0.53, "inbatch_neg_score": 3.0309, "inbatch_pos_score": 3.6367, "learning_rate": 2.073684210526316e-05, "loss": 3.4092, "norm_diff": 0.0452, "norm_loss": 0.0, "num_token_doc": 66.8892, "num_token_overlap": 11.6831, "num_token_query": 31.3847, "num_token_union": 65.1918, "num_word_context": 202.3613, "num_word_doc": 49.9212, "num_word_query": 23.3146, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20668.3587, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0312, "query_norm": 2.1569, "queue_k_norm": 2.2039, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3847, "sent_len_1": 66.8892, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.5462, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 121200 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4048, "doc_norm": 2.2028, "encoder_q-embeddings": 10266.9297, "encoder_q-layer.0": 7102.5566, "encoder_q-layer.1": 7682.7319, "encoder_q-layer.10": 14454.7207, "encoder_q-layer.11": 33931.0469, "encoder_q-layer.2": 8557.3574, "encoder_q-layer.3": 8954.5527, "encoder_q-layer.4": 9729.2568, "encoder_q-layer.5": 9997.3379, "encoder_q-layer.6": 11011.6338, "encoder_q-layer.7": 12334.915, "encoder_q-layer.8": 15030.0049, "encoder_q-layer.9": 13990.1357, "epoch": 0.53, "inbatch_neg_score": 3.0404, "inbatch_pos_score": 3.6641, "learning_rate": 2.0710526315789474e-05, "loss": 3.4048, "norm_diff": 0.0319, "norm_loss": 0.0, "num_token_doc": 66.8381, "num_token_overlap": 11.7104, "num_token_query": 31.4402, "num_token_union": 65.1952, "num_word_context": 202.569, "num_word_doc": 49.8751, "num_word_query": 23.3481, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21180.9047, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0391, "query_norm": 2.171, "queue_k_norm": 2.205, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4402, "sent_len_1": 66.8381, "sent_len_max_0": 127.9513, "sent_len_max_1": 207.6962, "stdk": 0.0493, "stdq": 0.0442, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 121300 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.42, "doc_norm": 2.2059, "encoder_q-embeddings": 9866.5049, "encoder_q-layer.0": 6713.0225, "encoder_q-layer.1": 7130.8574, "encoder_q-layer.10": 13605.0615, "encoder_q-layer.11": 33044.3242, "encoder_q-layer.2": 8153.5498, "encoder_q-layer.3": 8240.1611, "encoder_q-layer.4": 8742.2617, "encoder_q-layer.5": 8940.4453, "encoder_q-layer.6": 9974.5732, "encoder_q-layer.7": 11830.8623, "encoder_q-layer.8": 14166.1143, "encoder_q-layer.9": 12990.1631, "epoch": 0.53, "inbatch_neg_score": 3.0527, "inbatch_pos_score": 3.6504, "learning_rate": 2.068421052631579e-05, "loss": 3.42, "norm_diff": 0.0467, "norm_loss": 0.0, "num_token_doc": 66.8887, "num_token_overlap": 11.6607, "num_token_query": 31.3671, "num_token_union": 65.1798, "num_word_context": 202.3248, "num_word_doc": 49.8895, "num_word_query": 23.3025, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20448.274, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0527, "query_norm": 2.1592, "queue_k_norm": 2.2057, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3671, "sent_len_1": 66.8887, "sent_len_max_0": 127.97, "sent_len_max_1": 212.1788, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121400 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.41, "doc_norm": 2.2081, "encoder_q-embeddings": 9166.0068, "encoder_q-layer.0": 6355.3911, "encoder_q-layer.1": 6844.626, "encoder_q-layer.10": 13681.1982, "encoder_q-layer.11": 32278.8867, "encoder_q-layer.2": 7663.7559, "encoder_q-layer.3": 7873.4258, "encoder_q-layer.4": 8479.8477, "encoder_q-layer.5": 8876.1084, "encoder_q-layer.6": 9806.1152, "encoder_q-layer.7": 10963.3369, "encoder_q-layer.8": 14195.0264, "encoder_q-layer.9": 12894.2871, "epoch": 0.53, "inbatch_neg_score": 3.065, "inbatch_pos_score": 3.6797, "learning_rate": 2.0657894736842107e-05, "loss": 3.41, "norm_diff": 0.0457, "norm_loss": 0.0, "num_token_doc": 66.9235, "num_token_overlap": 11.6926, "num_token_query": 31.4517, "num_token_union": 65.2429, "num_word_context": 202.521, "num_word_doc": 49.9492, "num_word_query": 23.374, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19997.6221, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0645, "query_norm": 2.1624, "queue_k_norm": 2.207, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4517, "sent_len_1": 66.9235, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.0213, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121500 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3853, "doc_norm": 2.2097, "encoder_q-embeddings": 9788.0293, "encoder_q-layer.0": 6734.2266, "encoder_q-layer.1": 7243.9966, "encoder_q-layer.10": 14486.7637, "encoder_q-layer.11": 32130.9531, "encoder_q-layer.2": 8124.1572, "encoder_q-layer.3": 8262.334, "encoder_q-layer.4": 8744.8115, "encoder_q-layer.5": 9423.1777, "encoder_q-layer.6": 10387.7178, "encoder_q-layer.7": 12033.585, "encoder_q-layer.8": 14463.1221, "encoder_q-layer.9": 13190.7422, "epoch": 0.53, "inbatch_neg_score": 3.069, "inbatch_pos_score": 3.6797, "learning_rate": 2.0631578947368422e-05, "loss": 3.3853, "norm_diff": 0.0506, "norm_loss": 0.0, "num_token_doc": 66.6446, "num_token_overlap": 11.6791, "num_token_query": 31.3702, "num_token_union": 65.0337, "num_word_context": 202.2374, "num_word_doc": 49.7559, "num_word_query": 23.3062, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20420.0823, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0684, "query_norm": 2.1591, "queue_k_norm": 2.209, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3702, "sent_len_1": 66.6446, "sent_len_max_0": 127.9862, "sent_len_max_1": 206.7337, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121600 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.4118, "doc_norm": 2.2124, "encoder_q-embeddings": 9845.3955, "encoder_q-layer.0": 7021.0015, "encoder_q-layer.1": 7738.5869, "encoder_q-layer.10": 11924.1211, "encoder_q-layer.11": 29807.4297, "encoder_q-layer.2": 8587.3936, "encoder_q-layer.3": 8595.6094, "encoder_q-layer.4": 9320.5479, "encoder_q-layer.5": 9637.3369, "encoder_q-layer.6": 10229.9219, "encoder_q-layer.7": 11807.0107, "encoder_q-layer.8": 13201.1982, "encoder_q-layer.9": 11742.0586, "epoch": 0.53, "inbatch_neg_score": 3.0756, "inbatch_pos_score": 3.6914, "learning_rate": 2.0605263157894737e-05, "loss": 3.4118, "norm_diff": 0.0637, "norm_loss": 0.0, "num_token_doc": 66.6744, "num_token_overlap": 11.649, "num_token_query": 31.3468, "num_token_union": 65.0392, "num_word_context": 202.418, "num_word_doc": 49.7295, "num_word_query": 23.2593, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19331.2296, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0762, "query_norm": 2.1487, "queue_k_norm": 2.2113, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3468, "sent_len_1": 66.6744, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.6138, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 121700 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.41, "doc_norm": 2.2136, "encoder_q-embeddings": 10050.6035, "encoder_q-layer.0": 7199.8091, "encoder_q-layer.1": 7730.5146, "encoder_q-layer.10": 12795.9844, "encoder_q-layer.11": 30785.1914, "encoder_q-layer.2": 8632.9707, "encoder_q-layer.3": 9585.7051, "encoder_q-layer.4": 10161.7783, "encoder_q-layer.5": 10227.8574, "encoder_q-layer.6": 11777.418, "encoder_q-layer.7": 12638.584, "encoder_q-layer.8": 14742.2441, "encoder_q-layer.9": 13230.9316, "epoch": 0.53, "inbatch_neg_score": 3.079, "inbatch_pos_score": 3.6855, "learning_rate": 2.0578947368421052e-05, "loss": 3.41, "norm_diff": 0.07, "norm_loss": 0.0, "num_token_doc": 66.8174, "num_token_overlap": 11.6593, "num_token_query": 31.3591, "num_token_union": 65.1435, "num_word_context": 202.3981, "num_word_doc": 49.8592, "num_word_query": 23.302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20482.4709, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0781, "query_norm": 2.1436, "queue_k_norm": 2.2123, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3591, "sent_len_1": 66.8174, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.2713, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121800 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3982, "doc_norm": 2.2112, "encoder_q-embeddings": 9444.4297, "encoder_q-layer.0": 6692.8457, "encoder_q-layer.1": 7096.334, "encoder_q-layer.10": 12986.2021, "encoder_q-layer.11": 31576.9375, "encoder_q-layer.2": 7927.8022, "encoder_q-layer.3": 7819.3335, "encoder_q-layer.4": 8507.0664, "encoder_q-layer.5": 8578.9395, "encoder_q-layer.6": 9431.0654, "encoder_q-layer.7": 10704.7041, "encoder_q-layer.8": 13852.5, "encoder_q-layer.9": 11922.1582, "epoch": 0.53, "inbatch_neg_score": 3.0813, "inbatch_pos_score": 3.6953, "learning_rate": 2.055263157894737e-05, "loss": 3.3982, "norm_diff": 0.056, "norm_loss": 0.0, "num_token_doc": 66.7959, "num_token_overlap": 11.6855, "num_token_query": 31.3884, "num_token_union": 65.115, "num_word_context": 202.2859, "num_word_doc": 49.826, "num_word_query": 23.3178, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19372.9116, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.082, "query_norm": 2.1552, "queue_k_norm": 2.2126, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3884, "sent_len_1": 66.7959, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.7675, "stdk": 0.0492, "stdq": 0.0439, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 121900 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.4279, "doc_norm": 2.2127, "encoder_q-embeddings": 9890.0127, "encoder_q-layer.0": 6595.0039, "encoder_q-layer.1": 7225.9058, "encoder_q-layer.10": 13302.2021, "encoder_q-layer.11": 30902.9102, "encoder_q-layer.2": 8165.5537, "encoder_q-layer.3": 8460.9736, "encoder_q-layer.4": 8853.4814, "encoder_q-layer.5": 8854.0469, "encoder_q-layer.6": 9787.0586, "encoder_q-layer.7": 11092.3037, "encoder_q-layer.8": 13331.1777, "encoder_q-layer.9": 12582.3594, "epoch": 0.53, "inbatch_neg_score": 3.0817, "inbatch_pos_score": 3.6758, "learning_rate": 2.0526315789473685e-05, "loss": 3.4279, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.6488, "num_token_overlap": 11.6095, "num_token_query": 31.2512, "num_token_union": 65.014, "num_word_context": 202.2476, "num_word_doc": 49.7265, "num_word_query": 23.2091, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20018.3028, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.082, "query_norm": 2.1352, "queue_k_norm": 2.2138, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2512, "sent_len_1": 66.6488, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.4812, "stdk": 0.0492, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122000 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.427, "doc_norm": 2.2144, "encoder_q-embeddings": 9827.5508, "encoder_q-layer.0": 6629.1392, "encoder_q-layer.1": 7209.9985, "encoder_q-layer.10": 13235.9326, "encoder_q-layer.11": 31570.5547, "encoder_q-layer.2": 8065.2798, "encoder_q-layer.3": 8258.4463, "encoder_q-layer.4": 9034.541, "encoder_q-layer.5": 9116.709, "encoder_q-layer.6": 10507.6025, "encoder_q-layer.7": 12341.1816, "encoder_q-layer.8": 13496.5771, "encoder_q-layer.9": 12686.8936, "epoch": 0.53, "inbatch_neg_score": 3.0843, "inbatch_pos_score": 3.6836, "learning_rate": 2.05e-05, "loss": 3.427, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.6495, "num_token_overlap": 11.5979, "num_token_query": 31.2531, "num_token_union": 65.0318, "num_word_context": 202.2759, "num_word_doc": 49.7782, "num_word_query": 23.2101, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20261.5122, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.084, "query_norm": 2.1376, "queue_k_norm": 2.2156, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2531, "sent_len_1": 66.6495, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.0375, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122100 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.3954, "doc_norm": 2.2198, "encoder_q-embeddings": 9625.8496, "encoder_q-layer.0": 6752.2314, "encoder_q-layer.1": 7225.3335, "encoder_q-layer.10": 15419.7881, "encoder_q-layer.11": 33951.4492, "encoder_q-layer.2": 8154.6973, "encoder_q-layer.3": 8280.584, "encoder_q-layer.4": 9140.5928, "encoder_q-layer.5": 9415.875, "encoder_q-layer.6": 10780.2686, "encoder_q-layer.7": 11946.2285, "encoder_q-layer.8": 14715.3574, "encoder_q-layer.9": 13690.8057, "epoch": 0.53, "inbatch_neg_score": 3.0859, "inbatch_pos_score": 3.7031, "learning_rate": 2.047368421052632e-05, "loss": 3.3954, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.9555, "num_token_overlap": 11.723, "num_token_query": 31.4762, "num_token_union": 65.2338, "num_word_context": 202.6084, "num_word_doc": 49.9382, "num_word_query": 23.367, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20508.2666, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0859, "query_norm": 2.1449, "queue_k_norm": 2.2166, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4762, "sent_len_1": 66.9555, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.9187, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122200 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4058, "doc_norm": 2.2175, "encoder_q-embeddings": 9355.8018, "encoder_q-layer.0": 6431.9536, "encoder_q-layer.1": 6707.2915, "encoder_q-layer.10": 15219.4014, "encoder_q-layer.11": 34485.6562, "encoder_q-layer.2": 7559.4028, "encoder_q-layer.3": 7594.686, "encoder_q-layer.4": 8140.8618, "encoder_q-layer.5": 8407.708, "encoder_q-layer.6": 9621.0625, "encoder_q-layer.7": 11478.6162, "encoder_q-layer.8": 15086.1035, "encoder_q-layer.9": 14000.3232, "epoch": 0.53, "inbatch_neg_score": 3.0885, "inbatch_pos_score": 3.6934, "learning_rate": 2.044736842105263e-05, "loss": 3.4058, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.899, "num_token_overlap": 11.722, "num_token_query": 31.4493, "num_token_union": 65.1742, "num_word_context": 202.4061, "num_word_doc": 49.9085, "num_word_query": 23.351, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20141.0309, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0898, "query_norm": 2.1454, "queue_k_norm": 2.2203, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4493, "sent_len_1": 66.899, "sent_len_max_0": 127.9887, "sent_len_max_1": 211.485, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 122300 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4015, "doc_norm": 2.2213, "encoder_q-embeddings": 20209.0117, "encoder_q-layer.0": 13933.2822, "encoder_q-layer.1": 14820.3145, "encoder_q-layer.10": 27076.3438, "encoder_q-layer.11": 64414.7383, "encoder_q-layer.2": 16760.707, "encoder_q-layer.3": 17832.2246, "encoder_q-layer.4": 18998.9082, "encoder_q-layer.5": 20127.9727, "encoder_q-layer.6": 22581.502, "encoder_q-layer.7": 25579.6426, "encoder_q-layer.8": 29318.668, "encoder_q-layer.9": 28055.7852, "epoch": 0.53, "inbatch_neg_score": 3.0905, "inbatch_pos_score": 3.7129, "learning_rate": 2.042105263157895e-05, "loss": 3.4015, "norm_diff": 0.0658, "norm_loss": 0.0, "num_token_doc": 66.8277, "num_token_overlap": 11.6651, "num_token_query": 31.4362, "num_token_union": 65.2206, "num_word_context": 202.3361, "num_word_doc": 49.8559, "num_word_query": 23.36, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41708.2691, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 3.0898, "query_norm": 2.1555, "queue_k_norm": 2.2202, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4362, "sent_len_1": 66.8277, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.5025, "stdk": 0.0495, "stdq": 0.0439, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 122400 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.4222, "doc_norm": 2.2223, "encoder_q-embeddings": 9327.7812, "encoder_q-layer.0": 6485.2461, "encoder_q-layer.1": 6861.6572, "encoder_q-layer.10": 12804.7559, "encoder_q-layer.11": 30194.8672, "encoder_q-layer.2": 7586.8218, "encoder_q-layer.3": 7686.3164, "encoder_q-layer.4": 7915.2905, "encoder_q-layer.5": 8227.2207, "encoder_q-layer.6": 9514.6729, "encoder_q-layer.7": 11404.957, "encoder_q-layer.8": 13357.2598, "encoder_q-layer.9": 12205.7959, "epoch": 0.53, "inbatch_neg_score": 3.0921, "inbatch_pos_score": 3.7051, "learning_rate": 2.0394736842105264e-05, "loss": 3.4222, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.7655, "num_token_overlap": 11.6332, "num_token_query": 31.2402, "num_token_union": 65.0715, "num_word_context": 202.0398, "num_word_doc": 49.7785, "num_word_query": 23.2006, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18903.0232, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0938, "query_norm": 2.1452, "queue_k_norm": 2.22, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2402, "sent_len_1": 66.7655, "sent_len_max_0": 128.0, "sent_len_max_1": 210.0025, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122500 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.3937, "doc_norm": 2.2224, "encoder_q-embeddings": 10039.3789, "encoder_q-layer.0": 6790.8564, "encoder_q-layer.1": 7122.5303, "encoder_q-layer.10": 12766.0859, "encoder_q-layer.11": 31539.2656, "encoder_q-layer.2": 7852.6543, "encoder_q-layer.3": 8196.4082, "encoder_q-layer.4": 8686.8975, "encoder_q-layer.5": 9234.0234, "encoder_q-layer.6": 10327.0928, "encoder_q-layer.7": 11706.1846, "encoder_q-layer.8": 13845.4316, "encoder_q-layer.9": 12741.2129, "epoch": 0.53, "inbatch_neg_score": 3.0991, "inbatch_pos_score": 3.6953, "learning_rate": 2.0368421052631582e-05, "loss": 3.3937, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.9048, "num_token_overlap": 11.7005, "num_token_query": 31.4271, "num_token_union": 65.2101, "num_word_context": 202.4124, "num_word_doc": 49.9233, "num_word_query": 23.3431, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20274.1024, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.0977, "query_norm": 2.1437, "queue_k_norm": 2.2219, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4271, "sent_len_1": 66.9048, "sent_len_max_0": 127.965, "sent_len_max_1": 206.6562, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122600 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.4149, "doc_norm": 2.2181, "encoder_q-embeddings": 10740.4092, "encoder_q-layer.0": 7462.8696, "encoder_q-layer.1": 8219.3994, "encoder_q-layer.10": 13295.1064, "encoder_q-layer.11": 32210.2402, "encoder_q-layer.2": 8994.0908, "encoder_q-layer.3": 9282.8096, "encoder_q-layer.4": 10243.2129, "encoder_q-layer.5": 10629.2949, "encoder_q-layer.6": 11170.7686, "encoder_q-layer.7": 12015.417, "encoder_q-layer.8": 13530.8926, "encoder_q-layer.9": 12874.8467, "epoch": 0.53, "inbatch_neg_score": 3.1043, "inbatch_pos_score": 3.7031, "learning_rate": 2.0342105263157897e-05, "loss": 3.4149, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.7863, "num_token_overlap": 11.6723, "num_token_query": 31.3716, "num_token_union": 65.1285, "num_word_context": 202.1231, "num_word_doc": 49.8223, "num_word_query": 23.3089, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20913.6853, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1035, "query_norm": 2.1429, "queue_k_norm": 2.2222, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3716, "sent_len_1": 66.7863, "sent_len_max_0": 127.9475, "sent_len_max_1": 209.6163, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122700 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.4019, "doc_norm": 2.2251, "encoder_q-embeddings": 9828.251, "encoder_q-layer.0": 6863.7754, "encoder_q-layer.1": 7291.0835, "encoder_q-layer.10": 13486.1299, "encoder_q-layer.11": 33007.6523, "encoder_q-layer.2": 8037.3418, "encoder_q-layer.3": 8265.6152, "encoder_q-layer.4": 8821.6348, "encoder_q-layer.5": 9362.9824, "encoder_q-layer.6": 10091.1191, "encoder_q-layer.7": 12196.1455, "encoder_q-layer.8": 14490.3789, "encoder_q-layer.9": 13261.1689, "epoch": 0.53, "inbatch_neg_score": 3.1071, "inbatch_pos_score": 3.7168, "learning_rate": 2.0315789473684212e-05, "loss": 3.4019, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.7443, "num_token_overlap": 11.6822, "num_token_query": 31.3682, "num_token_union": 65.0917, "num_word_context": 202.2248, "num_word_doc": 49.8008, "num_word_query": 23.2953, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20544.49, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1074, "query_norm": 2.1528, "queue_k_norm": 2.2238, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3682, "sent_len_1": 66.7443, "sent_len_max_0": 127.93, "sent_len_max_1": 209.2925, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122800 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4098, "doc_norm": 2.2248, "encoder_q-embeddings": 10513.3691, "encoder_q-layer.0": 7577.2695, "encoder_q-layer.1": 7831.8481, "encoder_q-layer.10": 12325.1699, "encoder_q-layer.11": 30878.9023, "encoder_q-layer.2": 8517.2822, "encoder_q-layer.3": 8836.1807, "encoder_q-layer.4": 9161.1904, "encoder_q-layer.5": 9259.4277, "encoder_q-layer.6": 10721.4795, "encoder_q-layer.7": 10924.8672, "encoder_q-layer.8": 12928.4824, "encoder_q-layer.9": 12000.2412, "epoch": 0.53, "inbatch_neg_score": 3.1072, "inbatch_pos_score": 3.7148, "learning_rate": 2.0289473684210527e-05, "loss": 3.4098, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.5629, "num_token_overlap": 11.6896, "num_token_query": 31.3608, "num_token_union": 65.0011, "num_word_context": 202.4, "num_word_doc": 49.6889, "num_word_query": 23.3008, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19546.602, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1074, "query_norm": 2.148, "queue_k_norm": 2.224, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3608, "sent_len_1": 66.5629, "sent_len_max_0": 127.9788, "sent_len_max_1": 207.4263, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 122900 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4193, "doc_norm": 2.2266, "encoder_q-embeddings": 9637.1113, "encoder_q-layer.0": 7104.0659, "encoder_q-layer.1": 7241.0742, "encoder_q-layer.10": 12175.4365, "encoder_q-layer.11": 29846.0, "encoder_q-layer.2": 7818.2334, "encoder_q-layer.3": 7984.4653, "encoder_q-layer.4": 8698.5518, "encoder_q-layer.5": 8682.373, "encoder_q-layer.6": 9249.9072, "encoder_q-layer.7": 10623.8467, "encoder_q-layer.8": 13169.9893, "encoder_q-layer.9": 12234.0625, "epoch": 0.53, "inbatch_neg_score": 3.1121, "inbatch_pos_score": 3.7246, "learning_rate": 2.0263157894736842e-05, "loss": 3.4193, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.6296, "num_token_overlap": 11.6582, "num_token_query": 31.3391, "num_token_union": 65.0297, "num_word_context": 202.3283, "num_word_doc": 49.7471, "num_word_query": 23.2678, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19125.0008, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1113, "query_norm": 2.151, "queue_k_norm": 2.2252, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3391, "sent_len_1": 66.6296, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.925, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123000 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.4018, "doc_norm": 2.2225, "encoder_q-embeddings": 9171.0947, "encoder_q-layer.0": 6565.439, "encoder_q-layer.1": 7049.0479, "encoder_q-layer.10": 12490.0273, "encoder_q-layer.11": 29622.0078, "encoder_q-layer.2": 7557.1528, "encoder_q-layer.3": 8104.2319, "encoder_q-layer.4": 8205.8115, "encoder_q-layer.5": 8466.1025, "encoder_q-layer.6": 9490.4717, "encoder_q-layer.7": 10756.8428, "encoder_q-layer.8": 13522.335, "encoder_q-layer.9": 12362.2168, "epoch": 0.53, "inbatch_neg_score": 3.1201, "inbatch_pos_score": 3.7148, "learning_rate": 2.023684210526316e-05, "loss": 3.4018, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.8484, "num_token_overlap": 11.6552, "num_token_query": 31.3508, "num_token_union": 65.1428, "num_word_context": 202.2962, "num_word_doc": 49.8871, "num_word_query": 23.2884, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18899.3101, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1191, "query_norm": 2.1542, "queue_k_norm": 2.2251, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3508, "sent_len_1": 66.8484, "sent_len_max_0": 127.99, "sent_len_max_1": 210.2912, "stdk": 0.0491, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123100 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4085, "doc_norm": 2.2275, "encoder_q-embeddings": 9150.3838, "encoder_q-layer.0": 6274.9155, "encoder_q-layer.1": 6709.0864, "encoder_q-layer.10": 12633.2842, "encoder_q-layer.11": 31068.8184, "encoder_q-layer.2": 7493.8589, "encoder_q-layer.3": 7629.3398, "encoder_q-layer.4": 8240.4189, "encoder_q-layer.5": 9097.8008, "encoder_q-layer.6": 9792.6924, "encoder_q-layer.7": 10999.2031, "encoder_q-layer.8": 13330.9072, "encoder_q-layer.9": 12088.9326, "epoch": 0.53, "inbatch_neg_score": 3.1198, "inbatch_pos_score": 3.7227, "learning_rate": 2.0210526315789475e-05, "loss": 3.4085, "norm_diff": 0.0853, "norm_loss": 0.0, "num_token_doc": 66.6538, "num_token_overlap": 11.6688, "num_token_query": 31.3619, "num_token_union": 65.0255, "num_word_context": 202.0768, "num_word_doc": 49.7228, "num_word_query": 23.294, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19177.5312, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1211, "query_norm": 2.1422, "queue_k_norm": 2.2267, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3619, "sent_len_1": 66.6538, "sent_len_max_0": 128.0, "sent_len_max_1": 208.7038, "stdk": 0.0494, "stdq": 0.0425, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123200 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.4075, "doc_norm": 2.2252, "encoder_q-embeddings": 9921.3809, "encoder_q-layer.0": 6913.5332, "encoder_q-layer.1": 7329.8984, "encoder_q-layer.10": 12440.335, "encoder_q-layer.11": 29746.3145, "encoder_q-layer.2": 8209.2568, "encoder_q-layer.3": 8499.4365, "encoder_q-layer.4": 8863.6523, "encoder_q-layer.5": 9048.0732, "encoder_q-layer.6": 10163.6582, "encoder_q-layer.7": 11497.3301, "encoder_q-layer.8": 14509.6367, "encoder_q-layer.9": 12346.9092, "epoch": 0.54, "inbatch_neg_score": 3.127, "inbatch_pos_score": 3.7305, "learning_rate": 2.018421052631579e-05, "loss": 3.4075, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.845, "num_token_overlap": 11.6701, "num_token_query": 31.3334, "num_token_union": 65.1225, "num_word_context": 202.2414, "num_word_doc": 49.8729, "num_word_query": 23.2658, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19449.2904, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.127, "query_norm": 2.1533, "queue_k_norm": 2.228, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3334, "sent_len_1": 66.845, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.6325, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123300 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3974, "doc_norm": 2.2293, "encoder_q-embeddings": 10241.4375, "encoder_q-layer.0": 6968.5615, "encoder_q-layer.1": 7369.9082, "encoder_q-layer.10": 13623.2607, "encoder_q-layer.11": 31330.2578, "encoder_q-layer.2": 8186.3999, "encoder_q-layer.3": 8440.8379, "encoder_q-layer.4": 8843.0635, "encoder_q-layer.5": 9278.6553, "encoder_q-layer.6": 10287.7275, "encoder_q-layer.7": 12597.8057, "encoder_q-layer.8": 14250.1357, "encoder_q-layer.9": 13157.5986, "epoch": 0.54, "inbatch_neg_score": 3.1303, "inbatch_pos_score": 3.7539, "learning_rate": 2.0157894736842105e-05, "loss": 3.3974, "norm_diff": 0.0628, "norm_loss": 0.0, "num_token_doc": 66.8734, "num_token_overlap": 11.7047, "num_token_query": 31.4763, "num_token_union": 65.2122, "num_word_context": 201.9999, "num_word_doc": 49.9069, "num_word_query": 23.3857, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20267.9215, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1309, "query_norm": 2.1665, "queue_k_norm": 2.2283, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4763, "sent_len_1": 66.8734, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.845, "stdk": 0.0494, "stdq": 0.044, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123400 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.4052, "doc_norm": 2.2313, "encoder_q-embeddings": 10179.7002, "encoder_q-layer.0": 7054.7061, "encoder_q-layer.1": 7474.8325, "encoder_q-layer.10": 13529.7861, "encoder_q-layer.11": 31212.3965, "encoder_q-layer.2": 8089.9819, "encoder_q-layer.3": 8336.0518, "encoder_q-layer.4": 8719.3291, "encoder_q-layer.5": 8931.0801, "encoder_q-layer.6": 9880.5107, "encoder_q-layer.7": 11085.3887, "encoder_q-layer.8": 12998.3623, "encoder_q-layer.9": 12533.3242, "epoch": 0.54, "inbatch_neg_score": 3.1322, "inbatch_pos_score": 3.7344, "learning_rate": 2.013157894736842e-05, "loss": 3.4052, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.874, "num_token_overlap": 11.6816, "num_token_query": 31.3082, "num_token_union": 65.1316, "num_word_context": 202.4665, "num_word_doc": 49.9147, "num_word_query": 23.2432, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19971.9455, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1328, "query_norm": 2.1511, "queue_k_norm": 2.2298, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3082, "sent_len_1": 66.874, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.855, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123500 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.4165, "doc_norm": 2.2298, "encoder_q-embeddings": 10469.5605, "encoder_q-layer.0": 7328.3936, "encoder_q-layer.1": 7772.6152, "encoder_q-layer.10": 13079.1689, "encoder_q-layer.11": 29977.5938, "encoder_q-layer.2": 8471.7109, "encoder_q-layer.3": 8558.6484, "encoder_q-layer.4": 9336.6865, "encoder_q-layer.5": 9643.043, "encoder_q-layer.6": 10452.5293, "encoder_q-layer.7": 11187.2168, "encoder_q-layer.8": 12876.7393, "encoder_q-layer.9": 11976.0381, "epoch": 0.54, "inbatch_neg_score": 3.1384, "inbatch_pos_score": 3.7383, "learning_rate": 2.010526315789474e-05, "loss": 3.4165, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.7638, "num_token_overlap": 11.6327, "num_token_query": 31.3221, "num_token_union": 65.0878, "num_word_context": 202.2121, "num_word_doc": 49.7857, "num_word_query": 23.2602, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19542.6965, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1387, "query_norm": 2.1591, "queue_k_norm": 2.2308, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3221, "sent_len_1": 66.7638, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.3587, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123600 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.3971, "doc_norm": 2.2295, "encoder_q-embeddings": 10527.9854, "encoder_q-layer.0": 7030.5059, "encoder_q-layer.1": 7430.5752, "encoder_q-layer.10": 13304.6738, "encoder_q-layer.11": 31070.4297, "encoder_q-layer.2": 8179.3408, "encoder_q-layer.3": 8387.2686, "encoder_q-layer.4": 8801.9404, "encoder_q-layer.5": 9149.085, "encoder_q-layer.6": 10225.3799, "encoder_q-layer.7": 11811.2129, "encoder_q-layer.8": 13796.6611, "encoder_q-layer.9": 13059.6367, "epoch": 0.54, "inbatch_neg_score": 3.1403, "inbatch_pos_score": 3.7656, "learning_rate": 2.0078947368421053e-05, "loss": 3.3971, "norm_diff": 0.0616, "norm_loss": 0.0, "num_token_doc": 66.6786, "num_token_overlap": 11.6699, "num_token_query": 31.4136, "num_token_union": 65.0701, "num_word_context": 202.3601, "num_word_doc": 49.7266, "num_word_query": 23.3272, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20033.1585, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1406, "query_norm": 2.168, "queue_k_norm": 2.2308, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4136, "sent_len_1": 66.6786, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.52, "stdk": 0.0493, "stdq": 0.0439, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123700 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.4042, "doc_norm": 2.2317, "encoder_q-embeddings": 9653.7324, "encoder_q-layer.0": 6759.562, "encoder_q-layer.1": 7188.3228, "encoder_q-layer.10": 13320.1807, "encoder_q-layer.11": 30336.3008, "encoder_q-layer.2": 8030.002, "encoder_q-layer.3": 8430.3145, "encoder_q-layer.4": 8949.957, "encoder_q-layer.5": 9434.793, "encoder_q-layer.6": 10413.5586, "encoder_q-layer.7": 11602.3408, "encoder_q-layer.8": 14023.2812, "encoder_q-layer.9": 12367.5322, "epoch": 0.54, "inbatch_neg_score": 3.144, "inbatch_pos_score": 3.752, "learning_rate": 2.005263157894737e-05, "loss": 3.4042, "norm_diff": 0.079, "norm_loss": 0.0, "num_token_doc": 66.9003, "num_token_overlap": 11.6982, "num_token_query": 31.3632, "num_token_union": 65.0911, "num_word_context": 202.2351, "num_word_doc": 49.8934, "num_word_query": 23.3055, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19762.7436, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.1445, "query_norm": 2.1527, "queue_k_norm": 2.2326, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3632, "sent_len_1": 66.9003, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.9512, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123800 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.4173, "doc_norm": 2.2341, "encoder_q-embeddings": 4939.9102, "encoder_q-layer.0": 3563.783, "encoder_q-layer.1": 3785.9644, "encoder_q-layer.10": 6384.0454, "encoder_q-layer.11": 15162.6074, "encoder_q-layer.2": 4168.3716, "encoder_q-layer.3": 4165.709, "encoder_q-layer.4": 4488.0791, "encoder_q-layer.5": 4747.1689, "encoder_q-layer.6": 5042.3701, "encoder_q-layer.7": 5633.5483, "encoder_q-layer.8": 6618.4155, "encoder_q-layer.9": 6134.0859, "epoch": 0.54, "inbatch_neg_score": 3.1451, "inbatch_pos_score": 3.7441, "learning_rate": 2.0026315789473683e-05, "loss": 3.4173, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.669, "num_token_overlap": 11.6556, "num_token_query": 31.3084, "num_token_union": 65.0367, "num_word_context": 202.145, "num_word_doc": 49.75, "num_word_query": 23.2374, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9820.2257, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1445, "query_norm": 2.1517, "queue_k_norm": 2.2325, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3084, "sent_len_1": 66.669, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.4038, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 123900 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.4145, "doc_norm": 2.2346, "encoder_q-embeddings": 4845.2803, "encoder_q-layer.0": 3325.7852, "encoder_q-layer.1": 3511.9312, "encoder_q-layer.10": 7145.9082, "encoder_q-layer.11": 15240.2422, "encoder_q-layer.2": 3955.7305, "encoder_q-layer.3": 4007.7839, "encoder_q-layer.4": 4280.4404, "encoder_q-layer.5": 4542.479, "encoder_q-layer.6": 4984.5898, "encoder_q-layer.7": 5470.9746, "encoder_q-layer.8": 6623.062, "encoder_q-layer.9": 6436.9326, "epoch": 0.54, "inbatch_neg_score": 3.1504, "inbatch_pos_score": 3.7656, "learning_rate": 2e-05, "loss": 3.4145, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.5459, "num_token_overlap": 11.6167, "num_token_query": 31.2517, "num_token_union": 64.9474, "num_word_context": 202.001, "num_word_doc": 49.6662, "num_word_query": 23.2174, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9682.1076, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1504, "query_norm": 2.164, "queue_k_norm": 2.2337, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2517, "sent_len_1": 66.5459, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.8638, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124000 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.401, "doc_norm": 2.2336, "encoder_q-embeddings": 6043.54, "encoder_q-layer.0": 4105.3071, "encoder_q-layer.1": 4628.0103, "encoder_q-layer.10": 6558.8667, "encoder_q-layer.11": 15908.4033, "encoder_q-layer.2": 5309.7983, "encoder_q-layer.3": 6073.7676, "encoder_q-layer.4": 6049.9824, "encoder_q-layer.5": 6328.8608, "encoder_q-layer.6": 7184.3931, "encoder_q-layer.7": 6966.1309, "encoder_q-layer.8": 7768.3975, "encoder_q-layer.9": 6461.1426, "epoch": 0.54, "inbatch_neg_score": 3.1504, "inbatch_pos_score": 3.7422, "learning_rate": 1.9973684210526317e-05, "loss": 3.401, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.8328, "num_token_overlap": 11.6988, "num_token_query": 31.3816, "num_token_union": 65.1225, "num_word_context": 202.3403, "num_word_doc": 49.88, "num_word_query": 23.3141, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11143.0517, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1504, "query_norm": 2.1518, "queue_k_norm": 2.2343, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3816, "sent_len_1": 66.8328, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.08, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124100 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.3994, "doc_norm": 2.2393, "encoder_q-embeddings": 4849.3164, "encoder_q-layer.0": 3469.3306, "encoder_q-layer.1": 3703.9446, "encoder_q-layer.10": 6685.7065, "encoder_q-layer.11": 16052.4004, "encoder_q-layer.2": 4108.6465, "encoder_q-layer.3": 4125.6182, "encoder_q-layer.4": 4483.2363, "encoder_q-layer.5": 4662.8887, "encoder_q-layer.6": 5041.2144, "encoder_q-layer.7": 5797.8433, "encoder_q-layer.8": 6979.6812, "encoder_q-layer.9": 6596.2754, "epoch": 0.54, "inbatch_neg_score": 3.1564, "inbatch_pos_score": 3.752, "learning_rate": 1.994736842105263e-05, "loss": 3.3994, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.9704, "num_token_overlap": 11.7083, "num_token_query": 31.3957, "num_token_union": 65.2284, "num_word_context": 202.4249, "num_word_doc": 49.9702, "num_word_query": 23.3215, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10124.5183, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1562, "query_norm": 2.1592, "queue_k_norm": 2.236, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3957, "sent_len_1": 66.9704, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.9275, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124200 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.4129, "doc_norm": 2.2352, "encoder_q-embeddings": 4832.4204, "encoder_q-layer.0": 3385.9497, "encoder_q-layer.1": 3501.6038, "encoder_q-layer.10": 6119.2319, "encoder_q-layer.11": 14799.6729, "encoder_q-layer.2": 3965.3281, "encoder_q-layer.3": 4080.2283, "encoder_q-layer.4": 4330.9648, "encoder_q-layer.5": 4435.6362, "encoder_q-layer.6": 5042.168, "encoder_q-layer.7": 5747.1104, "encoder_q-layer.8": 6591.6416, "encoder_q-layer.9": 6054.2593, "epoch": 0.54, "inbatch_neg_score": 3.1614, "inbatch_pos_score": 3.7734, "learning_rate": 1.992105263157895e-05, "loss": 3.4129, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.7509, "num_token_overlap": 11.637, "num_token_query": 31.3358, "num_token_union": 65.0837, "num_word_context": 202.3194, "num_word_doc": 49.8013, "num_word_query": 23.2688, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9615.271, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1602, "query_norm": 2.1646, "queue_k_norm": 2.2361, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3358, "sent_len_1": 66.7509, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6863, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124300 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3948, "doc_norm": 2.2367, "encoder_q-embeddings": 4956.9028, "encoder_q-layer.0": 3395.1829, "encoder_q-layer.1": 3656.6565, "encoder_q-layer.10": 6912.1553, "encoder_q-layer.11": 15810.502, "encoder_q-layer.2": 4027.4324, "encoder_q-layer.3": 4195.4688, "encoder_q-layer.4": 4432.6089, "encoder_q-layer.5": 4712.7393, "encoder_q-layer.6": 5138.6572, "encoder_q-layer.7": 6086.3652, "encoder_q-layer.8": 7285.874, "encoder_q-layer.9": 6309.876, "epoch": 0.54, "inbatch_neg_score": 3.1671, "inbatch_pos_score": 3.7773, "learning_rate": 1.9894736842105265e-05, "loss": 3.3948, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.7165, "num_token_overlap": 11.6933, "num_token_query": 31.4719, "num_token_union": 65.0966, "num_word_context": 201.8125, "num_word_doc": 49.7737, "num_word_query": 23.3637, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9973.8718, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.166, "query_norm": 2.1686, "queue_k_norm": 2.2355, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4719, "sent_len_1": 66.7165, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.8175, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 124400 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3991, "doc_norm": 2.2389, "encoder_q-embeddings": 5117.7451, "encoder_q-layer.0": 3358.2144, "encoder_q-layer.1": 3551.9111, "encoder_q-layer.10": 6567.6528, "encoder_q-layer.11": 15494.1807, "encoder_q-layer.2": 3938.0911, "encoder_q-layer.3": 4078.9512, "encoder_q-layer.4": 4356.3418, "encoder_q-layer.5": 4458.6582, "encoder_q-layer.6": 5144.8945, "encoder_q-layer.7": 5793.1104, "encoder_q-layer.8": 6951.3105, "encoder_q-layer.9": 6239.8696, "epoch": 0.54, "inbatch_neg_score": 3.1712, "inbatch_pos_score": 3.7891, "learning_rate": 1.986842105263158e-05, "loss": 3.3991, "norm_diff": 0.0632, "norm_loss": 0.0, "num_token_doc": 66.8271, "num_token_overlap": 11.7029, "num_token_query": 31.4654, "num_token_union": 65.1445, "num_word_context": 202.3371, "num_word_doc": 49.8622, "num_word_query": 23.376, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9907.4989, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1719, "query_norm": 2.1757, "queue_k_norm": 2.2367, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4654, "sent_len_1": 66.8271, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.7587, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 124500 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.4026, "doc_norm": 2.2397, "encoder_q-embeddings": 4944.9854, "encoder_q-layer.0": 3570.7202, "encoder_q-layer.1": 3735.2505, "encoder_q-layer.10": 6615.3608, "encoder_q-layer.11": 15631.6514, "encoder_q-layer.2": 4305.5903, "encoder_q-layer.3": 4373.0566, "encoder_q-layer.4": 4651.9116, "encoder_q-layer.5": 4850.915, "encoder_q-layer.6": 5420.2803, "encoder_q-layer.7": 5885.4858, "encoder_q-layer.8": 6896.4473, "encoder_q-layer.9": 6160.5439, "epoch": 0.54, "inbatch_neg_score": 3.1802, "inbatch_pos_score": 3.7812, "learning_rate": 1.9842105263157895e-05, "loss": 3.4026, "norm_diff": 0.061, "norm_loss": 0.0, "num_token_doc": 66.7415, "num_token_overlap": 11.6878, "num_token_query": 31.4568, "num_token_union": 65.1449, "num_word_context": 202.4363, "num_word_doc": 49.8199, "num_word_query": 23.3781, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9977.604, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1797, "query_norm": 2.1787, "queue_k_norm": 2.2395, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4568, "sent_len_1": 66.7415, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.0275, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124600 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.401, "doc_norm": 2.2378, "encoder_q-embeddings": 4960.1274, "encoder_q-layer.0": 3338.8384, "encoder_q-layer.1": 3606.1924, "encoder_q-layer.10": 6978.3721, "encoder_q-layer.11": 16196.4912, "encoder_q-layer.2": 4110.4912, "encoder_q-layer.3": 4340.686, "encoder_q-layer.4": 4709.2588, "encoder_q-layer.5": 4783.6099, "encoder_q-layer.6": 5550.5181, "encoder_q-layer.7": 6405.7925, "encoder_q-layer.8": 8209.1357, "encoder_q-layer.9": 7415.9204, "epoch": 0.54, "inbatch_neg_score": 3.1874, "inbatch_pos_score": 3.7852, "learning_rate": 1.9815789473684213e-05, "loss": 3.401, "norm_diff": 0.0624, "norm_loss": 0.0, "num_token_doc": 66.7177, "num_token_overlap": 11.6629, "num_token_query": 31.3145, "num_token_union": 65.0486, "num_word_context": 202.0278, "num_word_doc": 49.7867, "num_word_query": 23.2462, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10629.9027, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1875, "query_norm": 2.1754, "queue_k_norm": 2.2386, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3145, "sent_len_1": 66.7177, "sent_len_max_0": 127.97, "sent_len_max_1": 209.8663, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 124700 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4037, "doc_norm": 2.2383, "encoder_q-embeddings": 4721.5713, "encoder_q-layer.0": 3270.5198, "encoder_q-layer.1": 3528.5427, "encoder_q-layer.10": 6510.8574, "encoder_q-layer.11": 15296.1465, "encoder_q-layer.2": 4030.981, "encoder_q-layer.3": 4159.6021, "encoder_q-layer.4": 4392.4785, "encoder_q-layer.5": 4457.0146, "encoder_q-layer.6": 4988.7217, "encoder_q-layer.7": 5462.0693, "encoder_q-layer.8": 6460.7593, "encoder_q-layer.9": 6166.4243, "epoch": 0.54, "inbatch_neg_score": 3.1978, "inbatch_pos_score": 3.8125, "learning_rate": 1.9789473684210528e-05, "loss": 3.4037, "norm_diff": 0.055, "norm_loss": 0.0, "num_token_doc": 66.7447, "num_token_overlap": 11.6749, "num_token_query": 31.3352, "num_token_union": 65.0857, "num_word_context": 202.2345, "num_word_doc": 49.7925, "num_word_query": 23.2628, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9769.5377, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.1973, "query_norm": 2.1833, "queue_k_norm": 2.2411, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3352, "sent_len_1": 66.7447, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.8975, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124800 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4092, "doc_norm": 2.2421, "encoder_q-embeddings": 5972.7114, "encoder_q-layer.0": 4132.7314, "encoder_q-layer.1": 4447.7314, "encoder_q-layer.10": 6646.0698, "encoder_q-layer.11": 15350.6641, "encoder_q-layer.2": 5028.1113, "encoder_q-layer.3": 5191.5859, "encoder_q-layer.4": 5503.8994, "encoder_q-layer.5": 6000.5449, "encoder_q-layer.6": 6421.9282, "encoder_q-layer.7": 6782.1592, "encoder_q-layer.8": 7191.1235, "encoder_q-layer.9": 6244.0972, "epoch": 0.54, "inbatch_neg_score": 3.2061, "inbatch_pos_score": 3.8145, "learning_rate": 1.9763157894736843e-05, "loss": 3.4092, "norm_diff": 0.0662, "norm_loss": 0.0, "num_token_doc": 66.8687, "num_token_overlap": 11.6792, "num_token_query": 31.2908, "num_token_union": 65.1346, "num_word_context": 202.3818, "num_word_doc": 49.9127, "num_word_query": 23.2315, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10535.6561, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.207, "query_norm": 2.1759, "queue_k_norm": 2.2423, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2908, "sent_len_1": 66.8687, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.825, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 124900 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.4192, "doc_norm": 2.2451, "encoder_q-embeddings": 5074.603, "encoder_q-layer.0": 3656.2854, "encoder_q-layer.1": 3876.2288, "encoder_q-layer.10": 6625.752, "encoder_q-layer.11": 15739.5879, "encoder_q-layer.2": 4321.2363, "encoder_q-layer.3": 4498.6567, "encoder_q-layer.4": 4741.9585, "encoder_q-layer.5": 5040.4937, "encoder_q-layer.6": 5478.8755, "encoder_q-layer.7": 5965.0703, "encoder_q-layer.8": 7098.8916, "encoder_q-layer.9": 6288.5596, "epoch": 0.54, "inbatch_neg_score": 3.2192, "inbatch_pos_score": 3.8359, "learning_rate": 1.9736842105263158e-05, "loss": 3.4192, "norm_diff": 0.05, "norm_loss": 0.0, "num_token_doc": 66.6087, "num_token_overlap": 11.6291, "num_token_query": 31.2865, "num_token_union": 65.0212, "num_word_context": 201.9299, "num_word_doc": 49.6852, "num_word_query": 23.2429, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10173.6344, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.2207, "query_norm": 2.195, "queue_k_norm": 2.2423, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2865, "sent_len_1": 66.6087, "sent_len_max_0": 127.9762, "sent_len_max_1": 210.9512, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 125000 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.4097, "doc_norm": 2.2415, "encoder_q-embeddings": 4709.4883, "encoder_q-layer.0": 3332.4832, "encoder_q-layer.1": 3491.5144, "encoder_q-layer.10": 6655.6567, "encoder_q-layer.11": 15619.1738, "encoder_q-layer.2": 3849.7661, "encoder_q-layer.3": 3939.1396, "encoder_q-layer.4": 4201.1016, "encoder_q-layer.5": 4354.0283, "encoder_q-layer.6": 4645.0938, "encoder_q-layer.7": 5331.6626, "encoder_q-layer.8": 6715.2383, "encoder_q-layer.9": 6380.1157, "epoch": 0.54, "inbatch_neg_score": 3.2271, "inbatch_pos_score": 3.8398, "learning_rate": 1.9710526315789473e-05, "loss": 3.4097, "norm_diff": 0.048, "norm_loss": 0.0, "num_token_doc": 66.6403, "num_token_overlap": 11.6754, "num_token_query": 31.4293, "num_token_union": 65.0558, "num_word_context": 202.3712, "num_word_doc": 49.727, "num_word_query": 23.3362, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9713.0541, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.2266, "query_norm": 2.1935, "queue_k_norm": 2.2438, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4293, "sent_len_1": 66.6403, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.3625, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 125100 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.4144, "doc_norm": 2.2458, "encoder_q-embeddings": 2498.6853, "encoder_q-layer.0": 1712.5969, "encoder_q-layer.1": 1867.8823, "encoder_q-layer.10": 3255.4819, "encoder_q-layer.11": 7741.2026, "encoder_q-layer.2": 2075.6135, "encoder_q-layer.3": 2141.7473, "encoder_q-layer.4": 2258.2927, "encoder_q-layer.5": 2269.8618, "encoder_q-layer.6": 2691.6433, "encoder_q-layer.7": 2828.3381, "encoder_q-layer.8": 3412.4326, "encoder_q-layer.9": 3166.457, "epoch": 0.54, "inbatch_neg_score": 3.2354, "inbatch_pos_score": 3.8438, "learning_rate": 1.968421052631579e-05, "loss": 3.4144, "norm_diff": 0.0507, "norm_loss": 0.0, "num_token_doc": 66.8352, "num_token_overlap": 11.693, "num_token_query": 31.352, "num_token_union": 65.11, "num_word_context": 202.2239, "num_word_doc": 49.8392, "num_word_query": 23.2631, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4991.9479, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2363, "query_norm": 2.1952, "queue_k_norm": 2.2469, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.352, "sent_len_1": 66.8352, "sent_len_max_0": 127.99, "sent_len_max_1": 210.515, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125200 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.4054, "doc_norm": 2.2457, "encoder_q-embeddings": 23638.2773, "encoder_q-layer.0": 15210.9873, "encoder_q-layer.1": 17158.6621, "encoder_q-layer.10": 3431.2603, "encoder_q-layer.11": 8022.7856, "encoder_q-layer.2": 20635.1855, "encoder_q-layer.3": 22659.6777, "encoder_q-layer.4": 23795.1191, "encoder_q-layer.5": 24913.5547, "encoder_q-layer.6": 23772.6836, "encoder_q-layer.7": 22412.709, "encoder_q-layer.8": 15274.7695, "encoder_q-layer.9": 4931.6479, "epoch": 0.54, "inbatch_neg_score": 3.2463, "inbatch_pos_score": 3.8457, "learning_rate": 1.9657894736842106e-05, "loss": 3.4054, "norm_diff": 0.0516, "norm_loss": 0.0, "num_token_doc": 66.7609, "num_token_overlap": 11.6873, "num_token_query": 31.4346, "num_token_union": 65.1381, "num_word_context": 202.3035, "num_word_doc": 49.8323, "num_word_query": 23.3464, "postclip_grad_norm": 1.0, "preclip_grad_norm": 29468.489, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 3.2461, "query_norm": 2.1941, "queue_k_norm": 2.248, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4346, "sent_len_1": 66.7609, "sent_len_max_0": 127.99, "sent_len_max_1": 208.895, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125300 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.3951, "doc_norm": 2.2504, "encoder_q-embeddings": 2568.0388, "encoder_q-layer.0": 1787.052, "encoder_q-layer.1": 1863.4839, "encoder_q-layer.10": 3661.9016, "encoder_q-layer.11": 8350.0195, "encoder_q-layer.2": 2086.4331, "encoder_q-layer.3": 2244.6382, "encoder_q-layer.4": 2349.0068, "encoder_q-layer.5": 2325.5659, "encoder_q-layer.6": 2559.0476, "encoder_q-layer.7": 2997.1726, "encoder_q-layer.8": 3536.4189, "encoder_q-layer.9": 3286.4102, "epoch": 0.54, "inbatch_neg_score": 3.255, "inbatch_pos_score": 3.8574, "learning_rate": 1.963157894736842e-05, "loss": 3.3951, "norm_diff": 0.0521, "norm_loss": 0.0, "num_token_doc": 66.7623, "num_token_overlap": 11.6522, "num_token_query": 31.3801, "num_token_union": 65.1383, "num_word_context": 202.2709, "num_word_doc": 49.8162, "num_word_query": 23.312, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5262.7951, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2539, "query_norm": 2.1983, "queue_k_norm": 2.2496, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3801, "sent_len_1": 66.7623, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.575, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125400 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.4019, "doc_norm": 2.2511, "encoder_q-embeddings": 2287.8513, "encoder_q-layer.0": 1673.8639, "encoder_q-layer.1": 1758.1382, "encoder_q-layer.10": 3481.0, "encoder_q-layer.11": 7947.6099, "encoder_q-layer.2": 2000.1702, "encoder_q-layer.3": 2021.4948, "encoder_q-layer.4": 2232.7153, "encoder_q-layer.5": 2289.4993, "encoder_q-layer.6": 2571.8108, "encoder_q-layer.7": 2877.1819, "encoder_q-layer.8": 3537.0657, "encoder_q-layer.9": 3270.2075, "epoch": 0.54, "inbatch_neg_score": 3.2585, "inbatch_pos_score": 3.8672, "learning_rate": 1.9605263157894736e-05, "loss": 3.4019, "norm_diff": 0.0578, "norm_loss": 0.0, "num_token_doc": 66.6707, "num_token_overlap": 11.6568, "num_token_query": 31.3621, "num_token_union": 65.0573, "num_word_context": 202.1439, "num_word_doc": 49.7437, "num_word_query": 23.2928, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5058.4434, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2578, "query_norm": 2.1933, "queue_k_norm": 2.2501, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3621, "sent_len_1": 66.6707, "sent_len_max_0": 127.9625, "sent_len_max_1": 208.8638, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125500 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.4188, "doc_norm": 2.2501, "encoder_q-embeddings": 2654.9812, "encoder_q-layer.0": 1895.344, "encoder_q-layer.1": 2074.4863, "encoder_q-layer.10": 3115.626, "encoder_q-layer.11": 7928.7241, "encoder_q-layer.2": 2429.9089, "encoder_q-layer.3": 2458.5615, "encoder_q-layer.4": 2530.4773, "encoder_q-layer.5": 2580.9028, "encoder_q-layer.6": 2828.7061, "encoder_q-layer.7": 3137.4282, "encoder_q-layer.8": 3577.0784, "encoder_q-layer.9": 3184.6042, "epoch": 0.55, "inbatch_neg_score": 3.2652, "inbatch_pos_score": 3.8652, "learning_rate": 1.957894736842105e-05, "loss": 3.4188, "norm_diff": 0.059, "norm_loss": 0.0, "num_token_doc": 66.7307, "num_token_overlap": 11.6548, "num_token_query": 31.3381, "num_token_union": 65.0661, "num_word_context": 202.109, "num_word_doc": 49.7908, "num_word_query": 23.2937, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5293.6622, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2656, "query_norm": 2.1911, "queue_k_norm": 2.253, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3381, "sent_len_1": 66.7307, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.9175, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125600 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4077, "doc_norm": 2.2539, "encoder_q-embeddings": 2294.4207, "encoder_q-layer.0": 1631.666, "encoder_q-layer.1": 1690.3511, "encoder_q-layer.10": 3238.7837, "encoder_q-layer.11": 7692.5435, "encoder_q-layer.2": 1897.4882, "encoder_q-layer.3": 1961.9202, "encoder_q-layer.4": 2074.0852, "encoder_q-layer.5": 2128.1851, "encoder_q-layer.6": 2381.9094, "encoder_q-layer.7": 2786.9448, "encoder_q-layer.8": 3275.4766, "encoder_q-layer.9": 3016.6335, "epoch": 0.55, "inbatch_neg_score": 3.268, "inbatch_pos_score": 3.877, "learning_rate": 1.955263157894737e-05, "loss": 3.4077, "norm_diff": 0.0592, "norm_loss": 0.0, "num_token_doc": 66.7536, "num_token_overlap": 11.7032, "num_token_query": 31.4824, "num_token_union": 65.1751, "num_word_context": 202.0938, "num_word_doc": 49.8049, "num_word_query": 23.3968, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4860.0074, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2695, "query_norm": 2.1947, "queue_k_norm": 2.2531, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4824, "sent_len_1": 66.7536, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.6975, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125700 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.4118, "doc_norm": 2.2538, "encoder_q-embeddings": 2326.8552, "encoder_q-layer.0": 1639.7384, "encoder_q-layer.1": 1726.9966, "encoder_q-layer.10": 3218.1948, "encoder_q-layer.11": 7885.4814, "encoder_q-layer.2": 1925.9071, "encoder_q-layer.3": 1954.4478, "encoder_q-layer.4": 2061.3088, "encoder_q-layer.5": 2142.5439, "encoder_q-layer.6": 2472.6294, "encoder_q-layer.7": 2841.2173, "encoder_q-layer.8": 3290.0688, "encoder_q-layer.9": 3023.8994, "epoch": 0.55, "inbatch_neg_score": 3.2725, "inbatch_pos_score": 3.8848, "learning_rate": 1.9526315789473685e-05, "loss": 3.4118, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7069, "num_token_overlap": 11.7015, "num_token_query": 31.3339, "num_token_union": 65.0063, "num_word_context": 202.0487, "num_word_doc": 49.7706, "num_word_query": 23.2801, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4872.6306, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2715, "query_norm": 2.1843, "queue_k_norm": 2.2553, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3339, "sent_len_1": 66.7069, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.095, "stdk": 0.0492, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125800 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.4022, "doc_norm": 2.2569, "encoder_q-embeddings": 2345.9373, "encoder_q-layer.0": 1650.4507, "encoder_q-layer.1": 1743.2058, "encoder_q-layer.10": 3305.0327, "encoder_q-layer.11": 7947.0166, "encoder_q-layer.2": 1939.2981, "encoder_q-layer.3": 2012.175, "encoder_q-layer.4": 2173.3579, "encoder_q-layer.5": 2157.5063, "encoder_q-layer.6": 2443.656, "encoder_q-layer.7": 2725.7554, "encoder_q-layer.8": 3331.5466, "encoder_q-layer.9": 3150.978, "epoch": 0.55, "inbatch_neg_score": 3.2756, "inbatch_pos_score": 3.8848, "learning_rate": 1.9500000000000003e-05, "loss": 3.4022, "norm_diff": 0.0652, "norm_loss": 0.0, "num_token_doc": 66.7098, "num_token_overlap": 11.6859, "num_token_query": 31.3282, "num_token_union": 65.0751, "num_word_context": 202.1225, "num_word_doc": 49.7787, "num_word_query": 23.2655, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4940.4902, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2773, "query_norm": 2.1918, "queue_k_norm": 2.2565, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3282, "sent_len_1": 66.7098, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.7837, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 125900 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4021, "doc_norm": 2.2544, "encoder_q-embeddings": 2662.8101, "encoder_q-layer.0": 1858.8577, "encoder_q-layer.1": 1983.9783, "encoder_q-layer.10": 3281.2607, "encoder_q-layer.11": 7664.6816, "encoder_q-layer.2": 2321.5442, "encoder_q-layer.3": 2451.8521, "encoder_q-layer.4": 2746.0325, "encoder_q-layer.5": 2847.2107, "encoder_q-layer.6": 3104.2935, "encoder_q-layer.7": 3399.7612, "encoder_q-layer.8": 3511.9241, "encoder_q-layer.9": 3085.4255, "epoch": 0.55, "inbatch_neg_score": 3.278, "inbatch_pos_score": 3.8809, "learning_rate": 1.9473684210526315e-05, "loss": 3.4021, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.7254, "num_token_overlap": 11.6723, "num_token_query": 31.2554, "num_token_union": 65.0064, "num_word_context": 202.0676, "num_word_doc": 49.754, "num_word_query": 23.1902, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5175.7498, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2773, "query_norm": 2.1826, "queue_k_norm": 2.2574, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2554, "sent_len_1": 66.7254, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.39, "stdk": 0.0491, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126000 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.3951, "doc_norm": 2.2589, "encoder_q-embeddings": 2508.8064, "encoder_q-layer.0": 1743.9476, "encoder_q-layer.1": 1947.5461, "encoder_q-layer.10": 3139.2605, "encoder_q-layer.11": 7683.6709, "encoder_q-layer.2": 2158.344, "encoder_q-layer.3": 2181.302, "encoder_q-layer.4": 2322.4021, "encoder_q-layer.5": 2303.3955, "encoder_q-layer.6": 2523.4512, "encoder_q-layer.7": 2836.5667, "encoder_q-layer.8": 3405.8406, "encoder_q-layer.9": 3074.3936, "epoch": 0.55, "inbatch_neg_score": 3.2789, "inbatch_pos_score": 3.875, "learning_rate": 1.9447368421052633e-05, "loss": 3.3951, "norm_diff": 0.0708, "norm_loss": 0.0, "num_token_doc": 66.7646, "num_token_overlap": 11.6819, "num_token_query": 31.3686, "num_token_union": 65.1206, "num_word_context": 202.111, "num_word_doc": 49.7668, "num_word_query": 23.2821, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5005.5394, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2793, "query_norm": 2.1881, "queue_k_norm": 2.2594, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3686, "sent_len_1": 66.7646, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.3913, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126100 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.4092, "doc_norm": 2.259, "encoder_q-embeddings": 2323.3977, "encoder_q-layer.0": 1610.5161, "encoder_q-layer.1": 1715.1813, "encoder_q-layer.10": 3134.4146, "encoder_q-layer.11": 7915.4619, "encoder_q-layer.2": 1922.2452, "encoder_q-layer.3": 1964.3043, "encoder_q-layer.4": 2141.7556, "encoder_q-layer.5": 2185.3064, "encoder_q-layer.6": 2471.311, "encoder_q-layer.7": 2862.3499, "encoder_q-layer.8": 3466.613, "encoder_q-layer.9": 3031.8457, "epoch": 0.55, "inbatch_neg_score": 3.2799, "inbatch_pos_score": 3.8887, "learning_rate": 1.9421052631578948e-05, "loss": 3.4092, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.7708, "num_token_overlap": 11.685, "num_token_query": 31.395, "num_token_union": 65.137, "num_word_context": 202.4795, "num_word_doc": 49.8506, "num_word_query": 23.3204, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4902.5466, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2793, "query_norm": 2.1861, "queue_k_norm": 2.2611, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.395, "sent_len_1": 66.7708, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.93, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 126200 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3957, "doc_norm": 2.2622, "encoder_q-embeddings": 2358.574, "encoder_q-layer.0": 1687.1057, "encoder_q-layer.1": 1788.0087, "encoder_q-layer.10": 2949.3953, "encoder_q-layer.11": 7377.2192, "encoder_q-layer.2": 2017.8719, "encoder_q-layer.3": 2018.491, "encoder_q-layer.4": 2215.3413, "encoder_q-layer.5": 2246.2725, "encoder_q-layer.6": 2520.1233, "encoder_q-layer.7": 2711.1541, "encoder_q-layer.8": 3168.2644, "encoder_q-layer.9": 2835.0952, "epoch": 0.55, "inbatch_neg_score": 3.2845, "inbatch_pos_score": 3.8945, "learning_rate": 1.9394736842105263e-05, "loss": 3.3957, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.7864, "num_token_overlap": 11.7269, "num_token_query": 31.4858, "num_token_union": 65.1596, "num_word_context": 202.1633, "num_word_doc": 49.8408, "num_word_query": 23.3915, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4740.6531, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2852, "query_norm": 2.1877, "queue_k_norm": 2.2615, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4858, "sent_len_1": 66.7864, "sent_len_max_0": 127.985, "sent_len_max_1": 208.735, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126300 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.4122, "doc_norm": 2.262, "encoder_q-embeddings": 3005.6548, "encoder_q-layer.0": 2046.6709, "encoder_q-layer.1": 2075.2524, "encoder_q-layer.10": 3124.6536, "encoder_q-layer.11": 7980.2686, "encoder_q-layer.2": 2320.4065, "encoder_q-layer.3": 2301.2424, "encoder_q-layer.4": 2352.467, "encoder_q-layer.5": 2385.9507, "encoder_q-layer.6": 2655.8862, "encoder_q-layer.7": 2950.5549, "encoder_q-layer.8": 3335.7288, "encoder_q-layer.9": 3028.6021, "epoch": 0.55, "inbatch_neg_score": 3.2877, "inbatch_pos_score": 3.8867, "learning_rate": 1.936842105263158e-05, "loss": 3.4122, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.67, "num_token_overlap": 11.6475, "num_token_query": 31.2818, "num_token_union": 65.0214, "num_word_context": 202.0014, "num_word_doc": 49.765, "num_word_query": 23.2337, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5203.0004, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2871, "query_norm": 2.187, "queue_k_norm": 2.2618, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2818, "sent_len_1": 66.67, "sent_len_max_0": 127.9975, "sent_len_max_1": 206.4125, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126400 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.3905, "doc_norm": 2.2662, "encoder_q-embeddings": 2417.5732, "encoder_q-layer.0": 1737.0264, "encoder_q-layer.1": 1754.4209, "encoder_q-layer.10": 3029.7095, "encoder_q-layer.11": 7445.7754, "encoder_q-layer.2": 1954.7919, "encoder_q-layer.3": 1962.8737, "encoder_q-layer.4": 2091.5791, "encoder_q-layer.5": 2193.4792, "encoder_q-layer.6": 2380.1411, "encoder_q-layer.7": 2788.6477, "encoder_q-layer.8": 3442.1541, "encoder_q-layer.9": 3114.7988, "epoch": 0.55, "inbatch_neg_score": 3.2899, "inbatch_pos_score": 3.8887, "learning_rate": 1.9342105263157896e-05, "loss": 3.3905, "norm_diff": 0.0835, "norm_loss": 0.0, "num_token_doc": 66.7795, "num_token_overlap": 11.7356, "num_token_query": 31.5273, "num_token_union": 65.1839, "num_word_context": 202.234, "num_word_doc": 49.832, "num_word_query": 23.4054, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4754.2152, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.291, "query_norm": 2.1828, "queue_k_norm": 2.2639, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5273, "sent_len_1": 66.7795, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.615, "stdk": 0.0495, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126500 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.4028, "doc_norm": 2.268, "encoder_q-embeddings": 4005.2334, "encoder_q-layer.0": 2705.9866, "encoder_q-layer.1": 2831.2366, "encoder_q-layer.10": 3307.9629, "encoder_q-layer.11": 7938.3584, "encoder_q-layer.2": 3153.8279, "encoder_q-layer.3": 3234.7537, "encoder_q-layer.4": 3506.3252, "encoder_q-layer.5": 3545.1631, "encoder_q-layer.6": 3738.2083, "encoder_q-layer.7": 3606.2266, "encoder_q-layer.8": 3787.144, "encoder_q-layer.9": 3197.125, "epoch": 0.55, "inbatch_neg_score": 3.2934, "inbatch_pos_score": 3.9082, "learning_rate": 1.931578947368421e-05, "loss": 3.4028, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.5681, "num_token_overlap": 11.6683, "num_token_query": 31.4114, "num_token_union": 65.0715, "num_word_context": 202.1542, "num_word_doc": 49.7, "num_word_query": 23.3365, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5998.9522, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.293, "query_norm": 2.1966, "queue_k_norm": 2.2651, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4114, "sent_len_1": 66.5681, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.83, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126600 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.397, "doc_norm": 2.267, "encoder_q-embeddings": 2485.3462, "encoder_q-layer.0": 1740.7, "encoder_q-layer.1": 1841.9454, "encoder_q-layer.10": 3349.7991, "encoder_q-layer.11": 7571.7344, "encoder_q-layer.2": 2116.8945, "encoder_q-layer.3": 2207.3887, "encoder_q-layer.4": 2369.2317, "encoder_q-layer.5": 2413.2227, "encoder_q-layer.6": 2708.5591, "encoder_q-layer.7": 2894.1191, "encoder_q-layer.8": 3439.8743, "encoder_q-layer.9": 3158.0923, "epoch": 0.55, "inbatch_neg_score": 3.296, "inbatch_pos_score": 3.916, "learning_rate": 1.9289473684210526e-05, "loss": 3.397, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.9296, "num_token_overlap": 11.6961, "num_token_query": 31.3772, "num_token_union": 65.182, "num_word_context": 202.4503, "num_word_doc": 49.9424, "num_word_query": 23.3026, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4892.6834, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.2949, "query_norm": 2.1926, "queue_k_norm": 2.2666, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3772, "sent_len_1": 66.9296, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.05, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 126700 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.4155, "doc_norm": 2.267, "encoder_q-embeddings": 2402.6626, "encoder_q-layer.0": 1722.3296, "encoder_q-layer.1": 1836.0123, "encoder_q-layer.10": 3434.8542, "encoder_q-layer.11": 8265.1885, "encoder_q-layer.2": 2125.6685, "encoder_q-layer.3": 2125.551, "encoder_q-layer.4": 2260.9072, "encoder_q-layer.5": 2281.2126, "encoder_q-layer.6": 2599.7058, "encoder_q-layer.7": 3016.8633, "encoder_q-layer.8": 3530.9907, "encoder_q-layer.9": 3287.8369, "epoch": 0.55, "inbatch_neg_score": 3.3006, "inbatch_pos_score": 3.8867, "learning_rate": 1.9263157894736845e-05, "loss": 3.4155, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7728, "num_token_overlap": 11.6702, "num_token_query": 31.3191, "num_token_union": 65.0867, "num_word_context": 202.2386, "num_word_doc": 49.8067, "num_word_query": 23.2518, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5193.378, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.3008, "query_norm": 2.1874, "queue_k_norm": 2.2668, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3191, "sent_len_1": 66.7728, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.0575, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 126800 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.4053, "doc_norm": 2.2666, "encoder_q-embeddings": 2865.3699, "encoder_q-layer.0": 1973.8743, "encoder_q-layer.1": 2014.493, "encoder_q-layer.10": 3239.2393, "encoder_q-layer.11": 7540.5127, "encoder_q-layer.2": 2161.6482, "encoder_q-layer.3": 2245.6855, "encoder_q-layer.4": 2422.5847, "encoder_q-layer.5": 2469.9155, "encoder_q-layer.6": 2752.6587, "encoder_q-layer.7": 3332.498, "encoder_q-layer.8": 3624.5188, "encoder_q-layer.9": 3111.1582, "epoch": 0.55, "inbatch_neg_score": 3.3049, "inbatch_pos_score": 3.9121, "learning_rate": 1.923684210526316e-05, "loss": 3.4053, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.7535, "num_token_overlap": 11.6475, "num_token_query": 31.39, "num_token_union": 65.168, "num_word_context": 202.3025, "num_word_doc": 49.8267, "num_word_query": 23.3232, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5102.9828, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.3047, "query_norm": 2.1944, "queue_k_norm": 2.2682, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.39, "sent_len_1": 66.7535, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.7663, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 126900 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4096, "doc_norm": 2.2672, "encoder_q-embeddings": 2370.7832, "encoder_q-layer.0": 1668.4424, "encoder_q-layer.1": 1747.6246, "encoder_q-layer.10": 3225.7119, "encoder_q-layer.11": 7927.6162, "encoder_q-layer.2": 1961.1018, "encoder_q-layer.3": 1971.3391, "encoder_q-layer.4": 2022.8931, "encoder_q-layer.5": 2120.7839, "encoder_q-layer.6": 2398.2805, "encoder_q-layer.7": 2753.231, "encoder_q-layer.8": 3221.0, "encoder_q-layer.9": 2937.7205, "epoch": 0.55, "inbatch_neg_score": 3.3095, "inbatch_pos_score": 3.9062, "learning_rate": 1.9210526315789474e-05, "loss": 3.4096, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.9508, "num_token_overlap": 11.6176, "num_token_query": 31.2585, "num_token_union": 65.2019, "num_word_context": 202.6135, "num_word_doc": 49.9701, "num_word_query": 23.2098, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4770.092, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.3105, "query_norm": 2.1943, "queue_k_norm": 2.2701, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2585, "sent_len_1": 66.9508, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.3975, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 127000 }, { "accuracy": 49.3164, "active_queue_size": 16384.0, "cl_loss": 3.3934, "doc_norm": 2.2667, "encoder_q-embeddings": 2453.2844, "encoder_q-layer.0": 1819.5447, "encoder_q-layer.1": 1907.5623, "encoder_q-layer.10": 3422.386, "encoder_q-layer.11": 8100.4106, "encoder_q-layer.2": 2190.4883, "encoder_q-layer.3": 2221.4834, "encoder_q-layer.4": 2389.6282, "encoder_q-layer.5": 2424.6599, "encoder_q-layer.6": 2714.3086, "encoder_q-layer.7": 3075.6814, "encoder_q-layer.8": 3567.5591, "encoder_q-layer.9": 3366.1946, "epoch": 0.55, "inbatch_neg_score": 3.3168, "inbatch_pos_score": 3.9121, "learning_rate": 1.918421052631579e-05, "loss": 3.3934, "norm_diff": 0.0648, "norm_loss": 0.0, "num_token_doc": 66.5591, "num_token_overlap": 11.6589, "num_token_query": 31.3455, "num_token_union": 64.9695, "num_word_context": 202.1085, "num_word_doc": 49.6698, "num_word_query": 23.2865, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5116.6044, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.3164, "query_norm": 2.2018, "queue_k_norm": 2.2705, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3455, "sent_len_1": 66.5591, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.8063, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 127100 }, { "accuracy": 48.6328, "active_queue_size": 16384.0, "cl_loss": 3.3939, "doc_norm": 2.2697, "encoder_q-embeddings": 5047.4893, "encoder_q-layer.0": 3423.3813, "encoder_q-layer.1": 3614.7666, "encoder_q-layer.10": 7600.6465, "encoder_q-layer.11": 16307.2783, "encoder_q-layer.2": 4052.906, "encoder_q-layer.3": 4264.9316, "encoder_q-layer.4": 4549.1396, "encoder_q-layer.5": 4675.6035, "encoder_q-layer.6": 5246.7417, "encoder_q-layer.7": 5788.7847, "encoder_q-layer.8": 7366.0137, "encoder_q-layer.9": 6943.6899, "epoch": 0.55, "inbatch_neg_score": 3.3187, "inbatch_pos_score": 3.9199, "learning_rate": 1.9157894736842104e-05, "loss": 3.3939, "norm_diff": 0.072, "norm_loss": 0.0, "num_token_doc": 66.738, "num_token_overlap": 11.6872, "num_token_query": 31.4147, "num_token_union": 65.0749, "num_word_context": 202.2035, "num_word_doc": 49.791, "num_word_query": 23.3295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10305.9002, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3184, "query_norm": 2.1977, "queue_k_norm": 2.27, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4147, "sent_len_1": 66.738, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.1125, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127200 }, { "accuracy": 53.3691, "active_queue_size": 16384.0, "cl_loss": 3.4054, "doc_norm": 2.2747, "encoder_q-embeddings": 4838.9644, "encoder_q-layer.0": 3353.5872, "encoder_q-layer.1": 3440.9878, "encoder_q-layer.10": 6632.9805, "encoder_q-layer.11": 15035.4365, "encoder_q-layer.2": 3855.8284, "encoder_q-layer.3": 3913.875, "encoder_q-layer.4": 4097.4595, "encoder_q-layer.5": 4203.6265, "encoder_q-layer.6": 4653.8828, "encoder_q-layer.7": 5382.4189, "encoder_q-layer.8": 6481.1768, "encoder_q-layer.9": 6316.1455, "epoch": 0.55, "inbatch_neg_score": 3.327, "inbatch_pos_score": 3.957, "learning_rate": 1.9131578947368423e-05, "loss": 3.4054, "norm_diff": 0.0647, "norm_loss": 0.0, "num_token_doc": 66.7567, "num_token_overlap": 11.676, "num_token_query": 31.4077, "num_token_union": 65.147, "num_word_context": 202.5377, "num_word_doc": 49.797, "num_word_query": 23.3312, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9607.5103, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3281, "query_norm": 2.21, "queue_k_norm": 2.2712, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4077, "sent_len_1": 66.7567, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.1937, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127300 }, { "accuracy": 52.2949, "active_queue_size": 16384.0, "cl_loss": 3.4103, "doc_norm": 2.2716, "encoder_q-embeddings": 4666.9521, "encoder_q-layer.0": 3235.3894, "encoder_q-layer.1": 3528.4395, "encoder_q-layer.10": 6529.0171, "encoder_q-layer.11": 15843.5254, "encoder_q-layer.2": 3951.8835, "encoder_q-layer.3": 3932.7341, "encoder_q-layer.4": 4149.5054, "encoder_q-layer.5": 4384.8545, "encoder_q-layer.6": 4729.3799, "encoder_q-layer.7": 5430.2158, "encoder_q-layer.8": 6961.5459, "encoder_q-layer.9": 6411.9702, "epoch": 0.55, "inbatch_neg_score": 3.3333, "inbatch_pos_score": 3.9414, "learning_rate": 1.9105263157894738e-05, "loss": 3.4103, "norm_diff": 0.0661, "norm_loss": 0.0, "num_token_doc": 66.574, "num_token_overlap": 11.6412, "num_token_query": 31.2498, "num_token_union": 64.9486, "num_word_context": 202.0792, "num_word_doc": 49.6604, "num_word_query": 23.2023, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9858.5677, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.332, "query_norm": 2.2055, "queue_k_norm": 2.2725, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2498, "sent_len_1": 66.574, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.8625, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127400 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.4093, "doc_norm": 2.2714, "encoder_q-embeddings": 5442.6914, "encoder_q-layer.0": 3989.3076, "encoder_q-layer.1": 4456.5674, "encoder_q-layer.10": 6572.6621, "encoder_q-layer.11": 15761.4502, "encoder_q-layer.2": 4904.5498, "encoder_q-layer.3": 5226.9492, "encoder_q-layer.4": 5532.7822, "encoder_q-layer.5": 5554.2021, "encoder_q-layer.6": 6242.7114, "encoder_q-layer.7": 6623.1494, "encoder_q-layer.8": 7360.6255, "encoder_q-layer.9": 6152.6074, "epoch": 0.55, "inbatch_neg_score": 3.3366, "inbatch_pos_score": 3.9434, "learning_rate": 1.9078947368421056e-05, "loss": 3.4093, "norm_diff": 0.0594, "norm_loss": 0.0, "num_token_doc": 66.9073, "num_token_overlap": 11.6732, "num_token_query": 31.3509, "num_token_union": 65.1858, "num_word_context": 202.4811, "num_word_doc": 49.9211, "num_word_query": 23.2805, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10737.4202, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3359, "query_norm": 2.2119, "queue_k_norm": 2.2743, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3509, "sent_len_1": 66.9073, "sent_len_max_0": 127.94, "sent_len_max_1": 209.3562, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 127500 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3954, "doc_norm": 2.2738, "encoder_q-embeddings": 4737.6255, "encoder_q-layer.0": 3393.8923, "encoder_q-layer.1": 3717.8931, "encoder_q-layer.10": 6103.0605, "encoder_q-layer.11": 15157.876, "encoder_q-layer.2": 4165.5352, "encoder_q-layer.3": 4259.0542, "encoder_q-layer.4": 4363.3843, "encoder_q-layer.5": 4363.73, "encoder_q-layer.6": 4735.1152, "encoder_q-layer.7": 5299.1631, "encoder_q-layer.8": 6366.5728, "encoder_q-layer.9": 5968.1655, "epoch": 0.55, "inbatch_neg_score": 3.3446, "inbatch_pos_score": 3.9551, "learning_rate": 1.9052631578947368e-05, "loss": 3.3954, "norm_diff": 0.0618, "norm_loss": 0.0, "num_token_doc": 66.7606, "num_token_overlap": 11.6993, "num_token_query": 31.4083, "num_token_union": 65.1174, "num_word_context": 202.4222, "num_word_doc": 49.8124, "num_word_query": 23.3443, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9600.3147, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3438, "query_norm": 2.212, "queue_k_norm": 2.2735, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4083, "sent_len_1": 66.7606, "sent_len_max_0": 127.975, "sent_len_max_1": 210.6775, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127600 }, { "accuracy": 53.0273, "active_queue_size": 16384.0, "cl_loss": 3.3899, "doc_norm": 2.2748, "encoder_q-embeddings": 5060.5239, "encoder_q-layer.0": 3450.3613, "encoder_q-layer.1": 3668.3152, "encoder_q-layer.10": 6502.208, "encoder_q-layer.11": 15572.3691, "encoder_q-layer.2": 4111.7104, "encoder_q-layer.3": 4408.1387, "encoder_q-layer.4": 4721.9136, "encoder_q-layer.5": 4985.1016, "encoder_q-layer.6": 5448.165, "encoder_q-layer.7": 6127.9263, "encoder_q-layer.8": 7366.2925, "encoder_q-layer.9": 6337.4473, "epoch": 0.55, "inbatch_neg_score": 3.3514, "inbatch_pos_score": 3.9727, "learning_rate": 1.9026315789473683e-05, "loss": 3.3899, "norm_diff": 0.0501, "norm_loss": 0.0, "num_token_doc": 66.7086, "num_token_overlap": 11.691, "num_token_query": 31.4133, "num_token_union": 65.0823, "num_word_context": 202.365, "num_word_doc": 49.7555, "num_word_query": 23.3298, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10026.6008, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3516, "query_norm": 2.2247, "queue_k_norm": 2.2737, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4133, "sent_len_1": 66.7086, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.3862, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 127700 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3922, "doc_norm": 2.2767, "encoder_q-embeddings": 4819.7173, "encoder_q-layer.0": 3351.5024, "encoder_q-layer.1": 3577.7148, "encoder_q-layer.10": 6176.7881, "encoder_q-layer.11": 15010.6309, "encoder_q-layer.2": 3995.9028, "encoder_q-layer.3": 4036.2708, "encoder_q-layer.4": 4254.7593, "encoder_q-layer.5": 4506.2334, "encoder_q-layer.6": 5004.9795, "encoder_q-layer.7": 5545.9302, "encoder_q-layer.8": 6978.5259, "encoder_q-layer.9": 6023.313, "epoch": 0.55, "inbatch_neg_score": 3.3641, "inbatch_pos_score": 3.9746, "learning_rate": 1.9e-05, "loss": 3.3922, "norm_diff": 0.0516, "norm_loss": 0.0, "num_token_doc": 66.7358, "num_token_overlap": 11.7, "num_token_query": 31.4896, "num_token_union": 65.1561, "num_word_context": 202.2779, "num_word_doc": 49.7951, "num_word_query": 23.4119, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9647.7872, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3652, "query_norm": 2.2251, "queue_k_norm": 2.2771, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4896, "sent_len_1": 66.7358, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.7475, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127800 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.3945, "doc_norm": 2.2763, "encoder_q-embeddings": 5806.6743, "encoder_q-layer.0": 4219.8125, "encoder_q-layer.1": 4590.6133, "encoder_q-layer.10": 7211.3174, "encoder_q-layer.11": 16195.7842, "encoder_q-layer.2": 5133.2188, "encoder_q-layer.3": 5588.0532, "encoder_q-layer.4": 6245.9663, "encoder_q-layer.5": 5796.7188, "encoder_q-layer.6": 5693.812, "encoder_q-layer.7": 6144.8535, "encoder_q-layer.8": 7574.7246, "encoder_q-layer.9": 6796.0879, "epoch": 0.55, "inbatch_neg_score": 3.3713, "inbatch_pos_score": 3.9629, "learning_rate": 1.8973684210526316e-05, "loss": 3.3945, "norm_diff": 0.0583, "norm_loss": 0.0, "num_token_doc": 66.7705, "num_token_overlap": 11.6913, "num_token_query": 31.4309, "num_token_union": 65.1476, "num_word_context": 202.2689, "num_word_doc": 49.8377, "num_word_query": 23.3603, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10952.0707, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3711, "query_norm": 2.218, "queue_k_norm": 2.278, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4309, "sent_len_1": 66.7705, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.0975, "stdk": 0.0493, "stdq": 0.0426, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 127900 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.4075, "doc_norm": 2.2797, "encoder_q-embeddings": 5027.8169, "encoder_q-layer.0": 3607.0029, "encoder_q-layer.1": 3947.0972, "encoder_q-layer.10": 6694.3345, "encoder_q-layer.11": 16347.5439, "encoder_q-layer.2": 4631.8115, "encoder_q-layer.3": 4800.085, "encoder_q-layer.4": 5083.1592, "encoder_q-layer.5": 5097.75, "encoder_q-layer.6": 5512.4854, "encoder_q-layer.7": 5978.459, "encoder_q-layer.8": 6828.9551, "encoder_q-layer.9": 6377.1621, "epoch": 0.56, "inbatch_neg_score": 3.3793, "inbatch_pos_score": 3.9961, "learning_rate": 1.8947368421052634e-05, "loss": 3.4075, "norm_diff": 0.0429, "norm_loss": 0.0, "num_token_doc": 66.7533, "num_token_overlap": 11.6779, "num_token_query": 31.3985, "num_token_union": 65.1431, "num_word_context": 202.4659, "num_word_doc": 49.8122, "num_word_query": 23.3204, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10444.8793, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3789, "query_norm": 2.2368, "queue_k_norm": 2.2791, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3985, "sent_len_1": 66.7533, "sent_len_max_0": 127.98, "sent_len_max_1": 207.8125, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128000 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3946, "doc_norm": 2.278, "encoder_q-embeddings": 5122.375, "encoder_q-layer.0": 3587.2654, "encoder_q-layer.1": 3856.2793, "encoder_q-layer.10": 6920.9097, "encoder_q-layer.11": 15928.2881, "encoder_q-layer.2": 4292.7822, "encoder_q-layer.3": 4311.874, "encoder_q-layer.4": 4709.2451, "encoder_q-layer.5": 4743.3804, "encoder_q-layer.6": 5232.5005, "encoder_q-layer.7": 6167.5586, "encoder_q-layer.8": 7433.8896, "encoder_q-layer.9": 6592.0186, "epoch": 0.56, "inbatch_neg_score": 3.3919, "inbatch_pos_score": 4.0078, "learning_rate": 1.892105263157895e-05, "loss": 3.3946, "norm_diff": 0.0345, "norm_loss": 0.0, "num_token_doc": 66.8474, "num_token_overlap": 11.6761, "num_token_query": 31.322, "num_token_union": 65.1187, "num_word_context": 202.2959, "num_word_doc": 49.8676, "num_word_query": 23.246, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10105.7588, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.3926, "query_norm": 2.2435, "queue_k_norm": 2.2801, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.322, "sent_len_1": 66.8474, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.1238, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128100 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.4011, "doc_norm": 2.2844, "encoder_q-embeddings": 4457.0635, "encoder_q-layer.0": 3184.8511, "encoder_q-layer.1": 3375.1089, "encoder_q-layer.10": 6089.646, "encoder_q-layer.11": 15222.6348, "encoder_q-layer.2": 3842.9937, "encoder_q-layer.3": 3815.7871, "encoder_q-layer.4": 3994.0261, "encoder_q-layer.5": 4146.5195, "encoder_q-layer.6": 4668.7227, "encoder_q-layer.7": 5144.0322, "encoder_q-layer.8": 6393.1328, "encoder_q-layer.9": 5806.1431, "epoch": 0.56, "inbatch_neg_score": 3.4003, "inbatch_pos_score": 4.0156, "learning_rate": 1.8894736842105264e-05, "loss": 3.4011, "norm_diff": 0.0425, "norm_loss": 0.0, "num_token_doc": 66.749, "num_token_overlap": 11.6859, "num_token_query": 31.3418, "num_token_union": 65.0751, "num_word_context": 202.4735, "num_word_doc": 49.7988, "num_word_query": 23.269, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9291.5902, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4004, "query_norm": 2.2418, "queue_k_norm": 2.2811, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3418, "sent_len_1": 66.749, "sent_len_max_0": 127.9613, "sent_len_max_1": 209.1538, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128200 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.4028, "doc_norm": 2.2833, "encoder_q-embeddings": 4987.6006, "encoder_q-layer.0": 3690.02, "encoder_q-layer.1": 3980.647, "encoder_q-layer.10": 6185.7344, "encoder_q-layer.11": 15088.0049, "encoder_q-layer.2": 4406.3628, "encoder_q-layer.3": 4368.3906, "encoder_q-layer.4": 4577.2422, "encoder_q-layer.5": 4897.4541, "encoder_q-layer.6": 5227.4717, "encoder_q-layer.7": 5851.8242, "encoder_q-layer.8": 6715.8677, "encoder_q-layer.9": 6366.5957, "epoch": 0.56, "inbatch_neg_score": 3.4119, "inbatch_pos_score": 4.0273, "learning_rate": 1.886842105263158e-05, "loss": 3.4028, "norm_diff": 0.0356, "norm_loss": 0.0, "num_token_doc": 66.8873, "num_token_overlap": 11.6647, "num_token_query": 31.344, "num_token_union": 65.1709, "num_word_context": 202.4507, "num_word_doc": 49.8886, "num_word_query": 23.293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9864.0613, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4121, "query_norm": 2.2477, "queue_k_norm": 2.2836, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.344, "sent_len_1": 66.8873, "sent_len_max_0": 127.9737, "sent_len_max_1": 211.0337, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128300 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3962, "doc_norm": 2.2855, "encoder_q-embeddings": 8092.8408, "encoder_q-layer.0": 5706.7686, "encoder_q-layer.1": 6744.4658, "encoder_q-layer.10": 7333.1289, "encoder_q-layer.11": 15833.9746, "encoder_q-layer.2": 7824.7534, "encoder_q-layer.3": 8597.2822, "encoder_q-layer.4": 8732.8516, "encoder_q-layer.5": 8970.5166, "encoder_q-layer.6": 9290.3662, "encoder_q-layer.7": 9844.792, "encoder_q-layer.8": 8678.4004, "encoder_q-layer.9": 6778.1948, "epoch": 0.56, "inbatch_neg_score": 3.4235, "inbatch_pos_score": 4.0391, "learning_rate": 1.8842105263157894e-05, "loss": 3.3962, "norm_diff": 0.0378, "norm_loss": 0.0, "num_token_doc": 66.772, "num_token_overlap": 11.6724, "num_token_query": 31.3256, "num_token_union": 65.0857, "num_word_context": 202.2734, "num_word_doc": 49.8296, "num_word_query": 23.2718, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13298.5695, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4238, "query_norm": 2.2477, "queue_k_norm": 2.2847, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3256, "sent_len_1": 66.772, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.2875, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128400 }, { "accuracy": 53.2227, "active_queue_size": 16384.0, "cl_loss": 3.4056, "doc_norm": 2.2881, "encoder_q-embeddings": 4791.3535, "encoder_q-layer.0": 3357.8406, "encoder_q-layer.1": 3492.8374, "encoder_q-layer.10": 6925.5771, "encoder_q-layer.11": 15077.9795, "encoder_q-layer.2": 3884.072, "encoder_q-layer.3": 3967.4614, "encoder_q-layer.4": 4228.7441, "encoder_q-layer.5": 4316.79, "encoder_q-layer.6": 4894.0835, "encoder_q-layer.7": 5666.0137, "encoder_q-layer.8": 6921.4902, "encoder_q-layer.9": 6451.1909, "epoch": 0.56, "inbatch_neg_score": 3.4359, "inbatch_pos_score": 4.0625, "learning_rate": 1.8815789473684213e-05, "loss": 3.4056, "norm_diff": 0.0333, "norm_loss": 0.0, "num_token_doc": 66.8365, "num_token_overlap": 11.6499, "num_token_query": 31.3034, "num_token_union": 65.1466, "num_word_context": 202.3949, "num_word_doc": 49.8469, "num_word_query": 23.2476, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9785.8373, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4355, "query_norm": 2.2548, "queue_k_norm": 2.2871, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3034, "sent_len_1": 66.8365, "sent_len_max_0": 127.99, "sent_len_max_1": 209.1, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128500 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3797, "doc_norm": 2.2901, "encoder_q-embeddings": 4956.5933, "encoder_q-layer.0": 3459.0679, "encoder_q-layer.1": 3538.9717, "encoder_q-layer.10": 6846.1943, "encoder_q-layer.11": 15889.1045, "encoder_q-layer.2": 3992.0327, "encoder_q-layer.3": 4252.0942, "encoder_q-layer.4": 4414.8662, "encoder_q-layer.5": 4502.6411, "encoder_q-layer.6": 4892.6133, "encoder_q-layer.7": 5690.2124, "encoder_q-layer.8": 6787.7505, "encoder_q-layer.9": 6376.4946, "epoch": 0.56, "inbatch_neg_score": 3.4487, "inbatch_pos_score": 4.0664, "learning_rate": 1.8789473684210528e-05, "loss": 3.3797, "norm_diff": 0.0336, "norm_loss": 0.0, "num_token_doc": 67.0603, "num_token_overlap": 11.7062, "num_token_query": 31.3482, "num_token_union": 65.247, "num_word_context": 202.4006, "num_word_doc": 50.0424, "num_word_query": 23.2727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9951.8332, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4473, "query_norm": 2.2564, "queue_k_norm": 2.2887, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3482, "sent_len_1": 67.0603, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.215, "stdk": 0.0494, "stdq": 0.0438, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128600 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3999, "doc_norm": 2.2916, "encoder_q-embeddings": 5303.9521, "encoder_q-layer.0": 3640.7642, "encoder_q-layer.1": 3835.1274, "encoder_q-layer.10": 6478.9717, "encoder_q-layer.11": 16044.4717, "encoder_q-layer.2": 4335.2847, "encoder_q-layer.3": 4503.0488, "encoder_q-layer.4": 4812.2324, "encoder_q-layer.5": 4875.2383, "encoder_q-layer.6": 5581.0913, "encoder_q-layer.7": 6387.6299, "encoder_q-layer.8": 7381.3564, "encoder_q-layer.9": 6407.667, "epoch": 0.56, "inbatch_neg_score": 3.4534, "inbatch_pos_score": 4.0547, "learning_rate": 1.8763157894736843e-05, "loss": 3.3999, "norm_diff": 0.0439, "norm_loss": 0.0, "num_token_doc": 66.6741, "num_token_overlap": 11.6386, "num_token_query": 31.3369, "num_token_union": 65.028, "num_word_context": 202.2695, "num_word_doc": 49.7785, "num_word_query": 23.2849, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10590.4213, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4531, "query_norm": 2.2477, "queue_k_norm": 2.2926, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3369, "sent_len_1": 66.6741, "sent_len_max_0": 127.9562, "sent_len_max_1": 207.6875, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 128700 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3995, "doc_norm": 2.2926, "encoder_q-embeddings": 4671.6279, "encoder_q-layer.0": 3331.2405, "encoder_q-layer.1": 3501.6167, "encoder_q-layer.10": 6616.437, "encoder_q-layer.11": 16467.0859, "encoder_q-layer.2": 4083.3218, "encoder_q-layer.3": 4121.6021, "encoder_q-layer.4": 4390.7769, "encoder_q-layer.5": 4558.5903, "encoder_q-layer.6": 5053.0669, "encoder_q-layer.7": 5893.7778, "encoder_q-layer.8": 6831.6504, "encoder_q-layer.9": 6530.0059, "epoch": 0.56, "inbatch_neg_score": 3.4595, "inbatch_pos_score": 4.0781, "learning_rate": 1.8736842105263158e-05, "loss": 3.3995, "norm_diff": 0.0539, "norm_loss": 0.0, "num_token_doc": 66.7305, "num_token_overlap": 11.639, "num_token_query": 31.3031, "num_token_union": 65.0459, "num_word_context": 202.2809, "num_word_doc": 49.7444, "num_word_query": 23.2395, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10176.8479, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4609, "query_norm": 2.2387, "queue_k_norm": 2.2915, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3031, "sent_len_1": 66.7305, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.9187, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128800 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3894, "doc_norm": 2.2941, "encoder_q-embeddings": 4925.4746, "encoder_q-layer.0": 3479.7417, "encoder_q-layer.1": 3700.323, "encoder_q-layer.10": 6605.2417, "encoder_q-layer.11": 15491.0098, "encoder_q-layer.2": 4168.5098, "encoder_q-layer.3": 4216.8687, "encoder_q-layer.4": 4482.0771, "encoder_q-layer.5": 4521.0425, "encoder_q-layer.6": 4908.895, "encoder_q-layer.7": 6132.3311, "encoder_q-layer.8": 6645.9106, "encoder_q-layer.9": 6211.1987, "epoch": 0.56, "inbatch_neg_score": 3.4635, "inbatch_pos_score": 4.0703, "learning_rate": 1.8710526315789476e-05, "loss": 3.3894, "norm_diff": 0.0611, "norm_loss": 0.0, "num_token_doc": 66.6943, "num_token_overlap": 11.6899, "num_token_query": 31.3858, "num_token_union": 65.0698, "num_word_context": 202.3534, "num_word_doc": 49.7792, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9884.9001, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4629, "query_norm": 2.233, "queue_k_norm": 2.2946, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3858, "sent_len_1": 66.6943, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.26, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 128900 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.4087, "doc_norm": 2.2961, "encoder_q-embeddings": 4497.5327, "encoder_q-layer.0": 3218.8081, "encoder_q-layer.1": 3406.8254, "encoder_q-layer.10": 6782.5972, "encoder_q-layer.11": 15581.9219, "encoder_q-layer.2": 3812.1702, "encoder_q-layer.3": 3929.0535, "encoder_q-layer.4": 4145.8188, "encoder_q-layer.5": 4218.3457, "encoder_q-layer.6": 4775.7622, "encoder_q-layer.7": 5745.9199, "encoder_q-layer.8": 6852.4214, "encoder_q-layer.9": 6532.1631, "epoch": 0.56, "inbatch_neg_score": 3.4692, "inbatch_pos_score": 4.0742, "learning_rate": 1.868421052631579e-05, "loss": 3.4087, "norm_diff": 0.0555, "norm_loss": 0.0, "num_token_doc": 66.7049, "num_token_overlap": 11.6719, "num_token_query": 31.3573, "num_token_union": 65.0463, "num_word_context": 202.2449, "num_word_doc": 49.7258, "num_word_query": 23.2941, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9643.3338, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4688, "query_norm": 2.2405, "queue_k_norm": 2.2964, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3573, "sent_len_1": 66.7049, "sent_len_max_0": 127.9988, "sent_len_max_1": 211.6587, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129000 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3916, "doc_norm": 2.3015, "encoder_q-embeddings": 4818.8594, "encoder_q-layer.0": 3407.2556, "encoder_q-layer.1": 3612.9302, "encoder_q-layer.10": 6648.7339, "encoder_q-layer.11": 15411.4395, "encoder_q-layer.2": 4027.0027, "encoder_q-layer.3": 4134.8271, "encoder_q-layer.4": 4417.8203, "encoder_q-layer.5": 4556.8657, "encoder_q-layer.6": 5178.1279, "encoder_q-layer.7": 5754.6392, "encoder_q-layer.8": 6731.2559, "encoder_q-layer.9": 6391.6836, "epoch": 0.56, "inbatch_neg_score": 3.4655, "inbatch_pos_score": 4.0859, "learning_rate": 1.865789473684211e-05, "loss": 3.3916, "norm_diff": 0.0647, "norm_loss": 0.0, "num_token_doc": 66.8034, "num_token_overlap": 11.674, "num_token_query": 31.3844, "num_token_union": 65.1403, "num_word_context": 202.265, "num_word_doc": 49.8781, "num_word_query": 23.3007, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9853.8994, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.4668, "query_norm": 2.2368, "queue_k_norm": 2.2982, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3844, "sent_len_1": 66.8034, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.4625, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 129100 }, { "accuracy": 52.3926, "active_queue_size": 16384.0, "cl_loss": 3.4057, "doc_norm": 2.3004, "encoder_q-embeddings": 9659.4326, "encoder_q-layer.0": 6808.9272, "encoder_q-layer.1": 7155.9746, "encoder_q-layer.10": 14011.8613, "encoder_q-layer.11": 34989.7656, "encoder_q-layer.2": 7903.6724, "encoder_q-layer.3": 8291.1123, "encoder_q-layer.4": 8809.2715, "encoder_q-layer.5": 9063.8203, "encoder_q-layer.6": 9933.2031, "encoder_q-layer.7": 11438.5908, "encoder_q-layer.8": 13701.5752, "encoder_q-layer.9": 13074.2266, "epoch": 0.56, "inbatch_neg_score": 3.4708, "inbatch_pos_score": 4.0859, "learning_rate": 1.863157894736842e-05, "loss": 3.4057, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.8811, "num_token_overlap": 11.614, "num_token_query": 31.2036, "num_token_union": 65.1126, "num_word_context": 202.3886, "num_word_doc": 49.8949, "num_word_query": 23.1745, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20542.2644, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4707, "query_norm": 2.2349, "queue_k_norm": 2.2989, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2036, "sent_len_1": 66.8811, "sent_len_max_0": 127.98, "sent_len_max_1": 208.7775, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129200 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.3926, "doc_norm": 2.3013, "encoder_q-embeddings": 10793.1377, "encoder_q-layer.0": 7454.96, "encoder_q-layer.1": 8220.3906, "encoder_q-layer.10": 13348.6768, "encoder_q-layer.11": 33488.5586, "encoder_q-layer.2": 9107.3584, "encoder_q-layer.3": 9619.8867, "encoder_q-layer.4": 9929.5283, "encoder_q-layer.5": 10187.1504, "encoder_q-layer.6": 11380.209, "encoder_q-layer.7": 13074.8115, "encoder_q-layer.8": 14613.4043, "encoder_q-layer.9": 13420.8779, "epoch": 0.56, "inbatch_neg_score": 3.4676, "inbatch_pos_score": 4.0586, "learning_rate": 1.8605263157894736e-05, "loss": 3.3926, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.7112, "num_token_overlap": 11.6867, "num_token_query": 31.4128, "num_token_union": 65.1111, "num_word_context": 202.4441, "num_word_doc": 49.7992, "num_word_query": 23.3328, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21379.6787, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4668, "query_norm": 2.2321, "queue_k_norm": 2.3016, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4128, "sent_len_1": 66.7112, "sent_len_max_0": 127.97, "sent_len_max_1": 207.8887, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 129300 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3839, "doc_norm": 2.3037, "encoder_q-embeddings": 9741.0264, "encoder_q-layer.0": 6851.707, "encoder_q-layer.1": 7112.2021, "encoder_q-layer.10": 14628.9746, "encoder_q-layer.11": 35624.0547, "encoder_q-layer.2": 8142.6221, "encoder_q-layer.3": 8667.4082, "encoder_q-layer.4": 9270.4521, "encoder_q-layer.5": 9617.4355, "encoder_q-layer.6": 10751.8096, "encoder_q-layer.7": 12840.2021, "encoder_q-layer.8": 15334.5586, "encoder_q-layer.9": 14162.3232, "epoch": 0.56, "inbatch_neg_score": 3.4699, "inbatch_pos_score": 4.0781, "learning_rate": 1.8578947368421054e-05, "loss": 3.3839, "norm_diff": 0.0701, "norm_loss": 0.0, "num_token_doc": 66.8419, "num_token_overlap": 11.6994, "num_token_query": 31.4524, "num_token_union": 65.1847, "num_word_context": 202.1472, "num_word_doc": 49.844, "num_word_query": 23.3668, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22057.6565, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4707, "query_norm": 2.2336, "queue_k_norm": 2.302, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4524, "sent_len_1": 66.8419, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.2388, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129400 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.3848, "doc_norm": 2.3028, "encoder_q-embeddings": 10499.0205, "encoder_q-layer.0": 7264.8008, "encoder_q-layer.1": 7843.5435, "encoder_q-layer.10": 13771.3643, "encoder_q-layer.11": 32763.3457, "encoder_q-layer.2": 8745.2998, "encoder_q-layer.3": 9172.2061, "encoder_q-layer.4": 9923.2354, "encoder_q-layer.5": 10066.9551, "encoder_q-layer.6": 11086.5381, "encoder_q-layer.7": 12443.5527, "encoder_q-layer.8": 13958.5674, "encoder_q-layer.9": 12617.248, "epoch": 0.56, "inbatch_neg_score": 3.4752, "inbatch_pos_score": 4.0703, "learning_rate": 1.855263157894737e-05, "loss": 3.3848, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.656, "num_token_overlap": 11.6692, "num_token_query": 31.3919, "num_token_union": 65.0889, "num_word_context": 202.1353, "num_word_doc": 49.7362, "num_word_query": 23.3216, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20460.3856, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4746, "query_norm": 2.2323, "queue_k_norm": 2.3031, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3919, "sent_len_1": 66.656, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.1662, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129500 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3995, "doc_norm": 2.3063, "encoder_q-embeddings": 9978.9238, "encoder_q-layer.0": 7159.7817, "encoder_q-layer.1": 7891.8447, "encoder_q-layer.10": 12444.5293, "encoder_q-layer.11": 31860.3125, "encoder_q-layer.2": 8943.1621, "encoder_q-layer.3": 9006.0273, "encoder_q-layer.4": 9209.8066, "encoder_q-layer.5": 9492.6689, "encoder_q-layer.6": 9970.2168, "encoder_q-layer.7": 10848.8975, "encoder_q-layer.8": 12811.3447, "encoder_q-layer.9": 11839.5947, "epoch": 0.56, "inbatch_neg_score": 3.4779, "inbatch_pos_score": 4.082, "learning_rate": 1.8526315789473687e-05, "loss": 3.3995, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.5614, "num_token_overlap": 11.654, "num_token_query": 31.3457, "num_token_union": 65.0161, "num_word_context": 202.1547, "num_word_doc": 49.6886, "num_word_query": 23.2618, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20054.5578, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4785, "query_norm": 2.2335, "queue_k_norm": 2.3045, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3457, "sent_len_1": 66.5614, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.5325, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129600 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.3905, "doc_norm": 2.3059, "encoder_q-embeddings": 11100.6611, "encoder_q-layer.0": 7787.1587, "encoder_q-layer.1": 8058.9966, "encoder_q-layer.10": 15813.3574, "encoder_q-layer.11": 35632.1914, "encoder_q-layer.2": 9153.9648, "encoder_q-layer.3": 9068.8945, "encoder_q-layer.4": 9562.7832, "encoder_q-layer.5": 9843.8428, "encoder_q-layer.6": 10928.0244, "encoder_q-layer.7": 12026.7422, "encoder_q-layer.8": 14738.7246, "encoder_q-layer.9": 13205.3447, "epoch": 0.56, "inbatch_neg_score": 3.4812, "inbatch_pos_score": 4.0703, "learning_rate": 1.85e-05, "loss": 3.3905, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.8996, "num_token_overlap": 11.7437, "num_token_query": 31.5173, "num_token_union": 65.2066, "num_word_context": 202.4675, "num_word_doc": 49.9307, "num_word_query": 23.406, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21925.8748, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4805, "query_norm": 2.2326, "queue_k_norm": 2.3065, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5173, "sent_len_1": 66.8996, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 129700 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.3982, "doc_norm": 2.3036, "encoder_q-embeddings": 9515.6172, "encoder_q-layer.0": 6758.2979, "encoder_q-layer.1": 7061.96, "encoder_q-layer.10": 13464.7686, "encoder_q-layer.11": 31064.7441, "encoder_q-layer.2": 7985.957, "encoder_q-layer.3": 8248.0811, "encoder_q-layer.4": 8980.4189, "encoder_q-layer.5": 9153.8789, "encoder_q-layer.6": 10208.3604, "encoder_q-layer.7": 11987.0146, "encoder_q-layer.8": 13405.0205, "encoder_q-layer.9": 13137.9873, "epoch": 0.56, "inbatch_neg_score": 3.4857, "inbatch_pos_score": 4.082, "learning_rate": 1.8473684210526317e-05, "loss": 3.3982, "norm_diff": 0.0676, "norm_loss": 0.0, "num_token_doc": 66.6159, "num_token_overlap": 11.6394, "num_token_query": 31.2372, "num_token_union": 64.9814, "num_word_context": 202.0805, "num_word_doc": 49.7249, "num_word_query": 23.1874, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19844.919, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4844, "query_norm": 2.2361, "queue_k_norm": 2.3056, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2372, "sent_len_1": 66.6159, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.0762, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129800 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3948, "doc_norm": 2.3065, "encoder_q-embeddings": 9133.873, "encoder_q-layer.0": 6262.8755, "encoder_q-layer.1": 6853.8218, "encoder_q-layer.10": 12831.3662, "encoder_q-layer.11": 31653.7422, "encoder_q-layer.2": 7720.5479, "encoder_q-layer.3": 8094.1709, "encoder_q-layer.4": 8463.1074, "encoder_q-layer.5": 8571.9639, "encoder_q-layer.6": 9648.0771, "encoder_q-layer.7": 11432.6016, "encoder_q-layer.8": 14146.3691, "encoder_q-layer.9": 13047.8584, "epoch": 0.56, "inbatch_neg_score": 3.4873, "inbatch_pos_score": 4.1055, "learning_rate": 1.8447368421052632e-05, "loss": 3.3948, "norm_diff": 0.0677, "norm_loss": 0.0, "num_token_doc": 66.7328, "num_token_overlap": 11.6877, "num_token_query": 31.3382, "num_token_union": 65.066, "num_word_context": 202.2275, "num_word_doc": 49.7932, "num_word_query": 23.2676, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19738.7548, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4863, "query_norm": 2.2388, "queue_k_norm": 2.3071, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3382, "sent_len_1": 66.7328, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.845, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 129900 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.4027, "doc_norm": 2.3104, "encoder_q-embeddings": 10812.3965, "encoder_q-layer.0": 7330.5835, "encoder_q-layer.1": 8133.623, "encoder_q-layer.10": 13648.459, "encoder_q-layer.11": 30588.7598, "encoder_q-layer.2": 8971.5449, "encoder_q-layer.3": 9443.0332, "encoder_q-layer.4": 9964.6641, "encoder_q-layer.5": 10205.2832, "encoder_q-layer.6": 10991.6689, "encoder_q-layer.7": 13005.2852, "encoder_q-layer.8": 14166.7188, "encoder_q-layer.9": 12383.6133, "epoch": 0.56, "inbatch_neg_score": 3.4909, "inbatch_pos_score": 4.0938, "learning_rate": 1.8421052631578947e-05, "loss": 3.4027, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.6481, "num_token_overlap": 11.6231, "num_token_query": 31.2701, "num_token_union": 65.0231, "num_word_context": 202.0479, "num_word_doc": 49.777, "num_word_query": 23.2202, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20530.8336, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4902, "query_norm": 2.2347, "queue_k_norm": 2.308, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2701, "sent_len_1": 66.6481, "sent_len_max_0": 127.995, "sent_len_max_1": 207.41, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130000 }, { "dev_runtime": 28.2411, "dev_samples_per_second": 1.133, "dev_steps_per_second": 0.035, "epoch": 0.56, "step": 130000, "test_accuracy": 94.2626953125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.33120325207710266, "test_doc_norm": 2.288938045501709, "test_inbatch_neg_score": 3.7773141860961914, "test_inbatch_pos_score": 4.800678253173828, "test_loss": 0.33120325207710266, "test_loss_align": 1.0418250560760498, "test_loss_unif": -19.427753448486328, "test_loss_unif_q@queue": -19.427753448486328, "test_norm_diff": 0.0330561101436615, "test_norm_loss": 0.0, "test_q@queue_neg_score": 3.486598491668701, "test_query_norm": 2.3219940662384033, "test_queue_k_norm": 2.3082642555236816, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044063322246074677, "test_stdq": 0.044176824390888214, "test_stdqueue_k": 0.04943143576383591, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.2411, "dev_samples_per_second": 1.133, "dev_steps_per_second": 0.035, "epoch": 0.56, "eval_beir-arguana_ndcg@10": 0.40749, "eval_beir-arguana_recall@10": 0.68492, "eval_beir-arguana_recall@100": 0.94737, "eval_beir-arguana_recall@20": 0.81721, "eval_beir-avg_ndcg@10": 0.4049155000000001, "eval_beir-avg_recall@10": 0.4735085, "eval_beir-avg_recall@100": 0.6515946666666667, "eval_beir-avg_recall@20": 0.5353600000000001, "eval_beir-cqadupstack_ndcg@10": 0.296505, "eval_beir-cqadupstack_recall@10": 0.3956549999999999, "eval_beir-cqadupstack_recall@100": 0.6316466666666667, "eval_beir-cqadupstack_recall@20": 0.4645700000000001, "eval_beir-fiqa_ndcg@10": 0.26687, "eval_beir-fiqa_recall@10": 0.32936, "eval_beir-fiqa_recall@100": 0.61644, "eval_beir-fiqa_recall@20": 0.41446, "eval_beir-nfcorpus_ndcg@10": 0.32229, "eval_beir-nfcorpus_recall@10": 0.15866, "eval_beir-nfcorpus_recall@100": 0.30881, "eval_beir-nfcorpus_recall@20": 0.19714, "eval_beir-nq_ndcg@10": 0.32597, "eval_beir-nq_recall@10": 0.51811, "eval_beir-nq_recall@100": 0.84031, "eval_beir-nq_recall@20": 0.63195, "eval_beir-quora_ndcg@10": 0.78661, "eval_beir-quora_recall@10": 0.8946, "eval_beir-quora_recall@100": 0.97978, "eval_beir-quora_recall@20": 0.93319, "eval_beir-scidocs_ndcg@10": 0.16402, "eval_beir-scidocs_recall@10": 0.17157, "eval_beir-scidocs_recall@100": 0.384, "eval_beir-scidocs_recall@20": 0.23217, "eval_beir-scifact_ndcg@10": 0.67765, "eval_beir-scifact_recall@10": 0.79133, "eval_beir-scifact_recall@100": 0.92156, "eval_beir-scifact_recall@20": 0.84411, "eval_beir-trec-covid_ndcg@10": 0.61533, "eval_beir-trec-covid_recall@10": 0.658, "eval_beir-trec-covid_recall@100": 0.4624, "eval_beir-trec-covid_recall@20": 0.611, "eval_beir-webis-touche2020_ndcg@10": 0.18642, "eval_beir-webis-touche2020_recall@10": 0.13288, "eval_beir-webis-touche2020_recall@100": 0.42363, "eval_beir-webis-touche2020_recall@20": 0.2078, "eval_senteval-avg_sts": 0.7372643028235695, "eval_senteval-sickr_spearman": 0.734060235869603, "eval_senteval-stsb_spearman": 0.7404683697775358, "step": 130000, "test_accuracy": 94.2626953125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.33120325207710266, "test_doc_norm": 2.288938045501709, "test_inbatch_neg_score": 3.7773141860961914, "test_inbatch_pos_score": 4.800678253173828, "test_loss": 0.33120325207710266, "test_loss_align": 1.0418250560760498, "test_loss_unif": -19.427753448486328, "test_loss_unif_q@queue": -19.427753448486328, "test_norm_diff": 0.0330561101436615, "test_norm_loss": 0.0, "test_q@queue_neg_score": 3.486598491668701, "test_query_norm": 2.3219940662384033, "test_queue_k_norm": 2.3082642555236816, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044063322246074677, "test_stdq": 0.044176824390888214, "test_stdqueue_k": 0.04943143576383591, "test_stdqueue_q": 0.0 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3935, "doc_norm": 2.3084, "encoder_q-embeddings": 9942.6973, "encoder_q-layer.0": 7162.8896, "encoder_q-layer.1": 7373.6245, "encoder_q-layer.10": 12580.4824, "encoder_q-layer.11": 31228.2539, "encoder_q-layer.2": 8187.4678, "encoder_q-layer.3": 8514.5439, "encoder_q-layer.4": 9295.7773, "encoder_q-layer.5": 9202.5752, "encoder_q-layer.6": 10392.7617, "encoder_q-layer.7": 11060.7852, "encoder_q-layer.8": 13204.6797, "encoder_q-layer.9": 11993.165, "epoch": 0.56, "inbatch_neg_score": 3.4935, "inbatch_pos_score": 4.0977, "learning_rate": 1.8394736842105266e-05, "loss": 3.3935, "norm_diff": 0.0682, "norm_loss": 0.0, "num_token_doc": 66.7388, "num_token_overlap": 11.6672, "num_token_query": 31.3709, "num_token_union": 65.1393, "num_word_context": 201.9375, "num_word_doc": 49.8076, "num_word_query": 23.3009, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19631.3861, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.4941, "query_norm": 2.2401, "queue_k_norm": 2.3091, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3709, "sent_len_1": 66.7388, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.52, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130100 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3893, "doc_norm": 2.3129, "encoder_q-embeddings": 19756.916, "encoder_q-layer.0": 14987.2461, "encoder_q-layer.1": 16161.8535, "encoder_q-layer.10": 13208.9453, "encoder_q-layer.11": 31247.5684, "encoder_q-layer.2": 18020.4434, "encoder_q-layer.3": 19604.6855, "encoder_q-layer.4": 22351.7578, "encoder_q-layer.5": 20499.6016, "encoder_q-layer.6": 22991.8379, "encoder_q-layer.7": 20385.7891, "encoder_q-layer.8": 17939.8477, "encoder_q-layer.9": 13106.4082, "epoch": 0.56, "inbatch_neg_score": 3.4985, "inbatch_pos_score": 4.1016, "learning_rate": 1.836842105263158e-05, "loss": 3.3893, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.6145, "num_token_overlap": 11.638, "num_token_query": 31.2798, "num_token_union": 64.9945, "num_word_context": 202.0641, "num_word_doc": 49.7095, "num_word_query": 23.2263, "postclip_grad_norm": 1.0, "preclip_grad_norm": 29494.9253, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 3.498, "query_norm": 2.2396, "queue_k_norm": 2.31, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2798, "sent_len_1": 66.6145, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.9363, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130200 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.3933, "doc_norm": 2.3103, "encoder_q-embeddings": 10248.2686, "encoder_q-layer.0": 6918.769, "encoder_q-layer.1": 7419.3022, "encoder_q-layer.10": 14483.9199, "encoder_q-layer.11": 32146.5625, "encoder_q-layer.2": 8271.8994, "encoder_q-layer.3": 8671.4639, "encoder_q-layer.4": 9093.5967, "encoder_q-layer.5": 9599.7627, "encoder_q-layer.6": 10699.0322, "encoder_q-layer.7": 11692.5225, "encoder_q-layer.8": 14301.1201, "encoder_q-layer.9": 12507.9502, "epoch": 0.57, "inbatch_neg_score": 3.5079, "inbatch_pos_score": 4.0938, "learning_rate": 1.8342105263157896e-05, "loss": 3.3933, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.9025, "num_token_overlap": 11.6672, "num_token_query": 31.3181, "num_token_union": 65.1704, "num_word_context": 202.4199, "num_word_doc": 49.9217, "num_word_query": 23.2389, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20307.9546, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.5078, "query_norm": 2.2434, "queue_k_norm": 2.3113, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3181, "sent_len_1": 66.9025, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.07, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 130300 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3876, "doc_norm": 2.3153, "encoder_q-embeddings": 10585.3867, "encoder_q-layer.0": 7424.7803, "encoder_q-layer.1": 8398.8701, "encoder_q-layer.10": 13071.7168, "encoder_q-layer.11": 31563.2344, "encoder_q-layer.2": 9593.6055, "encoder_q-layer.3": 9836.6162, "encoder_q-layer.4": 10412.8408, "encoder_q-layer.5": 10545.4131, "encoder_q-layer.6": 11803.4443, "encoder_q-layer.7": 12774.0918, "encoder_q-layer.8": 14792.4941, "encoder_q-layer.9": 12976.9277, "epoch": 0.57, "inbatch_neg_score": 3.5106, "inbatch_pos_score": 4.125, "learning_rate": 1.831578947368421e-05, "loss": 3.3876, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.8311, "num_token_overlap": 11.6866, "num_token_query": 31.3635, "num_token_union": 65.1251, "num_word_context": 202.2932, "num_word_doc": 49.8676, "num_word_query": 23.31, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21201.0863, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.5098, "query_norm": 2.2493, "queue_k_norm": 2.3119, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3635, "sent_len_1": 66.8311, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.1075, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130400 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3908, "doc_norm": 2.3113, "encoder_q-embeddings": 4930.4561, "encoder_q-layer.0": 3500.978, "encoder_q-layer.1": 3663.7546, "encoder_q-layer.10": 6792.1851, "encoder_q-layer.11": 16172.6758, "encoder_q-layer.2": 4083.8076, "encoder_q-layer.3": 4255.7593, "encoder_q-layer.4": 4515.3428, "encoder_q-layer.5": 4482.9629, "encoder_q-layer.6": 4978.8169, "encoder_q-layer.7": 5558.9473, "encoder_q-layer.8": 6732.0156, "encoder_q-layer.9": 6386.1714, "epoch": 0.57, "inbatch_neg_score": 3.522, "inbatch_pos_score": 4.1211, "learning_rate": 1.828947368421053e-05, "loss": 3.3908, "norm_diff": 0.0637, "norm_loss": 0.0, "num_token_doc": 66.83, "num_token_overlap": 11.6911, "num_token_query": 31.4138, "num_token_union": 65.1659, "num_word_context": 202.2563, "num_word_doc": 49.8485, "num_word_query": 23.3238, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10185.5393, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5215, "query_norm": 2.2477, "queue_k_norm": 2.3127, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4138, "sent_len_1": 66.83, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.42, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130500 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3964, "doc_norm": 2.3149, "encoder_q-embeddings": 5095.8359, "encoder_q-layer.0": 3608.1963, "encoder_q-layer.1": 3755.5249, "encoder_q-layer.10": 6130.3198, "encoder_q-layer.11": 15801.4639, "encoder_q-layer.2": 4168.877, "encoder_q-layer.3": 4446.1758, "encoder_q-layer.4": 4717.4307, "encoder_q-layer.5": 4853.0098, "encoder_q-layer.6": 5433.4399, "encoder_q-layer.7": 5756.207, "encoder_q-layer.8": 6492.2305, "encoder_q-layer.9": 6008.4995, "epoch": 0.57, "inbatch_neg_score": 3.5275, "inbatch_pos_score": 4.1484, "learning_rate": 1.8263157894736844e-05, "loss": 3.3964, "norm_diff": 0.0553, "norm_loss": 0.0, "num_token_doc": 66.8162, "num_token_overlap": 11.6884, "num_token_query": 31.4714, "num_token_union": 65.1842, "num_word_context": 202.4244, "num_word_doc": 49.8667, "num_word_query": 23.3894, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10150.4604, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5273, "query_norm": 2.2596, "queue_k_norm": 2.3127, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4714, "sent_len_1": 66.8162, "sent_len_max_0": 128.0, "sent_len_max_1": 209.4975, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0493, "stdqueue_q": 0.0, "step": 130600 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3955, "doc_norm": 2.3155, "encoder_q-embeddings": 4507.9395, "encoder_q-layer.0": 3160.3508, "encoder_q-layer.1": 3450.9543, "encoder_q-layer.10": 6097.2603, "encoder_q-layer.11": 15432.335, "encoder_q-layer.2": 3829.9771, "encoder_q-layer.3": 3910.9968, "encoder_q-layer.4": 4321.3931, "encoder_q-layer.5": 4316.0796, "encoder_q-layer.6": 5018.1562, "encoder_q-layer.7": 5366.7847, "encoder_q-layer.8": 6299.9819, "encoder_q-layer.9": 5823.2603, "epoch": 0.57, "inbatch_neg_score": 3.5313, "inbatch_pos_score": 4.1484, "learning_rate": 1.823684210526316e-05, "loss": 3.3955, "norm_diff": 0.0591, "norm_loss": 0.0, "num_token_doc": 66.7817, "num_token_overlap": 11.6773, "num_token_query": 31.2963, "num_token_union": 65.0526, "num_word_context": 202.4654, "num_word_doc": 49.8388, "num_word_query": 23.2265, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9612.2751, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5312, "query_norm": 2.2564, "queue_k_norm": 2.3142, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2963, "sent_len_1": 66.7817, "sent_len_max_0": 127.985, "sent_len_max_1": 208.6362, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130700 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3917, "doc_norm": 2.3155, "encoder_q-embeddings": 4923.5908, "encoder_q-layer.0": 3439.1689, "encoder_q-layer.1": 3672.7998, "encoder_q-layer.10": 6964.0781, "encoder_q-layer.11": 16144.1787, "encoder_q-layer.2": 4229.4883, "encoder_q-layer.3": 4346.3813, "encoder_q-layer.4": 4624.1787, "encoder_q-layer.5": 4908.001, "encoder_q-layer.6": 5530.1802, "encoder_q-layer.7": 5852.2271, "encoder_q-layer.8": 6717.3428, "encoder_q-layer.9": 6490.2222, "epoch": 0.57, "inbatch_neg_score": 3.5352, "inbatch_pos_score": 4.1523, "learning_rate": 1.8210526315789474e-05, "loss": 3.3917, "norm_diff": 0.0625, "norm_loss": 0.0, "num_token_doc": 66.815, "num_token_overlap": 11.6691, "num_token_query": 31.3642, "num_token_union": 65.15, "num_word_context": 202.2256, "num_word_doc": 49.8374, "num_word_query": 23.2807, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10161.4149, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5352, "query_norm": 2.253, "queue_k_norm": 2.3158, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3642, "sent_len_1": 66.815, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.3413, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130800 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3788, "doc_norm": 2.3166, "encoder_q-embeddings": 4836.9395, "encoder_q-layer.0": 3335.2012, "encoder_q-layer.1": 3631.0801, "encoder_q-layer.10": 6646.1396, "encoder_q-layer.11": 15835.8701, "encoder_q-layer.2": 4017.7378, "encoder_q-layer.3": 4079.0957, "encoder_q-layer.4": 4415.6597, "encoder_q-layer.5": 4405.7246, "encoder_q-layer.6": 4931.8247, "encoder_q-layer.7": 5764.6152, "encoder_q-layer.8": 6974.7793, "encoder_q-layer.9": 6380.4604, "epoch": 0.57, "inbatch_neg_score": 3.541, "inbatch_pos_score": 4.1484, "learning_rate": 1.818421052631579e-05, "loss": 3.3788, "norm_diff": 0.0598, "norm_loss": 0.0, "num_token_doc": 66.8707, "num_token_overlap": 11.7245, "num_token_query": 31.4189, "num_token_union": 65.1463, "num_word_context": 202.4517, "num_word_doc": 49.8718, "num_word_query": 23.3208, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10009.3118, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.541, "query_norm": 2.2568, "queue_k_norm": 2.3165, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4189, "sent_len_1": 66.8707, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.6662, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 130900 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.387, "doc_norm": 2.3204, "encoder_q-embeddings": 4653.2407, "encoder_q-layer.0": 3246.4934, "encoder_q-layer.1": 3437.4829, "encoder_q-layer.10": 6607.2222, "encoder_q-layer.11": 15479.417, "encoder_q-layer.2": 3876.1223, "encoder_q-layer.3": 4113.3721, "encoder_q-layer.4": 4302.1118, "encoder_q-layer.5": 4597.8154, "encoder_q-layer.6": 5325.9316, "encoder_q-layer.7": 6129.0752, "encoder_q-layer.8": 6939.3101, "encoder_q-layer.9": 6140.667, "epoch": 0.57, "inbatch_neg_score": 3.5496, "inbatch_pos_score": 4.1719, "learning_rate": 1.8157894736842107e-05, "loss": 3.387, "norm_diff": 0.0513, "norm_loss": 0.0, "num_token_doc": 66.6999, "num_token_overlap": 11.6835, "num_token_query": 31.4462, "num_token_union": 65.1089, "num_word_context": 202.3883, "num_word_doc": 49.7693, "num_word_query": 23.3609, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9817.1171, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5508, "query_norm": 2.2692, "queue_k_norm": 2.3185, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4462, "sent_len_1": 66.6999, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.5737, "stdk": 0.0495, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131000 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.3923, "doc_norm": 2.3167, "encoder_q-embeddings": 4859.168, "encoder_q-layer.0": 3410.4509, "encoder_q-layer.1": 3737.4429, "encoder_q-layer.10": 6904.582, "encoder_q-layer.11": 16432.6094, "encoder_q-layer.2": 4234.9043, "encoder_q-layer.3": 4469.2471, "encoder_q-layer.4": 4674.7026, "encoder_q-layer.5": 4972.7104, "encoder_q-layer.6": 5577.0854, "encoder_q-layer.7": 6197.0801, "encoder_q-layer.8": 7355.002, "encoder_q-layer.9": 6607.1958, "epoch": 0.57, "inbatch_neg_score": 3.5584, "inbatch_pos_score": 4.1562, "learning_rate": 1.8131578947368422e-05, "loss": 3.3923, "norm_diff": 0.0502, "norm_loss": 0.0, "num_token_doc": 66.723, "num_token_overlap": 11.6898, "num_token_query": 31.4395, "num_token_union": 65.1302, "num_word_context": 202.2218, "num_word_doc": 49.8079, "num_word_query": 23.3611, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10443.3701, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5586, "query_norm": 2.2665, "queue_k_norm": 2.3189, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4395, "sent_len_1": 66.723, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.47, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131100 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3977, "doc_norm": 2.3188, "encoder_q-embeddings": 8266.4795, "encoder_q-layer.0": 6273.6934, "encoder_q-layer.1": 7519.2847, "encoder_q-layer.10": 6490.0249, "encoder_q-layer.11": 15697.1201, "encoder_q-layer.2": 8745.6719, "encoder_q-layer.3": 9670.7354, "encoder_q-layer.4": 10328.0566, "encoder_q-layer.5": 11014.6846, "encoder_q-layer.6": 11903.7344, "encoder_q-layer.7": 13775.6592, "encoder_q-layer.8": 10979.7705, "encoder_q-layer.9": 6357.6201, "epoch": 0.57, "inbatch_neg_score": 3.5645, "inbatch_pos_score": 4.1641, "learning_rate": 1.810526315789474e-05, "loss": 3.3977, "norm_diff": 0.0466, "norm_loss": 0.0, "num_token_doc": 66.6242, "num_token_overlap": 11.6373, "num_token_query": 31.3062, "num_token_union": 64.9871, "num_word_context": 201.9622, "num_word_doc": 49.7039, "num_word_query": 23.2447, "postclip_grad_norm": 1.0, "preclip_grad_norm": 15121.4638, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5645, "query_norm": 2.2722, "queue_k_norm": 2.3218, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3062, "sent_len_1": 66.6242, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.9538, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 131200 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.394, "doc_norm": 2.3203, "encoder_q-embeddings": 4682.6758, "encoder_q-layer.0": 3255.4766, "encoder_q-layer.1": 3461.1479, "encoder_q-layer.10": 6362.6021, "encoder_q-layer.11": 16298.8936, "encoder_q-layer.2": 3896.0342, "encoder_q-layer.3": 4007.9587, "encoder_q-layer.4": 4268.2451, "encoder_q-layer.5": 4320.771, "encoder_q-layer.6": 5142.7642, "encoder_q-layer.7": 5616.165, "encoder_q-layer.8": 6859.0376, "encoder_q-layer.9": 6261.5693, "epoch": 0.57, "inbatch_neg_score": 3.5756, "inbatch_pos_score": 4.1797, "learning_rate": 1.8078947368421052e-05, "loss": 3.394, "norm_diff": 0.049, "norm_loss": 0.0, "num_token_doc": 66.954, "num_token_overlap": 11.6651, "num_token_query": 31.3119, "num_token_union": 65.1939, "num_word_context": 202.5039, "num_word_doc": 49.9406, "num_word_query": 23.2584, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10164.6677, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5762, "query_norm": 2.2712, "queue_k_norm": 2.3217, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3119, "sent_len_1": 66.954, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.0888, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131300 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3866, "doc_norm": 2.3227, "encoder_q-embeddings": 4747.2983, "encoder_q-layer.0": 3218.5391, "encoder_q-layer.1": 3440.864, "encoder_q-layer.10": 6429.1152, "encoder_q-layer.11": 15298.2148, "encoder_q-layer.2": 3938.292, "encoder_q-layer.3": 4080.7434, "encoder_q-layer.4": 4385.9209, "encoder_q-layer.5": 4490.6631, "encoder_q-layer.6": 5007.7803, "encoder_q-layer.7": 5824.7798, "encoder_q-layer.8": 6895.2656, "encoder_q-layer.9": 6397.0122, "epoch": 0.57, "inbatch_neg_score": 3.5846, "inbatch_pos_score": 4.1992, "learning_rate": 1.8052631578947367e-05, "loss": 3.3866, "norm_diff": 0.047, "norm_loss": 0.0, "num_token_doc": 66.8669, "num_token_overlap": 11.7027, "num_token_query": 31.4431, "num_token_union": 65.1992, "num_word_context": 202.4039, "num_word_doc": 49.8973, "num_word_query": 23.3585, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9730.8961, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.584, "query_norm": 2.2758, "queue_k_norm": 2.3236, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4431, "sent_len_1": 66.8669, "sent_len_max_0": 127.995, "sent_len_max_1": 209.1025, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131400 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3886, "doc_norm": 2.3271, "encoder_q-embeddings": 4968.4219, "encoder_q-layer.0": 3411.3774, "encoder_q-layer.1": 3627.3833, "encoder_q-layer.10": 6519.8184, "encoder_q-layer.11": 15829.293, "encoder_q-layer.2": 4067.3071, "encoder_q-layer.3": 4169.4814, "encoder_q-layer.4": 4607.5435, "encoder_q-layer.5": 4973.6729, "encoder_q-layer.6": 5191.8945, "encoder_q-layer.7": 6124.332, "encoder_q-layer.8": 7085.3276, "encoder_q-layer.9": 6518.811, "epoch": 0.57, "inbatch_neg_score": 3.5985, "inbatch_pos_score": 4.1992, "learning_rate": 1.8026315789473685e-05, "loss": 3.3886, "norm_diff": 0.0525, "norm_loss": 0.0, "num_token_doc": 66.7579, "num_token_overlap": 11.685, "num_token_query": 31.3056, "num_token_union": 65.0471, "num_word_context": 202.4381, "num_word_doc": 49.8364, "num_word_query": 23.2628, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10216.7924, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.5977, "query_norm": 2.2747, "queue_k_norm": 2.3252, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3056, "sent_len_1": 66.7579, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.7512, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131500 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3887, "doc_norm": 2.3291, "encoder_q-embeddings": 4908.0317, "encoder_q-layer.0": 3459.4341, "encoder_q-layer.1": 3691.2632, "encoder_q-layer.10": 6573.6396, "encoder_q-layer.11": 15994.7324, "encoder_q-layer.2": 4238.5088, "encoder_q-layer.3": 4478.5278, "encoder_q-layer.4": 4849.54, "encoder_q-layer.5": 5070.9297, "encoder_q-layer.6": 5497.2173, "encoder_q-layer.7": 5917.0786, "encoder_q-layer.8": 6868.5176, "encoder_q-layer.9": 6210.7427, "epoch": 0.57, "inbatch_neg_score": 3.61, "inbatch_pos_score": 4.2344, "learning_rate": 1.8e-05, "loss": 3.3887, "norm_diff": 0.0427, "norm_loss": 0.0, "num_token_doc": 66.9267, "num_token_overlap": 11.6836, "num_token_query": 31.3794, "num_token_union": 65.2146, "num_word_context": 202.3531, "num_word_doc": 49.9521, "num_word_query": 23.2904, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10398.8731, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6094, "query_norm": 2.2864, "queue_k_norm": 2.3268, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3794, "sent_len_1": 66.9267, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.5662, "stdk": 0.0496, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 131600 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.3968, "doc_norm": 2.3308, "encoder_q-embeddings": 4827.1035, "encoder_q-layer.0": 3346.6174, "encoder_q-layer.1": 3650.0232, "encoder_q-layer.10": 6493.292, "encoder_q-layer.11": 15844.3809, "encoder_q-layer.2": 4233.2773, "encoder_q-layer.3": 4262.8018, "encoder_q-layer.4": 4638.6572, "encoder_q-layer.5": 4792.9282, "encoder_q-layer.6": 5323.127, "encoder_q-layer.7": 5753.3691, "encoder_q-layer.8": 6887.7886, "encoder_q-layer.9": 6303.0513, "epoch": 0.57, "inbatch_neg_score": 3.6195, "inbatch_pos_score": 4.2227, "learning_rate": 1.797368421052632e-05, "loss": 3.3968, "norm_diff": 0.0581, "norm_loss": 0.0, "num_token_doc": 66.7882, "num_token_overlap": 11.6722, "num_token_query": 31.3606, "num_token_union": 65.1069, "num_word_context": 202.1566, "num_word_doc": 49.8346, "num_word_query": 23.2873, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10087.9298, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6191, "query_norm": 2.2727, "queue_k_norm": 2.3281, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3606, "sent_len_1": 66.7882, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.8137, "stdk": 0.0495, "stdq": 0.0426, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131700 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.4067, "doc_norm": 2.3285, "encoder_q-embeddings": 5002.936, "encoder_q-layer.0": 3504.6753, "encoder_q-layer.1": 3713.0601, "encoder_q-layer.10": 6436.5464, "encoder_q-layer.11": 16729.2227, "encoder_q-layer.2": 4103.3379, "encoder_q-layer.3": 4218.4409, "encoder_q-layer.4": 4361.645, "encoder_q-layer.5": 4494.0566, "encoder_q-layer.6": 5038.439, "encoder_q-layer.7": 5880.4854, "encoder_q-layer.8": 7044.6309, "encoder_q-layer.9": 6336.2612, "epoch": 0.57, "inbatch_neg_score": 3.6274, "inbatch_pos_score": 4.2344, "learning_rate": 1.7947368421052634e-05, "loss": 3.4067, "norm_diff": 0.0457, "norm_loss": 0.0, "num_token_doc": 66.8133, "num_token_overlap": 11.6334, "num_token_query": 31.2808, "num_token_union": 65.1365, "num_word_context": 202.6585, "num_word_doc": 49.8487, "num_word_query": 23.2298, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10214.8182, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.627, "query_norm": 2.2828, "queue_k_norm": 2.3292, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2808, "sent_len_1": 66.8133, "sent_len_max_0": 127.9963, "sent_len_max_1": 211.4062, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131800 }, { "accuracy": 53.5645, "active_queue_size": 16384.0, "cl_loss": 3.3915, "doc_norm": 2.3311, "encoder_q-embeddings": 4882.2368, "encoder_q-layer.0": 3421.4363, "encoder_q-layer.1": 3717.8291, "encoder_q-layer.10": 6762.5098, "encoder_q-layer.11": 16127.9414, "encoder_q-layer.2": 4297.8838, "encoder_q-layer.3": 4340.2607, "encoder_q-layer.4": 4561.3184, "encoder_q-layer.5": 4708.6719, "encoder_q-layer.6": 5126.3765, "encoder_q-layer.7": 5810.3706, "encoder_q-layer.8": 6423.1001, "encoder_q-layer.9": 6075.709, "epoch": 0.57, "inbatch_neg_score": 3.6355, "inbatch_pos_score": 4.2578, "learning_rate": 1.792105263157895e-05, "loss": 3.3915, "norm_diff": 0.0516, "norm_loss": 0.0, "num_token_doc": 66.8523, "num_token_overlap": 11.6945, "num_token_query": 31.3615, "num_token_union": 65.119, "num_word_context": 202.1642, "num_word_doc": 49.8649, "num_word_query": 23.2684, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9976.4009, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6367, "query_norm": 2.2796, "queue_k_norm": 2.3307, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3615, "sent_len_1": 66.8523, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.1, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 131900 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3723, "doc_norm": 2.3323, "encoder_q-embeddings": 5025.188, "encoder_q-layer.0": 3512.3164, "encoder_q-layer.1": 3675.8855, "encoder_q-layer.10": 7222.5757, "encoder_q-layer.11": 16297.0059, "encoder_q-layer.2": 4028.2288, "encoder_q-layer.3": 4189.4795, "encoder_q-layer.4": 4495.4175, "encoder_q-layer.5": 4789.4683, "encoder_q-layer.6": 5346.7993, "encoder_q-layer.7": 5704.3379, "encoder_q-layer.8": 6973.2002, "encoder_q-layer.9": 6775.6157, "epoch": 0.57, "inbatch_neg_score": 3.6402, "inbatch_pos_score": 4.2539, "learning_rate": 1.7894736842105264e-05, "loss": 3.3723, "norm_diff": 0.0477, "norm_loss": 0.0, "num_token_doc": 66.7849, "num_token_overlap": 11.7122, "num_token_query": 31.4727, "num_token_union": 65.1732, "num_word_context": 202.3886, "num_word_doc": 49.8286, "num_word_query": 23.3904, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10171.3607, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6406, "query_norm": 2.2846, "queue_k_norm": 2.3336, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4727, "sent_len_1": 66.7849, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.6225, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 132000 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.4054, "doc_norm": 2.3333, "encoder_q-embeddings": 4782.8125, "encoder_q-layer.0": 3287.168, "encoder_q-layer.1": 3505.8696, "encoder_q-layer.10": 6344.269, "encoder_q-layer.11": 15148.3789, "encoder_q-layer.2": 3849.6675, "encoder_q-layer.3": 3935.4021, "encoder_q-layer.4": 4201.3428, "encoder_q-layer.5": 4373.4888, "encoder_q-layer.6": 5066.1792, "encoder_q-layer.7": 5645.6196, "encoder_q-layer.8": 6530.4541, "encoder_q-layer.9": 6386.1064, "epoch": 0.57, "inbatch_neg_score": 3.6472, "inbatch_pos_score": 4.2461, "learning_rate": 1.786842105263158e-05, "loss": 3.4054, "norm_diff": 0.0582, "norm_loss": 0.0, "num_token_doc": 66.6085, "num_token_overlap": 11.6147, "num_token_query": 31.2133, "num_token_union": 64.9746, "num_word_context": 202.1061, "num_word_doc": 49.6763, "num_word_query": 23.1692, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9675.6529, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6465, "query_norm": 2.2752, "queue_k_norm": 2.3346, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2133, "sent_len_1": 66.6085, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.3262, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132100 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3845, "doc_norm": 2.3354, "encoder_q-embeddings": 5068.4639, "encoder_q-layer.0": 3412.4509, "encoder_q-layer.1": 3703.1455, "encoder_q-layer.10": 6922.397, "encoder_q-layer.11": 16481.0195, "encoder_q-layer.2": 4227.7759, "encoder_q-layer.3": 4406.8652, "encoder_q-layer.4": 4617.1743, "encoder_q-layer.5": 4760.9229, "encoder_q-layer.6": 5379.7144, "encoder_q-layer.7": 6140.8062, "encoder_q-layer.8": 7551.6353, "encoder_q-layer.9": 6655.4702, "epoch": 0.57, "inbatch_neg_score": 3.6492, "inbatch_pos_score": 4.2539, "learning_rate": 1.7842105263157897e-05, "loss": 3.3845, "norm_diff": 0.0607, "norm_loss": 0.0, "num_token_doc": 66.6917, "num_token_overlap": 11.6473, "num_token_query": 31.3555, "num_token_union": 65.1031, "num_word_context": 202.1596, "num_word_doc": 49.7611, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10500.7482, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6504, "query_norm": 2.2748, "queue_k_norm": 2.3364, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3555, "sent_len_1": 66.6917, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.6275, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132200 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3866, "doc_norm": 2.338, "encoder_q-embeddings": 4671.2451, "encoder_q-layer.0": 3146.6125, "encoder_q-layer.1": 3435.7454, "encoder_q-layer.10": 7193.7744, "encoder_q-layer.11": 17455.9141, "encoder_q-layer.2": 3901.8884, "encoder_q-layer.3": 4241.2202, "encoder_q-layer.4": 4553.3169, "encoder_q-layer.5": 4760.7568, "encoder_q-layer.6": 5409.5269, "encoder_q-layer.7": 6303.8779, "encoder_q-layer.8": 7476.1899, "encoder_q-layer.9": 6709.2651, "epoch": 0.57, "inbatch_neg_score": 3.6541, "inbatch_pos_score": 4.2695, "learning_rate": 1.7815789473684212e-05, "loss": 3.3866, "norm_diff": 0.0604, "norm_loss": 0.0, "num_token_doc": 66.6873, "num_token_overlap": 11.6846, "num_token_query": 31.3721, "num_token_union": 65.082, "num_word_context": 202.1139, "num_word_doc": 49.7493, "num_word_query": 23.2958, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10519.053, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6543, "query_norm": 2.2776, "queue_k_norm": 2.3378, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3721, "sent_len_1": 66.6873, "sent_len_max_0": 127.985, "sent_len_max_1": 209.855, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132300 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3998, "doc_norm": 2.3388, "encoder_q-embeddings": 4809.3633, "encoder_q-layer.0": 3312.6877, "encoder_q-layer.1": 3557.2021, "encoder_q-layer.10": 6301.0537, "encoder_q-layer.11": 15852.5869, "encoder_q-layer.2": 4105.1333, "encoder_q-layer.3": 4257.9097, "encoder_q-layer.4": 4658.0239, "encoder_q-layer.5": 4698.1401, "encoder_q-layer.6": 5120.438, "encoder_q-layer.7": 5724.8154, "encoder_q-layer.8": 6847.2627, "encoder_q-layer.9": 6099.6753, "epoch": 0.57, "inbatch_neg_score": 3.6574, "inbatch_pos_score": 4.2578, "learning_rate": 1.7789473684210527e-05, "loss": 3.3998, "norm_diff": 0.0669, "norm_loss": 0.0, "num_token_doc": 66.6559, "num_token_overlap": 11.6705, "num_token_query": 31.392, "num_token_union": 65.084, "num_word_context": 202.2493, "num_word_doc": 49.7583, "num_word_query": 23.3069, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10170.3772, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6562, "query_norm": 2.2719, "queue_k_norm": 2.3392, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.392, "sent_len_1": 66.6559, "sent_len_max_0": 127.9475, "sent_len_max_1": 208.5175, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132400 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3912, "doc_norm": 2.3405, "encoder_q-embeddings": 10268.584, "encoder_q-layer.0": 6996.6924, "encoder_q-layer.1": 7310.1392, "encoder_q-layer.10": 13179.8203, "encoder_q-layer.11": 32067.1484, "encoder_q-layer.2": 8034.3398, "encoder_q-layer.3": 8403.3184, "encoder_q-layer.4": 8961.168, "encoder_q-layer.5": 9686.2305, "encoder_q-layer.6": 11198.8643, "encoder_q-layer.7": 12363.9629, "encoder_q-layer.8": 13658.5615, "encoder_q-layer.9": 12250.0596, "epoch": 0.57, "inbatch_neg_score": 3.6634, "inbatch_pos_score": 4.2617, "learning_rate": 1.7763157894736842e-05, "loss": 3.3912, "norm_diff": 0.0665, "norm_loss": 0.0, "num_token_doc": 66.8547, "num_token_overlap": 11.6647, "num_token_query": 31.388, "num_token_union": 65.168, "num_word_context": 202.455, "num_word_doc": 49.9001, "num_word_query": 23.3277, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20303.4923, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.6621, "query_norm": 2.274, "queue_k_norm": 2.3403, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.388, "sent_len_1": 66.8547, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.9325, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132500 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3817, "doc_norm": 2.3406, "encoder_q-embeddings": 4799.9409, "encoder_q-layer.0": 3228.5066, "encoder_q-layer.1": 3382.4104, "encoder_q-layer.10": 6260.4165, "encoder_q-layer.11": 16996.9844, "encoder_q-layer.2": 3770.9297, "encoder_q-layer.3": 3980.969, "encoder_q-layer.4": 4221.8682, "encoder_q-layer.5": 4373.7295, "encoder_q-layer.6": 4980.748, "encoder_q-layer.7": 5589.543, "encoder_q-layer.8": 6807.0366, "encoder_q-layer.9": 6360.6621, "epoch": 0.58, "inbatch_neg_score": 3.6689, "inbatch_pos_score": 4.2812, "learning_rate": 1.773684210526316e-05, "loss": 3.3817, "norm_diff": 0.058, "norm_loss": 0.0, "num_token_doc": 66.7176, "num_token_overlap": 11.6595, "num_token_query": 31.3758, "num_token_union": 65.1408, "num_word_context": 202.3757, "num_word_doc": 49.7742, "num_word_query": 23.3179, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10000.7043, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.668, "query_norm": 2.2826, "queue_k_norm": 2.3415, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3758, "sent_len_1": 66.7176, "sent_len_max_0": 127.995, "sent_len_max_1": 209.7763, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132600 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3832, "doc_norm": 2.3467, "encoder_q-embeddings": 4972.5239, "encoder_q-layer.0": 3322.6492, "encoder_q-layer.1": 3505.2361, "encoder_q-layer.10": 6774.6016, "encoder_q-layer.11": 15977.4678, "encoder_q-layer.2": 3885.8474, "encoder_q-layer.3": 4098.8916, "encoder_q-layer.4": 4320.959, "encoder_q-layer.5": 4454.7383, "encoder_q-layer.6": 5041.9604, "encoder_q-layer.7": 6043.5137, "encoder_q-layer.8": 7002.8076, "encoder_q-layer.9": 6236.2251, "epoch": 0.58, "inbatch_neg_score": 3.6756, "inbatch_pos_score": 4.2969, "learning_rate": 1.7710526315789475e-05, "loss": 3.3832, "norm_diff": 0.0604, "norm_loss": 0.0, "num_token_doc": 66.8636, "num_token_overlap": 11.6855, "num_token_query": 31.3784, "num_token_union": 65.1987, "num_word_context": 202.554, "num_word_doc": 49.8725, "num_word_query": 23.305, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10069.5342, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6758, "query_norm": 2.2862, "queue_k_norm": 2.343, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3784, "sent_len_1": 66.8636, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.3963, "stdk": 0.0496, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132700 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3823, "doc_norm": 2.3456, "encoder_q-embeddings": 16486.125, "encoder_q-layer.0": 12756.4629, "encoder_q-layer.1": 11656.79, "encoder_q-layer.10": 6676.6772, "encoder_q-layer.11": 16742.2012, "encoder_q-layer.2": 13362.2812, "encoder_q-layer.3": 12572.8525, "encoder_q-layer.4": 12098.6328, "encoder_q-layer.5": 12286.3193, "encoder_q-layer.6": 12033.4668, "encoder_q-layer.7": 10507.6738, "encoder_q-layer.8": 8361.7178, "encoder_q-layer.9": 6373.0029, "epoch": 0.58, "inbatch_neg_score": 3.677, "inbatch_pos_score": 4.293, "learning_rate": 1.768421052631579e-05, "loss": 3.3823, "norm_diff": 0.0591, "norm_loss": 0.0, "num_token_doc": 66.674, "num_token_overlap": 11.6685, "num_token_query": 31.3465, "num_token_union": 65.0517, "num_word_context": 202.1066, "num_word_doc": 49.7515, "num_word_query": 23.2696, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18512.0337, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.6777, "query_norm": 2.2865, "queue_k_norm": 2.3457, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3465, "sent_len_1": 66.674, "sent_len_max_0": 127.99, "sent_len_max_1": 206.4112, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 132800 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.4042, "doc_norm": 2.3454, "encoder_q-embeddings": 5090.2559, "encoder_q-layer.0": 3517.259, "encoder_q-layer.1": 3728.1008, "encoder_q-layer.10": 6861.0635, "encoder_q-layer.11": 16192.4053, "encoder_q-layer.2": 4154.5483, "encoder_q-layer.3": 4203.8223, "encoder_q-layer.4": 4494.5527, "encoder_q-layer.5": 4482.98, "encoder_q-layer.6": 4797.4272, "encoder_q-layer.7": 5825.1216, "encoder_q-layer.8": 6840.8579, "encoder_q-layer.9": 6564.9673, "epoch": 0.58, "inbatch_neg_score": 3.6809, "inbatch_pos_score": 4.2812, "learning_rate": 1.7657894736842105e-05, "loss": 3.4042, "norm_diff": 0.0641, "norm_loss": 0.0, "num_token_doc": 66.6699, "num_token_overlap": 11.6897, "num_token_query": 31.3441, "num_token_union": 65.0144, "num_word_context": 202.2142, "num_word_doc": 49.7736, "num_word_query": 23.2701, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10282.5126, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6816, "query_norm": 2.2812, "queue_k_norm": 2.3455, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3441, "sent_len_1": 66.6699, "sent_len_max_0": 127.9887, "sent_len_max_1": 205.1662, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 132900 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.3931, "doc_norm": 2.3457, "encoder_q-embeddings": 4747.354, "encoder_q-layer.0": 3348.8462, "encoder_q-layer.1": 3518.7668, "encoder_q-layer.10": 6132.6816, "encoder_q-layer.11": 15670.3926, "encoder_q-layer.2": 3816.511, "encoder_q-layer.3": 3941.8967, "encoder_q-layer.4": 4203.9937, "encoder_q-layer.5": 4232.5015, "encoder_q-layer.6": 5049.9546, "encoder_q-layer.7": 5697.4155, "encoder_q-layer.8": 6801.3711, "encoder_q-layer.9": 6248.0981, "epoch": 0.58, "inbatch_neg_score": 3.6853, "inbatch_pos_score": 4.2891, "learning_rate": 1.763157894736842e-05, "loss": 3.3931, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.5817, "num_token_overlap": 11.6706, "num_token_query": 31.4274, "num_token_union": 65.0785, "num_word_context": 202.1236, "num_word_doc": 49.7009, "num_word_query": 23.3368, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9923.4934, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6855, "query_norm": 2.2764, "queue_k_norm": 2.3471, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4274, "sent_len_1": 66.5817, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.4313, "stdk": 0.0493, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133000 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3787, "doc_norm": 2.3479, "encoder_q-embeddings": 4653.0078, "encoder_q-layer.0": 3254.8596, "encoder_q-layer.1": 3447.981, "encoder_q-layer.10": 6503.231, "encoder_q-layer.11": 16863.1699, "encoder_q-layer.2": 3901.2686, "encoder_q-layer.3": 3981.6453, "encoder_q-layer.4": 4185.3774, "encoder_q-layer.5": 4312.2681, "encoder_q-layer.6": 4844.7832, "encoder_q-layer.7": 5698.7207, "encoder_q-layer.8": 6924.52, "encoder_q-layer.9": 6404.8042, "epoch": 0.58, "inbatch_neg_score": 3.6911, "inbatch_pos_score": 4.2969, "learning_rate": 1.760526315789474e-05, "loss": 3.3787, "norm_diff": 0.064, "norm_loss": 0.0, "num_token_doc": 66.8454, "num_token_overlap": 11.7014, "num_token_query": 31.4392, "num_token_union": 65.1919, "num_word_context": 202.2312, "num_word_doc": 49.8577, "num_word_query": 23.3582, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10135.5539, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6914, "query_norm": 2.2839, "queue_k_norm": 2.3487, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4392, "sent_len_1": 66.8454, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.3562, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 133100 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3864, "doc_norm": 2.3496, "encoder_q-embeddings": 5083.1782, "encoder_q-layer.0": 3586.4922, "encoder_q-layer.1": 3872.2139, "encoder_q-layer.10": 6444.0996, "encoder_q-layer.11": 15887.6768, "encoder_q-layer.2": 4284.7109, "encoder_q-layer.3": 4586.5322, "encoder_q-layer.4": 4860.5195, "encoder_q-layer.5": 4793.1094, "encoder_q-layer.6": 5336.8374, "encoder_q-layer.7": 5758.6216, "encoder_q-layer.8": 6974.689, "encoder_q-layer.9": 6479.249, "epoch": 0.58, "inbatch_neg_score": 3.6947, "inbatch_pos_score": 4.3047, "learning_rate": 1.7578947368421054e-05, "loss": 3.3864, "norm_diff": 0.0656, "norm_loss": 0.0, "num_token_doc": 66.9955, "num_token_overlap": 11.7165, "num_token_query": 31.4619, "num_token_union": 65.261, "num_word_context": 202.4968, "num_word_doc": 49.9735, "num_word_query": 23.3622, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10076.1313, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6934, "query_norm": 2.2841, "queue_k_norm": 2.3483, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4619, "sent_len_1": 66.9955, "sent_len_max_0": 127.985, "sent_len_max_1": 210.865, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133200 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3782, "doc_norm": 2.35, "encoder_q-embeddings": 4929.5215, "encoder_q-layer.0": 3399.2185, "encoder_q-layer.1": 3536.1272, "encoder_q-layer.10": 6344.793, "encoder_q-layer.11": 15708.4639, "encoder_q-layer.2": 4018.5449, "encoder_q-layer.3": 4278.8706, "encoder_q-layer.4": 4609.7515, "encoder_q-layer.5": 4833.3618, "encoder_q-layer.6": 5264.7339, "encoder_q-layer.7": 6033.7969, "encoder_q-layer.8": 6824.3794, "encoder_q-layer.9": 6102.7271, "epoch": 0.58, "inbatch_neg_score": 3.6948, "inbatch_pos_score": 4.293, "learning_rate": 1.7552631578947372e-05, "loss": 3.3782, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.7706, "num_token_overlap": 11.6961, "num_token_query": 31.4756, "num_token_union": 65.1844, "num_word_context": 202.1716, "num_word_doc": 49.7971, "num_word_query": 23.3656, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9996.8365, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6934, "query_norm": 2.2745, "queue_k_norm": 2.3514, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4756, "sent_len_1": 66.7706, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.9387, "stdk": 0.0493, "stdq": 0.0426, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 133300 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3755, "doc_norm": 2.3453, "encoder_q-embeddings": 4798.9722, "encoder_q-layer.0": 3336.3757, "encoder_q-layer.1": 3655.6638, "encoder_q-layer.10": 6348.6489, "encoder_q-layer.11": 15219.999, "encoder_q-layer.2": 4066.0581, "encoder_q-layer.3": 4222.8726, "encoder_q-layer.4": 4542.4072, "encoder_q-layer.5": 4486.5854, "encoder_q-layer.6": 4936.5942, "encoder_q-layer.7": 5367.5713, "encoder_q-layer.8": 6338.6904, "encoder_q-layer.9": 5936.29, "epoch": 0.58, "inbatch_neg_score": 3.6985, "inbatch_pos_score": 4.3125, "learning_rate": 1.7526315789473683e-05, "loss": 3.3755, "norm_diff": 0.0555, "norm_loss": 0.0, "num_token_doc": 66.6493, "num_token_overlap": 11.7031, "num_token_query": 31.4348, "num_token_union": 65.062, "num_word_context": 202.2754, "num_word_doc": 49.7379, "num_word_query": 23.3514, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9557.4193, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.6992, "query_norm": 2.2899, "queue_k_norm": 2.3507, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4348, "sent_len_1": 66.6493, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.4663, "stdk": 0.049, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133400 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3793, "doc_norm": 2.3528, "encoder_q-embeddings": 4735.9043, "encoder_q-layer.0": 3321.4727, "encoder_q-layer.1": 3554.7451, "encoder_q-layer.10": 6871.3438, "encoder_q-layer.11": 16635.3008, "encoder_q-layer.2": 3914.949, "encoder_q-layer.3": 4048.8787, "encoder_q-layer.4": 4396.7866, "encoder_q-layer.5": 4382.1133, "encoder_q-layer.6": 4970.7061, "encoder_q-layer.7": 6014.1196, "encoder_q-layer.8": 7059.5225, "encoder_q-layer.9": 6401.8716, "epoch": 0.58, "inbatch_neg_score": 3.7011, "inbatch_pos_score": 4.3125, "learning_rate": 1.75e-05, "loss": 3.3793, "norm_diff": 0.0709, "norm_loss": 0.0, "num_token_doc": 66.6919, "num_token_overlap": 11.6315, "num_token_query": 31.2433, "num_token_union": 65.0418, "num_word_context": 202.2218, "num_word_doc": 49.7732, "num_word_query": 23.2004, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10124.3521, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7012, "query_norm": 2.2818, "queue_k_norm": 2.3525, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2433, "sent_len_1": 66.6919, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.54, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133500 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3797, "doc_norm": 2.353, "encoder_q-embeddings": 11049.7764, "encoder_q-layer.0": 8420.2539, "encoder_q-layer.1": 9075.5039, "encoder_q-layer.10": 6469.6738, "encoder_q-layer.11": 15552.167, "encoder_q-layer.2": 10653.0674, "encoder_q-layer.3": 11759.96, "encoder_q-layer.4": 12309.668, "encoder_q-layer.5": 13605.3994, "encoder_q-layer.6": 15276.2959, "encoder_q-layer.7": 13563.2041, "encoder_q-layer.8": 10616.7588, "encoder_q-layer.9": 6615.2002, "epoch": 0.58, "inbatch_neg_score": 3.7075, "inbatch_pos_score": 4.3203, "learning_rate": 1.7473684210526317e-05, "loss": 3.3797, "norm_diff": 0.0606, "norm_loss": 0.0, "num_token_doc": 66.6873, "num_token_overlap": 11.6955, "num_token_query": 31.4512, "num_token_union": 65.1055, "num_word_context": 202.1756, "num_word_doc": 49.7779, "num_word_query": 23.3618, "postclip_grad_norm": 1.0, "preclip_grad_norm": 17254.4673, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.707, "query_norm": 2.2924, "queue_k_norm": 2.3534, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4512, "sent_len_1": 66.6873, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.5975, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133600 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3817, "doc_norm": 2.3548, "encoder_q-embeddings": 4819.7583, "encoder_q-layer.0": 3316.092, "encoder_q-layer.1": 3485.844, "encoder_q-layer.10": 6891.1084, "encoder_q-layer.11": 15164.665, "encoder_q-layer.2": 3942.4626, "encoder_q-layer.3": 4063.9507, "encoder_q-layer.4": 4193.6377, "encoder_q-layer.5": 4520.1494, "encoder_q-layer.6": 4978.0171, "encoder_q-layer.7": 5568.9912, "encoder_q-layer.8": 6710.7578, "encoder_q-layer.9": 6076.6025, "epoch": 0.58, "inbatch_neg_score": 3.7105, "inbatch_pos_score": 4.3203, "learning_rate": 1.7447368421052632e-05, "loss": 3.3817, "norm_diff": 0.0688, "norm_loss": 0.0, "num_token_doc": 66.8476, "num_token_overlap": 11.6867, "num_token_query": 31.3302, "num_token_union": 65.1173, "num_word_context": 202.1871, "num_word_doc": 49.8772, "num_word_query": 23.2768, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9777.4081, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7109, "query_norm": 2.286, "queue_k_norm": 2.3542, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3302, "sent_len_1": 66.8476, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.2113, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133700 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.3851, "doc_norm": 2.3569, "encoder_q-embeddings": 5115.9717, "encoder_q-layer.0": 3571.1738, "encoder_q-layer.1": 3828.0754, "encoder_q-layer.10": 7173.2383, "encoder_q-layer.11": 16589.6211, "encoder_q-layer.2": 4263.415, "encoder_q-layer.3": 4582.3433, "encoder_q-layer.4": 4796.4854, "encoder_q-layer.5": 4934.4888, "encoder_q-layer.6": 5616.2852, "encoder_q-layer.7": 5987.2944, "encoder_q-layer.8": 7114.7446, "encoder_q-layer.9": 6500.8379, "epoch": 0.58, "inbatch_neg_score": 3.7146, "inbatch_pos_score": 4.3086, "learning_rate": 1.742105263157895e-05, "loss": 3.3851, "norm_diff": 0.0764, "norm_loss": 0.0, "num_token_doc": 66.7895, "num_token_overlap": 11.6588, "num_token_query": 31.3177, "num_token_union": 65.1063, "num_word_context": 202.4218, "num_word_doc": 49.8596, "num_word_query": 23.2605, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10682.9357, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7148, "query_norm": 2.2805, "queue_k_norm": 2.3557, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3177, "sent_len_1": 66.7895, "sent_len_max_0": 127.975, "sent_len_max_1": 206.6788, "stdk": 0.0495, "stdq": 0.0427, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 133800 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3894, "doc_norm": 2.3566, "encoder_q-embeddings": 5014.7651, "encoder_q-layer.0": 3444.3042, "encoder_q-layer.1": 3678.9426, "encoder_q-layer.10": 6972.8423, "encoder_q-layer.11": 15702.7686, "encoder_q-layer.2": 4051.9778, "encoder_q-layer.3": 4309.0059, "encoder_q-layer.4": 4511.1274, "encoder_q-layer.5": 4751.0518, "encoder_q-layer.6": 5243.1167, "encoder_q-layer.7": 5927.915, "encoder_q-layer.8": 7122.9229, "encoder_q-layer.9": 6570.6191, "epoch": 0.58, "inbatch_neg_score": 3.7187, "inbatch_pos_score": 4.332, "learning_rate": 1.7394736842105265e-05, "loss": 3.3894, "norm_diff": 0.0657, "norm_loss": 0.0, "num_token_doc": 66.813, "num_token_overlap": 11.6923, "num_token_query": 31.4226, "num_token_union": 65.1635, "num_word_context": 202.2605, "num_word_doc": 49.8453, "num_word_query": 23.3531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10052.8608, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7188, "query_norm": 2.2909, "queue_k_norm": 2.3557, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4226, "sent_len_1": 66.813, "sent_len_max_0": 127.975, "sent_len_max_1": 209.2388, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 133900 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3943, "doc_norm": 2.3552, "encoder_q-embeddings": 5661.1123, "encoder_q-layer.0": 3861.5361, "encoder_q-layer.1": 4062.4326, "encoder_q-layer.10": 8035.0547, "encoder_q-layer.11": 17402.0957, "encoder_q-layer.2": 4611.8604, "encoder_q-layer.3": 4951.4629, "encoder_q-layer.4": 5229.6611, "encoder_q-layer.5": 5299.0078, "encoder_q-layer.6": 6262.5068, "encoder_q-layer.7": 6611.4053, "encoder_q-layer.8": 7636.8721, "encoder_q-layer.9": 7121.2725, "epoch": 0.58, "inbatch_neg_score": 3.7206, "inbatch_pos_score": 4.3281, "learning_rate": 1.736842105263158e-05, "loss": 3.3943, "norm_diff": 0.0664, "norm_loss": 0.0, "num_token_doc": 66.7707, "num_token_overlap": 11.6713, "num_token_query": 31.3418, "num_token_union": 65.1117, "num_word_context": 202.2499, "num_word_doc": 49.8417, "num_word_query": 23.2786, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11361.1858, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7207, "query_norm": 2.2888, "queue_k_norm": 2.3572, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3418, "sent_len_1": 66.7707, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.6712, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134000 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3793, "doc_norm": 2.3606, "encoder_q-embeddings": 4957.9902, "encoder_q-layer.0": 3399.7754, "encoder_q-layer.1": 3529.6252, "encoder_q-layer.10": 6425.4805, "encoder_q-layer.11": 15591.9463, "encoder_q-layer.2": 4015.572, "encoder_q-layer.3": 4211.6455, "encoder_q-layer.4": 4401.5195, "encoder_q-layer.5": 4472.585, "encoder_q-layer.6": 5130.708, "encoder_q-layer.7": 5738.6167, "encoder_q-layer.8": 6975.9922, "encoder_q-layer.9": 6368.4141, "epoch": 0.58, "inbatch_neg_score": 3.7237, "inbatch_pos_score": 4.3438, "learning_rate": 1.7342105263157895e-05, "loss": 3.3793, "norm_diff": 0.064, "norm_loss": 0.0, "num_token_doc": 66.729, "num_token_overlap": 11.7011, "num_token_query": 31.4452, "num_token_union": 65.1363, "num_word_context": 202.2521, "num_word_doc": 49.8053, "num_word_query": 23.3558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9965.5187, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7227, "query_norm": 2.2966, "queue_k_norm": 2.3584, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4452, "sent_len_1": 66.729, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.955, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134100 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.4054, "doc_norm": 2.3614, "encoder_q-embeddings": 5108.4414, "encoder_q-layer.0": 3578.8113, "encoder_q-layer.1": 3917.4966, "encoder_q-layer.10": 6943.6245, "encoder_q-layer.11": 15745.3633, "encoder_q-layer.2": 4331.3696, "encoder_q-layer.3": 4448.624, "encoder_q-layer.4": 4754.7271, "encoder_q-layer.5": 4778.8623, "encoder_q-layer.6": 5111.9175, "encoder_q-layer.7": 5676.3491, "encoder_q-layer.8": 6972.8311, "encoder_q-layer.9": 6351.7969, "epoch": 0.58, "inbatch_neg_score": 3.7309, "inbatch_pos_score": 4.3359, "learning_rate": 1.731578947368421e-05, "loss": 3.4054, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.7544, "num_token_overlap": 11.606, "num_token_query": 31.238, "num_token_union": 65.0802, "num_word_context": 202.2133, "num_word_doc": 49.8256, "num_word_query": 23.1926, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10210.1876, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7305, "query_norm": 2.2911, "queue_k_norm": 2.3594, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.238, "sent_len_1": 66.7544, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.79, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134200 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3851, "doc_norm": 2.3616, "encoder_q-embeddings": 4874.355, "encoder_q-layer.0": 3516.7832, "encoder_q-layer.1": 3736.6965, "encoder_q-layer.10": 6265.292, "encoder_q-layer.11": 15813.3145, "encoder_q-layer.2": 4066.4255, "encoder_q-layer.3": 4044.8547, "encoder_q-layer.4": 4333.1514, "encoder_q-layer.5": 4362.1211, "encoder_q-layer.6": 5096.5571, "encoder_q-layer.7": 5797.4922, "encoder_q-layer.8": 7160.5376, "encoder_q-layer.9": 6327.8467, "epoch": 0.58, "inbatch_neg_score": 3.7371, "inbatch_pos_score": 4.3594, "learning_rate": 1.728947368421053e-05, "loss": 3.3851, "norm_diff": 0.0608, "norm_loss": 0.0, "num_token_doc": 66.5776, "num_token_overlap": 11.6842, "num_token_query": 31.3993, "num_token_union": 65.0189, "num_word_context": 202.1574, "num_word_doc": 49.6701, "num_word_query": 23.3331, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10005.802, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7363, "query_norm": 2.3009, "queue_k_norm": 2.3593, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3993, "sent_len_1": 66.5776, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.7738, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134300 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3846, "doc_norm": 2.3625, "encoder_q-embeddings": 4717.8901, "encoder_q-layer.0": 3533.9512, "encoder_q-layer.1": 3768.262, "encoder_q-layer.10": 7826.9946, "encoder_q-layer.11": 16569.791, "encoder_q-layer.2": 4117.1777, "encoder_q-layer.3": 4200.9453, "encoder_q-layer.4": 4547.7583, "encoder_q-layer.5": 4882.6187, "encoder_q-layer.6": 5204.9209, "encoder_q-layer.7": 5938.2622, "encoder_q-layer.8": 7331.9824, "encoder_q-layer.9": 6932.541, "epoch": 0.58, "inbatch_neg_score": 3.7416, "inbatch_pos_score": 4.3516, "learning_rate": 1.7263157894736843e-05, "loss": 3.3846, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.6209, "num_token_overlap": 11.6863, "num_token_query": 31.3341, "num_token_union": 64.9956, "num_word_context": 202.1624, "num_word_doc": 49.718, "num_word_query": 23.2645, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10413.7527, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.7422, "query_norm": 2.2904, "queue_k_norm": 2.3608, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3341, "sent_len_1": 66.6209, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.7113, "stdk": 0.0495, "stdq": 0.0427, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134400 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3892, "doc_norm": 2.3647, "encoder_q-embeddings": 4756.3242, "encoder_q-layer.0": 3346.7812, "encoder_q-layer.1": 3556.4856, "encoder_q-layer.10": 7390.2969, "encoder_q-layer.11": 16273.4395, "encoder_q-layer.2": 4051.7136, "encoder_q-layer.3": 4226.835, "encoder_q-layer.4": 4614.2261, "encoder_q-layer.5": 4723.2832, "encoder_q-layer.6": 5059.5693, "encoder_q-layer.7": 5929.1392, "encoder_q-layer.8": 7102.2217, "encoder_q-layer.9": 6989.6646, "epoch": 0.58, "inbatch_neg_score": 3.7514, "inbatch_pos_score": 4.3633, "learning_rate": 1.723684210526316e-05, "loss": 3.3892, "norm_diff": 0.0622, "norm_loss": 0.0, "num_token_doc": 66.7555, "num_token_overlap": 11.695, "num_token_query": 31.4205, "num_token_union": 65.1159, "num_word_context": 202.3512, "num_word_doc": 49.8069, "num_word_query": 23.3435, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10329.3346, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.752, "query_norm": 2.3025, "queue_k_norm": 2.3629, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4205, "sent_len_1": 66.7555, "sent_len_max_0": 127.98, "sent_len_max_1": 210.1987, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 134500 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.4008, "doc_norm": 2.3624, "encoder_q-embeddings": 9382.8037, "encoder_q-layer.0": 6728.1133, "encoder_q-layer.1": 7078.457, "encoder_q-layer.10": 13452.7285, "encoder_q-layer.11": 31100.4941, "encoder_q-layer.2": 7827.5688, "encoder_q-layer.3": 8273.8682, "encoder_q-layer.4": 8921.5518, "encoder_q-layer.5": 9140.4707, "encoder_q-layer.6": 9521.8174, "encoder_q-layer.7": 11046.8379, "encoder_q-layer.8": 13605.9922, "encoder_q-layer.9": 13160.6406, "epoch": 0.58, "inbatch_neg_score": 3.7604, "inbatch_pos_score": 4.3555, "learning_rate": 1.7210526315789473e-05, "loss": 3.4008, "norm_diff": 0.063, "norm_loss": 0.0, "num_token_doc": 66.6413, "num_token_overlap": 11.661, "num_token_query": 31.4258, "num_token_union": 65.0872, "num_word_context": 202.5041, "num_word_doc": 49.7202, "num_word_query": 23.347, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19541.6944, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.7598, "query_norm": 2.2994, "queue_k_norm": 2.3626, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4258, "sent_len_1": 66.6413, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.3688, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134600 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.3757, "doc_norm": 2.3637, "encoder_q-embeddings": 9504.374, "encoder_q-layer.0": 6811.0024, "encoder_q-layer.1": 7126.6519, "encoder_q-layer.10": 12666.626, "encoder_q-layer.11": 31545.4512, "encoder_q-layer.2": 8037.6685, "encoder_q-layer.3": 7923.085, "encoder_q-layer.4": 8421.1758, "encoder_q-layer.5": 8768.6367, "encoder_q-layer.6": 9710.8652, "encoder_q-layer.7": 11109.6475, "encoder_q-layer.8": 13120.5557, "encoder_q-layer.9": 12493.4512, "epoch": 0.58, "inbatch_neg_score": 3.7688, "inbatch_pos_score": 4.3633, "learning_rate": 1.718421052631579e-05, "loss": 3.3757, "norm_diff": 0.0596, "norm_loss": 0.0, "num_token_doc": 66.821, "num_token_overlap": 11.7006, "num_token_query": 31.4763, "num_token_union": 65.2044, "num_word_context": 202.4705, "num_word_doc": 49.8803, "num_word_query": 23.3824, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19692.5303, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.7676, "query_norm": 2.304, "queue_k_norm": 2.3642, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4763, "sent_len_1": 66.821, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.8113, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134700 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.3935, "doc_norm": 2.3643, "encoder_q-embeddings": 10928.585, "encoder_q-layer.0": 7795.3784, "encoder_q-layer.1": 8465.8438, "encoder_q-layer.10": 11913.0195, "encoder_q-layer.11": 31019.1934, "encoder_q-layer.2": 9572.416, "encoder_q-layer.3": 9880.7295, "encoder_q-layer.4": 10451.0137, "encoder_q-layer.5": 11071.2734, "encoder_q-layer.6": 11399.1289, "encoder_q-layer.7": 12813.4316, "encoder_q-layer.8": 14021.5186, "encoder_q-layer.9": 11986.3555, "epoch": 0.58, "inbatch_neg_score": 3.7749, "inbatch_pos_score": 4.375, "learning_rate": 1.7157894736842107e-05, "loss": 3.3935, "norm_diff": 0.0603, "norm_loss": 0.0, "num_token_doc": 66.8376, "num_token_overlap": 11.6958, "num_token_query": 31.358, "num_token_union": 65.1261, "num_word_context": 202.4146, "num_word_doc": 49.8821, "num_word_query": 23.3008, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20931.8304, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.7754, "query_norm": 2.3039, "queue_k_norm": 2.3649, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.358, "sent_len_1": 66.8376, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.0462, "stdk": 0.0493, "stdq": 0.0428, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134800 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.3778, "doc_norm": 2.3675, "encoder_q-embeddings": 10050.7129, "encoder_q-layer.0": 7051.3828, "encoder_q-layer.1": 7530.6797, "encoder_q-layer.10": 13206.3184, "encoder_q-layer.11": 32231.082, "encoder_q-layer.2": 8399.5693, "encoder_q-layer.3": 8523.1445, "encoder_q-layer.4": 9540.127, "encoder_q-layer.5": 9331.8105, "encoder_q-layer.6": 10670.0391, "encoder_q-layer.7": 11207.9668, "encoder_q-layer.8": 12811.0684, "encoder_q-layer.9": 12263.8037, "epoch": 0.59, "inbatch_neg_score": 3.7843, "inbatch_pos_score": 4.3867, "learning_rate": 1.713157894736842e-05, "loss": 3.3778, "norm_diff": 0.0569, "norm_loss": 0.0, "num_token_doc": 66.8849, "num_token_overlap": 11.6774, "num_token_query": 31.3636, "num_token_union": 65.1901, "num_word_context": 202.4214, "num_word_doc": 49.9335, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20679.6405, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 3.7852, "query_norm": 2.3106, "queue_k_norm": 2.3666, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3636, "sent_len_1": 66.8849, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.2637, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 134900 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3855, "doc_norm": 2.3679, "encoder_q-embeddings": 2311.7974, "encoder_q-layer.0": 1628.6985, "encoder_q-layer.1": 1710.5941, "encoder_q-layer.10": 3241.8513, "encoder_q-layer.11": 7766.4712, "encoder_q-layer.2": 1911.6466, "encoder_q-layer.3": 1979.3357, "encoder_q-layer.4": 2152.0417, "encoder_q-layer.5": 2162.9504, "encoder_q-layer.6": 2472.6758, "encoder_q-layer.7": 2795.0552, "encoder_q-layer.8": 3388.3298, "encoder_q-layer.9": 3066.4473, "epoch": 0.59, "inbatch_neg_score": 3.7949, "inbatch_pos_score": 4.4062, "learning_rate": 1.7105263157894737e-05, "loss": 3.3855, "norm_diff": 0.0519, "norm_loss": 0.0, "num_token_doc": 66.7552, "num_token_overlap": 11.6895, "num_token_query": 31.4899, "num_token_union": 65.1648, "num_word_context": 202.1301, "num_word_doc": 49.8214, "num_word_query": 23.4075, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4874.6061, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.7949, "query_norm": 2.3161, "queue_k_norm": 2.3679, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4899, "sent_len_1": 66.7552, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.31, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135000 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3836, "doc_norm": 2.3675, "encoder_q-embeddings": 2912.6042, "encoder_q-layer.0": 2040.7074, "encoder_q-layer.1": 2158.2375, "encoder_q-layer.10": 3249.3374, "encoder_q-layer.11": 7930.543, "encoder_q-layer.2": 2393.5315, "encoder_q-layer.3": 2555.7974, "encoder_q-layer.4": 2778.6565, "encoder_q-layer.5": 2894.969, "encoder_q-layer.6": 3013.1672, "encoder_q-layer.7": 3268.1184, "encoder_q-layer.8": 3551.2551, "encoder_q-layer.9": 3139.7783, "epoch": 0.59, "inbatch_neg_score": 3.8034, "inbatch_pos_score": 4.4141, "learning_rate": 1.707894736842105e-05, "loss": 3.3836, "norm_diff": 0.0476, "norm_loss": 0.0, "num_token_doc": 66.7821, "num_token_overlap": 11.7093, "num_token_query": 31.3929, "num_token_union": 65.0893, "num_word_context": 202.3192, "num_word_doc": 49.82, "num_word_query": 23.3259, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5401.3635, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8027, "query_norm": 2.3199, "queue_k_norm": 2.3694, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3929, "sent_len_1": 66.7821, "sent_len_max_0": 127.9988, "sent_len_max_1": 210.3575, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135100 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3689, "doc_norm": 2.3706, "encoder_q-embeddings": 2427.4976, "encoder_q-layer.0": 1701.7535, "encoder_q-layer.1": 1793.0718, "encoder_q-layer.10": 3790.0701, "encoder_q-layer.11": 8621.2666, "encoder_q-layer.2": 2007.0153, "encoder_q-layer.3": 2057.7339, "encoder_q-layer.4": 2161.5608, "encoder_q-layer.5": 2261.4814, "encoder_q-layer.6": 2554.551, "encoder_q-layer.7": 3056.1165, "encoder_q-layer.8": 3854.6001, "encoder_q-layer.9": 3442.3674, "epoch": 0.59, "inbatch_neg_score": 3.8136, "inbatch_pos_score": 4.4141, "learning_rate": 1.705263157894737e-05, "loss": 3.3689, "norm_diff": 0.0492, "norm_loss": 0.0, "num_token_doc": 66.8597, "num_token_overlap": 11.7041, "num_token_query": 31.393, "num_token_union": 65.1407, "num_word_context": 202.2186, "num_word_doc": 49.855, "num_word_query": 23.3057, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5348.9493, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8125, "query_norm": 2.3214, "queue_k_norm": 2.3711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.393, "sent_len_1": 66.8597, "sent_len_max_0": 127.9688, "sent_len_max_1": 211.8237, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135200 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3922, "doc_norm": 2.3727, "encoder_q-embeddings": 2454.3179, "encoder_q-layer.0": 1651.3618, "encoder_q-layer.1": 1749.2015, "encoder_q-layer.10": 3305.3149, "encoder_q-layer.11": 8022.2148, "encoder_q-layer.2": 1940.3477, "encoder_q-layer.3": 1996.3892, "encoder_q-layer.4": 2093.6401, "encoder_q-layer.5": 2128.3784, "encoder_q-layer.6": 2338.0479, "encoder_q-layer.7": 2880.4641, "encoder_q-layer.8": 3300.4424, "encoder_q-layer.9": 3036.7737, "epoch": 0.59, "inbatch_neg_score": 3.8208, "inbatch_pos_score": 4.4219, "learning_rate": 1.7026315789473685e-05, "loss": 3.3922, "norm_diff": 0.0578, "norm_loss": 0.0, "num_token_doc": 66.7298, "num_token_overlap": 11.6494, "num_token_query": 31.2575, "num_token_union": 65.043, "num_word_context": 202.2051, "num_word_doc": 49.7909, "num_word_query": 23.2192, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4961.1381, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8203, "query_norm": 2.3149, "queue_k_norm": 2.3725, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2575, "sent_len_1": 66.7298, "sent_len_max_0": 127.995, "sent_len_max_1": 210.7675, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135300 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3919, "doc_norm": 2.3734, "encoder_q-embeddings": 2488.6819, "encoder_q-layer.0": 1733.7275, "encoder_q-layer.1": 1866.4489, "encoder_q-layer.10": 3171.8599, "encoder_q-layer.11": 8071.7568, "encoder_q-layer.2": 2002.1868, "encoder_q-layer.3": 2115.2947, "encoder_q-layer.4": 2170.0037, "encoder_q-layer.5": 2226.9922, "encoder_q-layer.6": 2528.3552, "encoder_q-layer.7": 2852.0659, "encoder_q-layer.8": 3358.1938, "encoder_q-layer.9": 3086.4133, "epoch": 0.59, "inbatch_neg_score": 3.8277, "inbatch_pos_score": 4.4336, "learning_rate": 1.7000000000000003e-05, "loss": 3.3919, "norm_diff": 0.0543, "norm_loss": 0.0, "num_token_doc": 66.8105, "num_token_overlap": 11.6461, "num_token_query": 31.3, "num_token_union": 65.1112, "num_word_context": 202.2779, "num_word_doc": 49.8359, "num_word_query": 23.2359, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5042.6811, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8281, "query_norm": 2.3191, "queue_k_norm": 2.3745, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3, "sent_len_1": 66.8105, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.7325, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 135400 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3879, "doc_norm": 2.378, "encoder_q-embeddings": 2385.9841, "encoder_q-layer.0": 1679.6964, "encoder_q-layer.1": 1800.1562, "encoder_q-layer.10": 3344.2092, "encoder_q-layer.11": 8350.126, "encoder_q-layer.2": 1957.7891, "encoder_q-layer.3": 2007.7726, "encoder_q-layer.4": 2188.5244, "encoder_q-layer.5": 2217.2288, "encoder_q-layer.6": 2469.1509, "encoder_q-layer.7": 2754.4849, "encoder_q-layer.8": 3373.8997, "encoder_q-layer.9": 3145.5508, "epoch": 0.59, "inbatch_neg_score": 3.8313, "inbatch_pos_score": 4.4492, "learning_rate": 1.6973684210526318e-05, "loss": 3.3879, "norm_diff": 0.0568, "norm_loss": 0.0, "num_token_doc": 66.9217, "num_token_overlap": 11.6754, "num_token_query": 31.338, "num_token_union": 65.1972, "num_word_context": 202.4595, "num_word_doc": 49.9488, "num_word_query": 23.2688, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5127.9245, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.832, "query_norm": 2.3213, "queue_k_norm": 2.3759, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.338, "sent_len_1": 66.9217, "sent_len_max_0": 127.99, "sent_len_max_1": 206.885, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135500 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3934, "doc_norm": 2.3792, "encoder_q-embeddings": 2414.2742, "encoder_q-layer.0": 1655.5042, "encoder_q-layer.1": 1764.8917, "encoder_q-layer.10": 3407.8843, "encoder_q-layer.11": 8543.6641, "encoder_q-layer.2": 1964.3851, "encoder_q-layer.3": 2033.4081, "encoder_q-layer.4": 2136.5032, "encoder_q-layer.5": 2198.019, "encoder_q-layer.6": 2448.3909, "encoder_q-layer.7": 3010.1711, "encoder_q-layer.8": 3577.248, "encoder_q-layer.9": 3320.2327, "epoch": 0.59, "inbatch_neg_score": 3.8356, "inbatch_pos_score": 4.4531, "learning_rate": 1.694736842105263e-05, "loss": 3.3934, "norm_diff": 0.0531, "norm_loss": 0.0, "num_token_doc": 66.6081, "num_token_overlap": 11.6752, "num_token_query": 31.4385, "num_token_union": 65.0722, "num_word_context": 202.1656, "num_word_doc": 49.704, "num_word_query": 23.3558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5166.8532, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8359, "query_norm": 2.3261, "queue_k_norm": 2.3765, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4385, "sent_len_1": 66.6081, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.2812, "stdk": 0.0495, "stdq": 0.0439, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135600 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3761, "doc_norm": 2.3775, "encoder_q-embeddings": 2342.9377, "encoder_q-layer.0": 1714.9288, "encoder_q-layer.1": 1822.7697, "encoder_q-layer.10": 3290.874, "encoder_q-layer.11": 8031.8521, "encoder_q-layer.2": 2004.6996, "encoder_q-layer.3": 2035.5839, "encoder_q-layer.4": 2163.5928, "encoder_q-layer.5": 2256.6938, "encoder_q-layer.6": 2553.833, "encoder_q-layer.7": 2828.6567, "encoder_q-layer.8": 3428.073, "encoder_q-layer.9": 3077.1333, "epoch": 0.59, "inbatch_neg_score": 3.8406, "inbatch_pos_score": 4.4453, "learning_rate": 1.6921052631578948e-05, "loss": 3.3761, "norm_diff": 0.0593, "norm_loss": 0.0, "num_token_doc": 66.7296, "num_token_overlap": 11.6833, "num_token_query": 31.3577, "num_token_union": 65.0987, "num_word_context": 202.4684, "num_word_doc": 49.7786, "num_word_query": 23.2937, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5036.8701, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8398, "query_norm": 2.3182, "queue_k_norm": 2.3785, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3577, "sent_len_1": 66.7296, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.9162, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135700 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3781, "doc_norm": 2.3792, "encoder_q-embeddings": 2524.6565, "encoder_q-layer.0": 1763.449, "encoder_q-layer.1": 1837.3958, "encoder_q-layer.10": 3387.499, "encoder_q-layer.11": 7910.4365, "encoder_q-layer.2": 2077.7471, "encoder_q-layer.3": 2084.2825, "encoder_q-layer.4": 2269.8711, "encoder_q-layer.5": 2335.5845, "encoder_q-layer.6": 2560.9492, "encoder_q-layer.7": 2956.5618, "encoder_q-layer.8": 3372.5415, "encoder_q-layer.9": 3172.5879, "epoch": 0.59, "inbatch_neg_score": 3.8472, "inbatch_pos_score": 4.4648, "learning_rate": 1.6894736842105263e-05, "loss": 3.3781, "norm_diff": 0.0584, "norm_loss": 0.0, "num_token_doc": 66.8438, "num_token_overlap": 11.6762, "num_token_query": 31.2782, "num_token_union": 65.1144, "num_word_context": 202.5734, "num_word_doc": 49.9049, "num_word_query": 23.2377, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5094.7518, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8477, "query_norm": 2.3208, "queue_k_norm": 2.379, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2782, "sent_len_1": 66.8438, "sent_len_max_0": 127.9788, "sent_len_max_1": 207.98, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135800 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.3913, "doc_norm": 2.3808, "encoder_q-embeddings": 2616.8828, "encoder_q-layer.0": 1862.6284, "encoder_q-layer.1": 1991.759, "encoder_q-layer.10": 3480.615, "encoder_q-layer.11": 8068.7212, "encoder_q-layer.2": 2179.9524, "encoder_q-layer.3": 2245.4988, "encoder_q-layer.4": 2397.8403, "encoder_q-layer.5": 2464.6624, "encoder_q-layer.6": 2747.6492, "encoder_q-layer.7": 3103.866, "encoder_q-layer.8": 3507.0994, "encoder_q-layer.9": 3217.0554, "epoch": 0.59, "inbatch_neg_score": 3.8497, "inbatch_pos_score": 4.4492, "learning_rate": 1.686842105263158e-05, "loss": 3.3913, "norm_diff": 0.0689, "norm_loss": 0.0, "num_token_doc": 66.6276, "num_token_overlap": 11.6641, "num_token_query": 31.3412, "num_token_union": 64.9931, "num_word_context": 201.9337, "num_word_doc": 49.6995, "num_word_query": 23.2727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5295.0356, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8496, "query_norm": 2.3119, "queue_k_norm": 2.3804, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3412, "sent_len_1": 66.6276, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.1113, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 135900 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3841, "doc_norm": 2.3827, "encoder_q-embeddings": 2435.7253, "encoder_q-layer.0": 1678.5707, "encoder_q-layer.1": 1765.6323, "encoder_q-layer.10": 3671.6038, "encoder_q-layer.11": 8131.835, "encoder_q-layer.2": 1979.6078, "encoder_q-layer.3": 2039.4911, "encoder_q-layer.4": 2238.657, "encoder_q-layer.5": 2257.7695, "encoder_q-layer.6": 2456.4194, "encoder_q-layer.7": 2834.4949, "encoder_q-layer.8": 3494.8127, "encoder_q-layer.9": 3382.7795, "epoch": 0.59, "inbatch_neg_score": 3.8555, "inbatch_pos_score": 4.4531, "learning_rate": 1.6842105263157896e-05, "loss": 3.3841, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.8209, "num_token_overlap": 11.6732, "num_token_query": 31.2681, "num_token_union": 65.0837, "num_word_context": 202.3461, "num_word_doc": 49.8655, "num_word_query": 23.2196, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5074.7172, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8555, "query_norm": 2.3092, "queue_k_norm": 2.3829, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2681, "sent_len_1": 66.8209, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.2163, "stdk": 0.0494, "stdq": 0.0426, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 136000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3751, "doc_norm": 2.3848, "encoder_q-embeddings": 2526.8486, "encoder_q-layer.0": 1769.0468, "encoder_q-layer.1": 1879.9716, "encoder_q-layer.10": 3101.5688, "encoder_q-layer.11": 7669.1387, "encoder_q-layer.2": 2097.1626, "encoder_q-layer.3": 2165.0112, "encoder_q-layer.4": 2310.5168, "encoder_q-layer.5": 2332.8574, "encoder_q-layer.6": 2536.7715, "encoder_q-layer.7": 2764.981, "encoder_q-layer.8": 3182.72, "encoder_q-layer.9": 3036.4282, "epoch": 0.59, "inbatch_neg_score": 3.8578, "inbatch_pos_score": 4.4766, "learning_rate": 1.681578947368421e-05, "loss": 3.3751, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.8658, "num_token_overlap": 11.6918, "num_token_query": 31.304, "num_token_union": 65.1024, "num_word_context": 202.4458, "num_word_doc": 49.8989, "num_word_query": 23.2394, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4898.2347, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8574, "query_norm": 2.3189, "queue_k_norm": 2.3844, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.304, "sent_len_1": 66.8658, "sent_len_max_0": 128.0, "sent_len_max_1": 208.12, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 136100 }, { "accuracy": 48.3887, "active_queue_size": 16384.0, "cl_loss": 3.3844, "doc_norm": 2.3826, "encoder_q-embeddings": 2547.2839, "encoder_q-layer.0": 1781.8099, "encoder_q-layer.1": 1862.9017, "encoder_q-layer.10": 4206.8848, "encoder_q-layer.11": 8993.5674, "encoder_q-layer.2": 2118.6565, "encoder_q-layer.3": 2224.834, "encoder_q-layer.4": 2343.811, "encoder_q-layer.5": 2439.9216, "encoder_q-layer.6": 2655.946, "encoder_q-layer.7": 3208.5959, "encoder_q-layer.8": 3775.8142, "encoder_q-layer.9": 3578.5974, "epoch": 0.59, "inbatch_neg_score": 3.8615, "inbatch_pos_score": 4.4609, "learning_rate": 1.6789473684210526e-05, "loss": 3.3844, "norm_diff": 0.0631, "norm_loss": 0.0, "num_token_doc": 66.784, "num_token_overlap": 11.6878, "num_token_query": 31.4209, "num_token_union": 65.148, "num_word_context": 202.4083, "num_word_doc": 49.8258, "num_word_query": 23.3418, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5501.8915, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8613, "query_norm": 2.3195, "queue_k_norm": 2.3833, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4209, "sent_len_1": 66.784, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6225, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136200 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3728, "doc_norm": 2.3857, "encoder_q-embeddings": 2788.3789, "encoder_q-layer.0": 1964.7202, "encoder_q-layer.1": 2066.7468, "encoder_q-layer.10": 3182.7102, "encoder_q-layer.11": 8081.9941, "encoder_q-layer.2": 2285.8311, "encoder_q-layer.3": 2315.0708, "encoder_q-layer.4": 2560.8049, "encoder_q-layer.5": 2831.5012, "encoder_q-layer.6": 3037.4971, "encoder_q-layer.7": 3317.8154, "encoder_q-layer.8": 3523.2236, "encoder_q-layer.9": 3121.6021, "epoch": 0.59, "inbatch_neg_score": 3.8597, "inbatch_pos_score": 4.4805, "learning_rate": 1.676315789473684e-05, "loss": 3.3728, "norm_diff": 0.0645, "norm_loss": 0.0, "num_token_doc": 66.6465, "num_token_overlap": 11.6842, "num_token_query": 31.4292, "num_token_union": 65.0499, "num_word_context": 202.1108, "num_word_doc": 49.7057, "num_word_query": 23.3497, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5268.3689, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8594, "query_norm": 2.3213, "queue_k_norm": 2.3867, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4292, "sent_len_1": 66.6465, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.2237, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 136300 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3665, "doc_norm": 2.3864, "encoder_q-embeddings": 2448.2661, "encoder_q-layer.0": 1756.9089, "encoder_q-layer.1": 1889.1652, "encoder_q-layer.10": 3343.8611, "encoder_q-layer.11": 7638.3579, "encoder_q-layer.2": 2147.0112, "encoder_q-layer.3": 2206.9133, "encoder_q-layer.4": 2378.8989, "encoder_q-layer.5": 2475.928, "encoder_q-layer.6": 2852.7478, "encoder_q-layer.7": 3347.2966, "encoder_q-layer.8": 3390.9592, "encoder_q-layer.9": 3018.6157, "epoch": 0.59, "inbatch_neg_score": 3.8627, "inbatch_pos_score": 4.4805, "learning_rate": 1.673684210526316e-05, "loss": 3.3665, "norm_diff": 0.0661, "norm_loss": 0.0, "num_token_doc": 66.9098, "num_token_overlap": 11.7022, "num_token_query": 31.4167, "num_token_union": 65.1991, "num_word_context": 202.5253, "num_word_doc": 49.9119, "num_word_query": 23.3313, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5032.4491, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8633, "query_norm": 2.3203, "queue_k_norm": 2.3866, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4167, "sent_len_1": 66.9098, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.9538, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136400 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3734, "doc_norm": 2.3881, "encoder_q-embeddings": 2550.6946, "encoder_q-layer.0": 1781.3168, "encoder_q-layer.1": 1932.965, "encoder_q-layer.10": 3049.2986, "encoder_q-layer.11": 7594.5806, "encoder_q-layer.2": 2157.5454, "encoder_q-layer.3": 2296.6072, "encoder_q-layer.4": 2532.6045, "encoder_q-layer.5": 2563.0762, "encoder_q-layer.6": 2798.7754, "encoder_q-layer.7": 3031.5166, "encoder_q-layer.8": 3402.7715, "encoder_q-layer.9": 3055.8655, "epoch": 0.59, "inbatch_neg_score": 3.8662, "inbatch_pos_score": 4.4766, "learning_rate": 1.6710526315789475e-05, "loss": 3.3734, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.8776, "num_token_overlap": 11.7208, "num_token_query": 31.3693, "num_token_union": 65.1511, "num_word_context": 202.3529, "num_word_doc": 49.9391, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4957.1306, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8672, "query_norm": 2.3144, "queue_k_norm": 2.3873, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3693, "sent_len_1": 66.8776, "sent_len_max_0": 127.96, "sent_len_max_1": 208.1838, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136500 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3748, "doc_norm": 2.3866, "encoder_q-embeddings": 3096.2412, "encoder_q-layer.0": 2172.9343, "encoder_q-layer.1": 2360.8281, "encoder_q-layer.10": 3115.5903, "encoder_q-layer.11": 7782.9937, "encoder_q-layer.2": 2698.9292, "encoder_q-layer.3": 2854.1125, "encoder_q-layer.4": 3080.2695, "encoder_q-layer.5": 3193.5371, "encoder_q-layer.6": 3457.4641, "encoder_q-layer.7": 3535.4119, "encoder_q-layer.8": 3602.6875, "encoder_q-layer.9": 3061.8494, "epoch": 0.59, "inbatch_neg_score": 3.8689, "inbatch_pos_score": 4.4766, "learning_rate": 1.668421052631579e-05, "loss": 3.3748, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.9448, "num_token_overlap": 11.6949, "num_token_query": 31.4068, "num_token_union": 65.2095, "num_word_context": 202.58, "num_word_doc": 49.9652, "num_word_query": 23.3237, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5546.2588, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8691, "query_norm": 2.3145, "queue_k_norm": 2.3879, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4068, "sent_len_1": 66.9448, "sent_len_max_0": 127.985, "sent_len_max_1": 210.8475, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136600 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3752, "doc_norm": 2.3876, "encoder_q-embeddings": 2749.0806, "encoder_q-layer.0": 1940.0515, "encoder_q-layer.1": 2255.647, "encoder_q-layer.10": 3279.7739, "encoder_q-layer.11": 7778.8184, "encoder_q-layer.2": 2568.9834, "encoder_q-layer.3": 2720.5967, "encoder_q-layer.4": 2920.3159, "encoder_q-layer.5": 2944.3672, "encoder_q-layer.6": 2895.6892, "encoder_q-layer.7": 3104.4934, "encoder_q-layer.8": 3394.7871, "encoder_q-layer.9": 3122.866, "epoch": 0.59, "inbatch_neg_score": 3.8702, "inbatch_pos_score": 4.4805, "learning_rate": 1.6657894736842105e-05, "loss": 3.3752, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.6921, "num_token_overlap": 11.684, "num_token_query": 31.4136, "num_token_union": 65.0807, "num_word_context": 202.1604, "num_word_doc": 49.7564, "num_word_query": 23.3381, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5275.5024, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.8711, "query_norm": 2.3155, "queue_k_norm": 2.3888, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4136, "sent_len_1": 66.6921, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.985, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136700 }, { "accuracy": 52.7344, "active_queue_size": 16384.0, "cl_loss": 3.3783, "doc_norm": 2.3883, "encoder_q-embeddings": 2516.9956, "encoder_q-layer.0": 1792.4955, "encoder_q-layer.1": 1979.465, "encoder_q-layer.10": 3525.9209, "encoder_q-layer.11": 7572.7617, "encoder_q-layer.2": 2292.8342, "encoder_q-layer.3": 2470.7007, "encoder_q-layer.4": 2608.293, "encoder_q-layer.5": 2716.1255, "encoder_q-layer.6": 2895.6514, "encoder_q-layer.7": 3374.4375, "encoder_q-layer.8": 3622.4106, "encoder_q-layer.9": 3098.4854, "epoch": 0.59, "inbatch_neg_score": 3.8741, "inbatch_pos_score": 4.5, "learning_rate": 1.6631578947368423e-05, "loss": 3.3783, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.8535, "num_token_overlap": 11.714, "num_token_query": 31.3536, "num_token_union": 65.1348, "num_word_context": 202.2771, "num_word_doc": 49.8957, "num_word_query": 23.2905, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5125.548, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.873, "query_norm": 2.3196, "queue_k_norm": 2.3908, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3536, "sent_len_1": 66.8535, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.05, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 136800 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.3823, "doc_norm": 2.3888, "encoder_q-embeddings": 2541.9873, "encoder_q-layer.0": 1838.3162, "encoder_q-layer.1": 1928.9355, "encoder_q-layer.10": 3133.4771, "encoder_q-layer.11": 7770.8735, "encoder_q-layer.2": 2209.7446, "encoder_q-layer.3": 2224.1125, "encoder_q-layer.4": 2450.7708, "encoder_q-layer.5": 2330.0894, "encoder_q-layer.6": 2530.2976, "encoder_q-layer.7": 2903.6077, "encoder_q-layer.8": 3302.3301, "encoder_q-layer.9": 2977.4399, "epoch": 0.59, "inbatch_neg_score": 3.8749, "inbatch_pos_score": 4.4844, "learning_rate": 1.6605263157894738e-05, "loss": 3.3823, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.7623, "num_token_overlap": 11.6612, "num_token_query": 31.3776, "num_token_union": 65.1408, "num_word_context": 202.2268, "num_word_doc": 49.8311, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4969.8991, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 3.875, "query_norm": 2.3201, "queue_k_norm": 2.3911, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3776, "sent_len_1": 66.7623, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.65, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 136900 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3816, "doc_norm": 2.3952, "encoder_q-embeddings": 5166.6523, "encoder_q-layer.0": 3908.8406, "encoder_q-layer.1": 3975.3022, "encoder_q-layer.10": 6849.0986, "encoder_q-layer.11": 16248.1621, "encoder_q-layer.2": 4393.9146, "encoder_q-layer.3": 4386.9604, "encoder_q-layer.4": 4736.8003, "encoder_q-layer.5": 4870.4346, "encoder_q-layer.6": 5388.2598, "encoder_q-layer.7": 5967.5854, "encoder_q-layer.8": 7285.3892, "encoder_q-layer.9": 6475.1958, "epoch": 0.59, "inbatch_neg_score": 3.885, "inbatch_pos_score": 4.4922, "learning_rate": 1.6578947368421053e-05, "loss": 3.3816, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.9076, "num_token_overlap": 11.6967, "num_token_query": 31.4122, "num_token_union": 65.2155, "num_word_context": 202.5987, "num_word_doc": 49.9098, "num_word_query": 23.3212, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10502.1854, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8828, "query_norm": 2.3231, "queue_k_norm": 2.3921, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4122, "sent_len_1": 66.9076, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.8462, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137000 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3779, "doc_norm": 2.3929, "encoder_q-embeddings": 4694.3174, "encoder_q-layer.0": 3278.8518, "encoder_q-layer.1": 3516.092, "encoder_q-layer.10": 6151.625, "encoder_q-layer.11": 15224.6582, "encoder_q-layer.2": 3920.7463, "encoder_q-layer.3": 4093.5181, "encoder_q-layer.4": 4324.7563, "encoder_q-layer.5": 4553.8789, "encoder_q-layer.6": 5142.4214, "encoder_q-layer.7": 5395.3916, "encoder_q-layer.8": 6420.9258, "encoder_q-layer.9": 5872.6519, "epoch": 0.59, "inbatch_neg_score": 3.8858, "inbatch_pos_score": 4.5117, "learning_rate": 1.6552631578947368e-05, "loss": 3.3779, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.6939, "num_token_overlap": 11.6719, "num_token_query": 31.4251, "num_token_union": 65.1021, "num_word_context": 202.2221, "num_word_doc": 49.7553, "num_word_query": 23.3476, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9471.2088, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8848, "query_norm": 2.3274, "queue_k_norm": 2.3926, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4251, "sent_len_1": 66.6939, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.0613, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137100 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3682, "doc_norm": 2.3929, "encoder_q-embeddings": 4830.8057, "encoder_q-layer.0": 3308.2888, "encoder_q-layer.1": 3496.6499, "encoder_q-layer.10": 6644.3091, "encoder_q-layer.11": 16636.3965, "encoder_q-layer.2": 3980.4265, "encoder_q-layer.3": 4140.4453, "encoder_q-layer.4": 4559.6348, "encoder_q-layer.5": 4668.4248, "encoder_q-layer.6": 5184.4448, "encoder_q-layer.7": 6061.0474, "encoder_q-layer.8": 7067.4111, "encoder_q-layer.9": 6513.564, "epoch": 0.6, "inbatch_neg_score": 3.8881, "inbatch_pos_score": 4.4922, "learning_rate": 1.6526315789473683e-05, "loss": 3.3682, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.701, "num_token_overlap": 11.6942, "num_token_query": 31.4137, "num_token_union": 65.0858, "num_word_context": 201.9656, "num_word_doc": 49.745, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10271.5735, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8867, "query_norm": 2.3242, "queue_k_norm": 2.3937, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4137, "sent_len_1": 66.701, "sent_len_max_0": 127.9613, "sent_len_max_1": 210.5925, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137200 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3678, "doc_norm": 2.3959, "encoder_q-embeddings": 4883.1016, "encoder_q-layer.0": 3329.8181, "encoder_q-layer.1": 3491.8716, "encoder_q-layer.10": 6724.8394, "encoder_q-layer.11": 16554.9824, "encoder_q-layer.2": 3910.7739, "encoder_q-layer.3": 3962.5889, "encoder_q-layer.4": 4239.6318, "encoder_q-layer.5": 4412.6191, "encoder_q-layer.6": 5158.8447, "encoder_q-layer.7": 5527.4912, "encoder_q-layer.8": 6862.8765, "encoder_q-layer.9": 6455.4619, "epoch": 0.6, "inbatch_neg_score": 3.8963, "inbatch_pos_score": 4.5078, "learning_rate": 1.65e-05, "loss": 3.3678, "norm_diff": 0.0708, "norm_loss": 0.0, "num_token_doc": 66.8026, "num_token_overlap": 11.7173, "num_token_query": 31.3823, "num_token_union": 65.1043, "num_word_context": 202.3947, "num_word_doc": 49.8492, "num_word_query": 23.3037, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10274.5158, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.8965, "query_norm": 2.3251, "queue_k_norm": 2.3954, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3823, "sent_len_1": 66.8026, "sent_len_max_0": 127.985, "sent_len_max_1": 207.2862, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 137300 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3819, "doc_norm": 2.3956, "encoder_q-embeddings": 5036.937, "encoder_q-layer.0": 3551.2278, "encoder_q-layer.1": 3831.0095, "encoder_q-layer.10": 6860.9341, "encoder_q-layer.11": 16105.7363, "encoder_q-layer.2": 4317.0942, "encoder_q-layer.3": 4341.2803, "encoder_q-layer.4": 4811.5874, "encoder_q-layer.5": 4523.4111, "encoder_q-layer.6": 5170.2485, "encoder_q-layer.7": 5983.2568, "encoder_q-layer.8": 7265.4526, "encoder_q-layer.9": 6303.2246, "epoch": 0.6, "inbatch_neg_score": 3.9001, "inbatch_pos_score": 4.5078, "learning_rate": 1.6473684210526316e-05, "loss": 3.3819, "norm_diff": 0.0689, "norm_loss": 0.0, "num_token_doc": 66.6466, "num_token_overlap": 11.6398, "num_token_query": 31.3136, "num_token_union": 65.0486, "num_word_context": 202.0745, "num_word_doc": 49.7371, "num_word_query": 23.2614, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10258.2686, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9004, "query_norm": 2.3267, "queue_k_norm": 2.3959, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3136, "sent_len_1": 66.6466, "sent_len_max_0": 127.99, "sent_len_max_1": 207.8825, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137400 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3717, "doc_norm": 2.3993, "encoder_q-embeddings": 4731.0112, "encoder_q-layer.0": 3300.8823, "encoder_q-layer.1": 3583.7852, "encoder_q-layer.10": 6297.585, "encoder_q-layer.11": 15748.9619, "encoder_q-layer.2": 4013.8579, "encoder_q-layer.3": 4142.2861, "encoder_q-layer.4": 4473.8574, "encoder_q-layer.5": 4880.4917, "encoder_q-layer.6": 5365.1035, "encoder_q-layer.7": 6266.3291, "encoder_q-layer.8": 6958.1919, "encoder_q-layer.9": 6326.2241, "epoch": 0.6, "inbatch_neg_score": 3.9032, "inbatch_pos_score": 4.5156, "learning_rate": 1.6447368421052635e-05, "loss": 3.3717, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7551, "num_token_overlap": 11.6985, "num_token_query": 31.4271, "num_token_union": 65.1421, "num_word_context": 202.5008, "num_word_doc": 49.8397, "num_word_query": 23.3411, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10035.9464, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9023, "query_norm": 2.3298, "queue_k_norm": 2.3967, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4271, "sent_len_1": 66.7551, "sent_len_max_0": 127.9838, "sent_len_max_1": 206.56, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137500 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.381, "doc_norm": 2.3984, "encoder_q-embeddings": 6075.8379, "encoder_q-layer.0": 4360.9829, "encoder_q-layer.1": 4809.4741, "encoder_q-layer.10": 6349.4624, "encoder_q-layer.11": 15176.2363, "encoder_q-layer.2": 5373.7007, "encoder_q-layer.3": 5579.7432, "encoder_q-layer.4": 6591.9272, "encoder_q-layer.5": 6725.4141, "encoder_q-layer.6": 6977.5166, "encoder_q-layer.7": 6707.7163, "encoder_q-layer.8": 7057.103, "encoder_q-layer.9": 6430.1821, "epoch": 0.6, "inbatch_neg_score": 3.9111, "inbatch_pos_score": 4.5156, "learning_rate": 1.642105263157895e-05, "loss": 3.381, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.7738, "num_token_overlap": 11.6391, "num_token_query": 31.2886, "num_token_union": 65.0832, "num_word_context": 202.3532, "num_word_doc": 49.8658, "num_word_query": 23.2369, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11006.9404, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9102, "query_norm": 2.3303, "queue_k_norm": 2.3968, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2886, "sent_len_1": 66.7738, "sent_len_max_0": 127.96, "sent_len_max_1": 207.7587, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137600 }, { "accuracy": 48.1934, "active_queue_size": 16384.0, "cl_loss": 3.3788, "doc_norm": 2.3979, "encoder_q-embeddings": 4739.2778, "encoder_q-layer.0": 3449.3044, "encoder_q-layer.1": 3744.3394, "encoder_q-layer.10": 7462.9766, "encoder_q-layer.11": 16671.4785, "encoder_q-layer.2": 4124.7134, "encoder_q-layer.3": 4182.3633, "encoder_q-layer.4": 4464.8062, "encoder_q-layer.5": 4438.5815, "encoder_q-layer.6": 5087.9663, "encoder_q-layer.7": 6209.5176, "encoder_q-layer.8": 7176.4761, "encoder_q-layer.9": 6749.8203, "epoch": 0.6, "inbatch_neg_score": 3.9195, "inbatch_pos_score": 4.5117, "learning_rate": 1.639473684210526e-05, "loss": 3.3788, "norm_diff": 0.0652, "norm_loss": 0.0, "num_token_doc": 66.7959, "num_token_overlap": 11.6861, "num_token_query": 31.357, "num_token_union": 65.1323, "num_word_context": 202.2947, "num_word_doc": 49.8617, "num_word_query": 23.2855, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10357.8695, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.918, "query_norm": 2.3326, "queue_k_norm": 2.3985, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.357, "sent_len_1": 66.7959, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.2063, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137700 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.369, "doc_norm": 2.3999, "encoder_q-embeddings": 5677.4839, "encoder_q-layer.0": 3883.9292, "encoder_q-layer.1": 4345.019, "encoder_q-layer.10": 6692.1382, "encoder_q-layer.11": 15845.4453, "encoder_q-layer.2": 4868.8257, "encoder_q-layer.3": 5211.7598, "encoder_q-layer.4": 5425.2461, "encoder_q-layer.5": 5292.5469, "encoder_q-layer.6": 5645.6606, "encoder_q-layer.7": 6198.4302, "encoder_q-layer.8": 6971.5366, "encoder_q-layer.9": 5984.4927, "epoch": 0.6, "inbatch_neg_score": 3.9251, "inbatch_pos_score": 4.5312, "learning_rate": 1.636842105263158e-05, "loss": 3.369, "norm_diff": 0.064, "norm_loss": 0.0, "num_token_doc": 66.678, "num_token_overlap": 11.7047, "num_token_query": 31.429, "num_token_union": 65.0812, "num_word_context": 202.3275, "num_word_doc": 49.758, "num_word_query": 23.3323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10457.5082, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9258, "query_norm": 2.3359, "queue_k_norm": 2.3995, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.429, "sent_len_1": 66.678, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.6375, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 137800 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3851, "doc_norm": 2.4014, "encoder_q-embeddings": 5279.6304, "encoder_q-layer.0": 3793.6443, "encoder_q-layer.1": 3918.656, "encoder_q-layer.10": 6479.0488, "encoder_q-layer.11": 16068.4414, "encoder_q-layer.2": 4510.4492, "encoder_q-layer.3": 4712.6738, "encoder_q-layer.4": 5307.7148, "encoder_q-layer.5": 5733.2632, "encoder_q-layer.6": 6250.605, "encoder_q-layer.7": 6696.9316, "encoder_q-layer.8": 7299.2554, "encoder_q-layer.9": 6352.7324, "epoch": 0.6, "inbatch_neg_score": 3.9313, "inbatch_pos_score": 4.5352, "learning_rate": 1.6342105263157894e-05, "loss": 3.3851, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.7728, "num_token_overlap": 11.6746, "num_token_query": 31.3762, "num_token_union": 65.13, "num_word_context": 202.2479, "num_word_doc": 49.8115, "num_word_query": 23.3039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10748.9244, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9316, "query_norm": 2.3327, "queue_k_norm": 2.4009, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3762, "sent_len_1": 66.7728, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.115, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 137900 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3821, "doc_norm": 2.4018, "encoder_q-embeddings": 4552.7168, "encoder_q-layer.0": 3251.9856, "encoder_q-layer.1": 3507.5405, "encoder_q-layer.10": 6974.8594, "encoder_q-layer.11": 15675.4932, "encoder_q-layer.2": 3979.6616, "encoder_q-layer.3": 4077.1323, "encoder_q-layer.4": 4245.939, "encoder_q-layer.5": 4506.9873, "encoder_q-layer.6": 5078.0215, "encoder_q-layer.7": 5933.9033, "encoder_q-layer.8": 7330.2573, "encoder_q-layer.9": 6587.6577, "epoch": 0.6, "inbatch_neg_score": 3.9414, "inbatch_pos_score": 4.5664, "learning_rate": 1.6315789473684213e-05, "loss": 3.3821, "norm_diff": 0.0564, "norm_loss": 0.0, "num_token_doc": 66.6074, "num_token_overlap": 11.6618, "num_token_query": 31.3951, "num_token_union": 65.0558, "num_word_context": 202.1966, "num_word_doc": 49.7622, "num_word_query": 23.3369, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9848.6402, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9414, "query_norm": 2.3454, "queue_k_norm": 2.4015, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3951, "sent_len_1": 66.6074, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.1337, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 138000 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3712, "doc_norm": 2.4008, "encoder_q-embeddings": 4767.4248, "encoder_q-layer.0": 3340.4612, "encoder_q-layer.1": 3590.022, "encoder_q-layer.10": 6381.2056, "encoder_q-layer.11": 15520.5059, "encoder_q-layer.2": 4134.2891, "encoder_q-layer.3": 4504.7803, "encoder_q-layer.4": 4796.6201, "encoder_q-layer.5": 4797.7832, "encoder_q-layer.6": 5210.6528, "encoder_q-layer.7": 5722.6753, "encoder_q-layer.8": 6662.2461, "encoder_q-layer.9": 6054.3638, "epoch": 0.6, "inbatch_neg_score": 3.9517, "inbatch_pos_score": 4.5781, "learning_rate": 1.6289473684210528e-05, "loss": 3.3712, "norm_diff": 0.0498, "norm_loss": 0.0, "num_token_doc": 66.7186, "num_token_overlap": 11.7278, "num_token_query": 31.5167, "num_token_union": 65.1579, "num_word_context": 202.2618, "num_word_doc": 49.7962, "num_word_query": 23.4025, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9999.0142, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9512, "query_norm": 2.351, "queue_k_norm": 2.4017, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5167, "sent_len_1": 66.7186, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.655, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 138100 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3873, "doc_norm": 2.4046, "encoder_q-embeddings": 4662.7568, "encoder_q-layer.0": 3302.0461, "encoder_q-layer.1": 3663.8704, "encoder_q-layer.10": 6218.3809, "encoder_q-layer.11": 15283.6709, "encoder_q-layer.2": 3908.0061, "encoder_q-layer.3": 4189.6655, "encoder_q-layer.4": 4474.0107, "encoder_q-layer.5": 4644.4829, "encoder_q-layer.6": 5025.019, "encoder_q-layer.7": 5782.2485, "encoder_q-layer.8": 6573.8818, "encoder_q-layer.9": 6129.165, "epoch": 0.6, "inbatch_neg_score": 3.9618, "inbatch_pos_score": 4.5781, "learning_rate": 1.6263157894736843e-05, "loss": 3.3873, "norm_diff": 0.0476, "norm_loss": 0.0, "num_token_doc": 66.6566, "num_token_overlap": 11.667, "num_token_query": 31.3613, "num_token_union": 65.0597, "num_word_context": 202.1786, "num_word_doc": 49.7389, "num_word_query": 23.2922, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9776.8806, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9629, "query_norm": 2.357, "queue_k_norm": 2.4046, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3613, "sent_len_1": 66.6566, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.7575, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 138200 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3821, "doc_norm": 2.4046, "encoder_q-embeddings": 4906.5454, "encoder_q-layer.0": 3412.5996, "encoder_q-layer.1": 3621.5488, "encoder_q-layer.10": 6535.356, "encoder_q-layer.11": 15734.6484, "encoder_q-layer.2": 4055.9897, "encoder_q-layer.3": 4224.0117, "encoder_q-layer.4": 4339.5908, "encoder_q-layer.5": 4407.1978, "encoder_q-layer.6": 5008.459, "encoder_q-layer.7": 5658.7739, "encoder_q-layer.8": 6580.5508, "encoder_q-layer.9": 6135.7134, "epoch": 0.6, "inbatch_neg_score": 3.9724, "inbatch_pos_score": 4.582, "learning_rate": 1.6236842105263158e-05, "loss": 3.3821, "norm_diff": 0.0529, "norm_loss": 0.0, "num_token_doc": 66.8515, "num_token_overlap": 11.6935, "num_token_query": 31.3963, "num_token_union": 65.1702, "num_word_context": 202.2918, "num_word_doc": 49.8896, "num_word_query": 23.3182, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9832.2632, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9727, "query_norm": 2.3517, "queue_k_norm": 2.4048, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3963, "sent_len_1": 66.8515, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.5863, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 138300 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3721, "doc_norm": 2.409, "encoder_q-embeddings": 7171.1572, "encoder_q-layer.0": 5363.1338, "encoder_q-layer.1": 5765.2432, "encoder_q-layer.10": 6144.4307, "encoder_q-layer.11": 15766.3672, "encoder_q-layer.2": 7085.7153, "encoder_q-layer.3": 7554.978, "encoder_q-layer.4": 8174.1738, "encoder_q-layer.5": 8964.8457, "encoder_q-layer.6": 7947.0161, "encoder_q-layer.7": 7564.6143, "encoder_q-layer.8": 7248.5522, "encoder_q-layer.9": 6186.3984, "epoch": 0.6, "inbatch_neg_score": 3.9839, "inbatch_pos_score": 4.5938, "learning_rate": 1.6210526315789473e-05, "loss": 3.3721, "norm_diff": 0.0515, "norm_loss": 0.0, "num_token_doc": 66.7841, "num_token_overlap": 11.6873, "num_token_query": 31.3982, "num_token_union": 65.1318, "num_word_context": 202.5057, "num_word_doc": 49.857, "num_word_query": 23.3163, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12486.1481, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9824, "query_norm": 2.3575, "queue_k_norm": 2.4073, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3982, "sent_len_1": 66.7841, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.3812, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 138400 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3721, "doc_norm": 2.408, "encoder_q-embeddings": 4786.4429, "encoder_q-layer.0": 3389.6006, "encoder_q-layer.1": 3613.3542, "encoder_q-layer.10": 6189.9209, "encoder_q-layer.11": 15091.3525, "encoder_q-layer.2": 4024.0183, "encoder_q-layer.3": 4152.7329, "encoder_q-layer.4": 4317.1997, "encoder_q-layer.5": 4444.2554, "encoder_q-layer.6": 4843.3086, "encoder_q-layer.7": 5767.2959, "encoder_q-layer.8": 6705.999, "encoder_q-layer.9": 6037.0356, "epoch": 0.6, "inbatch_neg_score": 3.9943, "inbatch_pos_score": 4.6016, "learning_rate": 1.618421052631579e-05, "loss": 3.3721, "norm_diff": 0.0452, "norm_loss": 0.0, "num_token_doc": 66.8625, "num_token_overlap": 11.7036, "num_token_query": 31.3724, "num_token_union": 65.1595, "num_word_context": 202.2446, "num_word_doc": 49.894, "num_word_query": 23.3058, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9725.7202, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 3.9941, "query_norm": 2.3628, "queue_k_norm": 2.4074, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3724, "sent_len_1": 66.8625, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.7512, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 138500 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3772, "doc_norm": 2.4088, "encoder_q-embeddings": 5069.6992, "encoder_q-layer.0": 3546.5007, "encoder_q-layer.1": 3779.093, "encoder_q-layer.10": 6650.293, "encoder_q-layer.11": 16609.0312, "encoder_q-layer.2": 4244.6123, "encoder_q-layer.3": 4158.9844, "encoder_q-layer.4": 4483.0337, "encoder_q-layer.5": 4587.4927, "encoder_q-layer.6": 5194.2349, "encoder_q-layer.7": 6043.8711, "encoder_q-layer.8": 7297.0522, "encoder_q-layer.9": 6504.2007, "epoch": 0.6, "inbatch_neg_score": 4.0033, "inbatch_pos_score": 4.6016, "learning_rate": 1.6157894736842106e-05, "loss": 3.3772, "norm_diff": 0.048, "norm_loss": 0.0, "num_token_doc": 66.8367, "num_token_overlap": 11.6816, "num_token_query": 31.4122, "num_token_union": 65.1841, "num_word_context": 202.41, "num_word_doc": 49.8561, "num_word_query": 23.3358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10592.5743, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0039, "query_norm": 2.3609, "queue_k_norm": 2.41, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4122, "sent_len_1": 66.8367, "sent_len_max_0": 127.9988, "sent_len_max_1": 210.0575, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 138600 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3807, "doc_norm": 2.4103, "encoder_q-embeddings": 4784.6841, "encoder_q-layer.0": 3459.7214, "encoder_q-layer.1": 3634.0688, "encoder_q-layer.10": 6277.1064, "encoder_q-layer.11": 15471.7744, "encoder_q-layer.2": 4082.2139, "encoder_q-layer.3": 4145.6318, "encoder_q-layer.4": 4458.6523, "encoder_q-layer.5": 4489.7271, "encoder_q-layer.6": 4803.1572, "encoder_q-layer.7": 5427.897, "encoder_q-layer.8": 6343.7437, "encoder_q-layer.9": 6031.8354, "epoch": 0.6, "inbatch_neg_score": 4.0091, "inbatch_pos_score": 4.6133, "learning_rate": 1.613157894736842e-05, "loss": 3.3807, "norm_diff": 0.0538, "norm_loss": 0.0, "num_token_doc": 66.8663, "num_token_overlap": 11.6851, "num_token_query": 31.3855, "num_token_union": 65.1569, "num_word_context": 202.1184, "num_word_doc": 49.868, "num_word_query": 23.3306, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9890.4131, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0078, "query_norm": 2.3565, "queue_k_norm": 2.4109, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3855, "sent_len_1": 66.8663, "sent_len_max_0": 127.995, "sent_len_max_1": 209.025, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 138700 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3815, "doc_norm": 2.412, "encoder_q-embeddings": 4944.0649, "encoder_q-layer.0": 3368.2637, "encoder_q-layer.1": 3493.3672, "encoder_q-layer.10": 6622.2617, "encoder_q-layer.11": 16245.2197, "encoder_q-layer.2": 3849.51, "encoder_q-layer.3": 3999.7119, "encoder_q-layer.4": 4257.8755, "encoder_q-layer.5": 4547.3823, "encoder_q-layer.6": 4907.0107, "encoder_q-layer.7": 5786.7607, "encoder_q-layer.8": 6904.0962, "encoder_q-layer.9": 6191.9067, "epoch": 0.6, "inbatch_neg_score": 4.0146, "inbatch_pos_score": 4.625, "learning_rate": 1.6105263157894736e-05, "loss": 3.3815, "norm_diff": 0.0556, "norm_loss": 0.0, "num_token_doc": 66.8063, "num_token_overlap": 11.6442, "num_token_query": 31.2903, "num_token_union": 65.1332, "num_word_context": 202.3485, "num_word_doc": 49.8276, "num_word_query": 23.245, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10259.1529, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0156, "query_norm": 2.3564, "queue_k_norm": 2.4138, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2903, "sent_len_1": 66.8063, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.9787, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 138800 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3746, "doc_norm": 2.4124, "encoder_q-embeddings": 4653.8457, "encoder_q-layer.0": 3253.8193, "encoder_q-layer.1": 3483.998, "encoder_q-layer.10": 6736.8691, "encoder_q-layer.11": 16393.0684, "encoder_q-layer.2": 3784.2239, "encoder_q-layer.3": 3859.0723, "encoder_q-layer.4": 4217.3374, "encoder_q-layer.5": 4409.9717, "encoder_q-layer.6": 4902.8423, "encoder_q-layer.7": 5636.8833, "encoder_q-layer.8": 6927.23, "encoder_q-layer.9": 6275.1094, "epoch": 0.6, "inbatch_neg_score": 4.0225, "inbatch_pos_score": 4.6211, "learning_rate": 1.6078947368421054e-05, "loss": 3.3746, "norm_diff": 0.0564, "norm_loss": 0.0, "num_token_doc": 66.812, "num_token_overlap": 11.7046, "num_token_query": 31.4097, "num_token_union": 65.1586, "num_word_context": 202.3132, "num_word_doc": 49.8485, "num_word_query": 23.3351, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10131.3521, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0234, "query_norm": 2.356, "queue_k_norm": 2.4154, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4097, "sent_len_1": 66.812, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.5412, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 138900 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3797, "doc_norm": 2.4145, "encoder_q-embeddings": 10141.5039, "encoder_q-layer.0": 7022.7158, "encoder_q-layer.1": 7340.1533, "encoder_q-layer.10": 13451.0439, "encoder_q-layer.11": 31488.498, "encoder_q-layer.2": 8207.4854, "encoder_q-layer.3": 8489.8594, "encoder_q-layer.4": 9091.2988, "encoder_q-layer.5": 9631.6592, "encoder_q-layer.6": 11129.7441, "encoder_q-layer.7": 12210.7754, "encoder_q-layer.8": 14054.1074, "encoder_q-layer.9": 12913.6211, "epoch": 0.6, "inbatch_neg_score": 4.0271, "inbatch_pos_score": 4.6445, "learning_rate": 1.605263157894737e-05, "loss": 3.3797, "norm_diff": 0.0558, "norm_loss": 0.0, "num_token_doc": 66.6273, "num_token_overlap": 11.666, "num_token_query": 31.4571, "num_token_union": 65.0891, "num_word_context": 202.5521, "num_word_doc": 49.7303, "num_word_query": 23.3577, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20475.8072, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.0273, "query_norm": 2.3587, "queue_k_norm": 2.4157, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4571, "sent_len_1": 66.6273, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.4725, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 139000 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3758, "doc_norm": 2.4167, "encoder_q-embeddings": 9184.9053, "encoder_q-layer.0": 6545.0171, "encoder_q-layer.1": 7158.5903, "encoder_q-layer.10": 12867.4824, "encoder_q-layer.11": 31690.3223, "encoder_q-layer.2": 8186.332, "encoder_q-layer.3": 8418.0381, "encoder_q-layer.4": 8992.5801, "encoder_q-layer.5": 9289.4141, "encoder_q-layer.6": 10265.3662, "encoder_q-layer.7": 11507.2129, "encoder_q-layer.8": 13171.8428, "encoder_q-layer.9": 12322.042, "epoch": 0.6, "inbatch_neg_score": 4.0285, "inbatch_pos_score": 4.6484, "learning_rate": 1.6026315789473684e-05, "loss": 3.3758, "norm_diff": 0.059, "norm_loss": 0.0, "num_token_doc": 66.713, "num_token_overlap": 11.6769, "num_token_query": 31.4101, "num_token_union": 65.135, "num_word_context": 202.3157, "num_word_doc": 49.7895, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20088.5657, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.0273, "query_norm": 2.3578, "queue_k_norm": 2.4171, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4101, "sent_len_1": 66.713, "sent_len_max_0": 128.0, "sent_len_max_1": 207.8325, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 139100 }, { "accuracy": 52.7344, "active_queue_size": 16384.0, "cl_loss": 3.3863, "doc_norm": 2.4164, "encoder_q-embeddings": 4581.5947, "encoder_q-layer.0": 3230.4622, "encoder_q-layer.1": 3451.6301, "encoder_q-layer.10": 6367.2568, "encoder_q-layer.11": 15599.9863, "encoder_q-layer.2": 3754.4385, "encoder_q-layer.3": 3949.3926, "encoder_q-layer.4": 4227.8867, "encoder_q-layer.5": 4283.5659, "encoder_q-layer.6": 4855.2583, "encoder_q-layer.7": 5476.4194, "encoder_q-layer.8": 6322.4795, "encoder_q-layer.9": 6048.8716, "epoch": 0.6, "inbatch_neg_score": 4.0331, "inbatch_pos_score": 4.6484, "learning_rate": 1.6000000000000003e-05, "loss": 3.3863, "norm_diff": 0.0653, "norm_loss": 0.0, "num_token_doc": 66.7263, "num_token_overlap": 11.6523, "num_token_query": 31.3363, "num_token_union": 65.094, "num_word_context": 202.1983, "num_word_doc": 49.8, "num_word_query": 23.2526, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9749.9885, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0352, "query_norm": 2.3511, "queue_k_norm": 2.4189, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3363, "sent_len_1": 66.7263, "sent_len_max_0": 127.965, "sent_len_max_1": 208.1975, "stdk": 0.0493, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 139200 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3887, "doc_norm": 2.4192, "encoder_q-embeddings": 4645.3301, "encoder_q-layer.0": 3268.9558, "encoder_q-layer.1": 3506.0791, "encoder_q-layer.10": 7226.9409, "encoder_q-layer.11": 17266.2344, "encoder_q-layer.2": 3970.8013, "encoder_q-layer.3": 4164.2021, "encoder_q-layer.4": 4505.1724, "encoder_q-layer.5": 4572.8037, "encoder_q-layer.6": 5076.687, "encoder_q-layer.7": 6046.2686, "encoder_q-layer.8": 7062.0483, "encoder_q-layer.9": 6827.3574, "epoch": 0.6, "inbatch_neg_score": 4.0383, "inbatch_pos_score": 4.6367, "learning_rate": 1.5973684210526314e-05, "loss": 3.3887, "norm_diff": 0.0596, "norm_loss": 0.0, "num_token_doc": 66.8687, "num_token_overlap": 11.6564, "num_token_query": 31.414, "num_token_union": 65.2181, "num_word_context": 202.5746, "num_word_doc": 49.8921, "num_word_query": 23.3354, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10580.531, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0352, "query_norm": 2.3596, "queue_k_norm": 2.4202, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.414, "sent_len_1": 66.8687, "sent_len_max_0": 127.9387, "sent_len_max_1": 208.8475, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 139300 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.372, "doc_norm": 2.4201, "encoder_q-embeddings": 4814.9839, "encoder_q-layer.0": 3302.8875, "encoder_q-layer.1": 3591.6072, "encoder_q-layer.10": 7382.3916, "encoder_q-layer.11": 17548.0605, "encoder_q-layer.2": 3955.9451, "encoder_q-layer.3": 4089.5469, "encoder_q-layer.4": 4462.8955, "encoder_q-layer.5": 4455.4961, "encoder_q-layer.6": 5071.2959, "encoder_q-layer.7": 6189.6646, "encoder_q-layer.8": 7641.3774, "encoder_q-layer.9": 7086.9712, "epoch": 0.6, "inbatch_neg_score": 4.0384, "inbatch_pos_score": 4.668, "learning_rate": 1.5947368421052633e-05, "loss": 3.372, "norm_diff": 0.0577, "norm_loss": 0.0, "num_token_doc": 66.8548, "num_token_overlap": 11.7211, "num_token_query": 31.499, "num_token_union": 65.1963, "num_word_context": 202.0145, "num_word_doc": 49.8643, "num_word_query": 23.4015, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10613.6862, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0391, "query_norm": 2.3625, "queue_k_norm": 2.422, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.499, "sent_len_1": 66.8548, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.8713, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 139400 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3868, "doc_norm": 2.4237, "encoder_q-embeddings": 4892.2354, "encoder_q-layer.0": 3391.3462, "encoder_q-layer.1": 3605.1355, "encoder_q-layer.10": 6788.1138, "encoder_q-layer.11": 16751.082, "encoder_q-layer.2": 3996.8557, "encoder_q-layer.3": 3969.7644, "encoder_q-layer.4": 4227.6948, "encoder_q-layer.5": 4285.3809, "encoder_q-layer.6": 4845.1025, "encoder_q-layer.7": 5541.2173, "encoder_q-layer.8": 7116.1807, "encoder_q-layer.9": 6392.8984, "epoch": 0.61, "inbatch_neg_score": 4.0396, "inbatch_pos_score": 4.6523, "learning_rate": 1.5921052631578948e-05, "loss": 3.3868, "norm_diff": 0.0677, "norm_loss": 0.0, "num_token_doc": 66.7912, "num_token_overlap": 11.6367, "num_token_query": 31.2521, "num_token_union": 65.0743, "num_word_context": 202.4762, "num_word_doc": 49.8531, "num_word_query": 23.2089, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10397.3753, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0391, "query_norm": 2.356, "queue_k_norm": 2.4223, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2521, "sent_len_1": 66.7912, "sent_len_max_0": 127.975, "sent_len_max_1": 207.0975, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 139500 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3864, "doc_norm": 2.4248, "encoder_q-embeddings": 5166.9131, "encoder_q-layer.0": 3507.0664, "encoder_q-layer.1": 3667.1714, "encoder_q-layer.10": 7147.2734, "encoder_q-layer.11": 16774.9922, "encoder_q-layer.2": 4167.6113, "encoder_q-layer.3": 4138.9517, "encoder_q-layer.4": 4413.4858, "encoder_q-layer.5": 4471.7563, "encoder_q-layer.6": 5116.0581, "encoder_q-layer.7": 5779.896, "encoder_q-layer.8": 6778.1504, "encoder_q-layer.9": 6667.0913, "epoch": 0.61, "inbatch_neg_score": 4.0457, "inbatch_pos_score": 4.6641, "learning_rate": 1.5894736842105266e-05, "loss": 3.3864, "norm_diff": 0.0653, "norm_loss": 0.0, "num_token_doc": 66.7962, "num_token_overlap": 11.667, "num_token_query": 31.3346, "num_token_union": 65.1003, "num_word_context": 202.1182, "num_word_doc": 49.8258, "num_word_query": 23.2674, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10348.2431, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0469, "query_norm": 2.3596, "queue_k_norm": 2.423, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3346, "sent_len_1": 66.7962, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.6813, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 139600 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3852, "doc_norm": 2.4249, "encoder_q-embeddings": 7043.7715, "encoder_q-layer.0": 5050.5439, "encoder_q-layer.1": 5445.2188, "encoder_q-layer.10": 6404.1343, "encoder_q-layer.11": 15702.9717, "encoder_q-layer.2": 6255.6865, "encoder_q-layer.3": 6621.0874, "encoder_q-layer.4": 7271.3691, "encoder_q-layer.5": 7469.4209, "encoder_q-layer.6": 8432.7295, "encoder_q-layer.7": 8261.4971, "encoder_q-layer.8": 8427.9209, "encoder_q-layer.9": 6813.7842, "epoch": 0.61, "inbatch_neg_score": 4.0478, "inbatch_pos_score": 4.6562, "learning_rate": 1.586842105263158e-05, "loss": 3.3852, "norm_diff": 0.0664, "norm_loss": 0.0, "num_token_doc": 66.8466, "num_token_overlap": 11.6931, "num_token_query": 31.4747, "num_token_union": 65.2062, "num_word_context": 202.5182, "num_word_doc": 49.8675, "num_word_query": 23.3852, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12274.0186, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0469, "query_norm": 2.3584, "queue_k_norm": 2.4238, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4747, "sent_len_1": 66.8466, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.1, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 139700 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3768, "doc_norm": 2.4252, "encoder_q-embeddings": 4914.918, "encoder_q-layer.0": 3541.3796, "encoder_q-layer.1": 3791.6743, "encoder_q-layer.10": 6800.1987, "encoder_q-layer.11": 16443.877, "encoder_q-layer.2": 4172.1035, "encoder_q-layer.3": 4285.2539, "encoder_q-layer.4": 4566.8433, "encoder_q-layer.5": 4749.3032, "encoder_q-layer.6": 5457.3511, "encoder_q-layer.7": 6739.1855, "encoder_q-layer.8": 7559.312, "encoder_q-layer.9": 6557.8193, "epoch": 0.61, "inbatch_neg_score": 4.0495, "inbatch_pos_score": 4.6562, "learning_rate": 1.5842105263157896e-05, "loss": 3.3768, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.6917, "num_token_overlap": 11.6584, "num_token_query": 31.3496, "num_token_union": 65.0617, "num_word_context": 202.1358, "num_word_doc": 49.7604, "num_word_query": 23.2812, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10552.1151, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0508, "query_norm": 2.3552, "queue_k_norm": 2.4255, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3496, "sent_len_1": 66.6917, "sent_len_max_0": 127.9688, "sent_len_max_1": 211.3812, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 139800 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3658, "doc_norm": 2.4259, "encoder_q-embeddings": 4613.9399, "encoder_q-layer.0": 3187.2559, "encoder_q-layer.1": 3433.5151, "encoder_q-layer.10": 6225.0146, "encoder_q-layer.11": 15866.0361, "encoder_q-layer.2": 3794.6528, "encoder_q-layer.3": 3911.9851, "encoder_q-layer.4": 4188.4766, "encoder_q-layer.5": 4379.1763, "encoder_q-layer.6": 4837.0112, "encoder_q-layer.7": 5638.5923, "encoder_q-layer.8": 6656.8081, "encoder_q-layer.9": 6160.498, "epoch": 0.61, "inbatch_neg_score": 4.0524, "inbatch_pos_score": 4.6719, "learning_rate": 1.581578947368421e-05, "loss": 3.3658, "norm_diff": 0.0667, "norm_loss": 0.0, "num_token_doc": 66.7967, "num_token_overlap": 11.7033, "num_token_query": 31.4613, "num_token_union": 65.159, "num_word_context": 202.3181, "num_word_doc": 49.8318, "num_word_query": 23.3741, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9835.2049, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0547, "query_norm": 2.3593, "queue_k_norm": 2.4258, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4613, "sent_len_1": 66.7967, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.1188, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 139900 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3657, "doc_norm": 2.4259, "encoder_q-embeddings": 4888.6733, "encoder_q-layer.0": 3485.4539, "encoder_q-layer.1": 3675.3755, "encoder_q-layer.10": 6746.4482, "encoder_q-layer.11": 16243.6094, "encoder_q-layer.2": 4061.6907, "encoder_q-layer.3": 4242.6611, "encoder_q-layer.4": 4677.9233, "encoder_q-layer.5": 4769.0312, "encoder_q-layer.6": 5301.2358, "encoder_q-layer.7": 5644.5684, "encoder_q-layer.8": 7160.105, "encoder_q-layer.9": 6346.9434, "epoch": 0.61, "inbatch_neg_score": 4.054, "inbatch_pos_score": 4.6758, "learning_rate": 1.5789473684210526e-05, "loss": 3.3657, "norm_diff": 0.0609, "norm_loss": 0.0, "num_token_doc": 66.8119, "num_token_overlap": 11.7316, "num_token_query": 31.5709, "num_token_union": 65.2125, "num_word_context": 202.2262, "num_word_doc": 49.8644, "num_word_query": 23.4734, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10409.2385, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0547, "query_norm": 2.365, "queue_k_norm": 2.428, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5709, "sent_len_1": 66.8119, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.145, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140000 }, { "dev_runtime": 29.1423, "dev_samples_per_second": 1.098, "dev_steps_per_second": 0.034, "epoch": 0.61, "step": 140000, "test_accuracy": 94.287109375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.32711660861968994, "test_doc_norm": 2.4096617698669434, "test_inbatch_neg_score": 4.3501129150390625, "test_inbatch_pos_score": 5.376894474029541, "test_loss": 0.32711660861968994, "test_loss_align": 1.0202200412750244, "test_loss_unif": -27.722368240356445, "test_loss_unif_q@queue": -27.722368240356445, "test_norm_diff": 0.030867435038089752, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.055730819702148, "test_query_norm": 2.4405291080474854, "test_queue_k_norm": 2.4279847145080566, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043999843299388885, "test_stdq": 0.04440036788582802, "test_stdqueue_k": 0.04948699101805687, "test_stdqueue_q": 0.0 }, { "dev_runtime": 29.1423, "dev_samples_per_second": 1.098, "dev_steps_per_second": 0.034, "epoch": 0.61, "eval_beir-arguana_ndcg@10": 0.41128, "eval_beir-arguana_recall@10": 0.69203, "eval_beir-arguana_recall@100": 0.95164, "eval_beir-arguana_recall@20": 0.82504, "eval_beir-avg_ndcg@10": 0.40538474999999996, "eval_beir-avg_recall@10": 0.4755915833333334, "eval_beir-avg_recall@100": 0.65489525, "eval_beir-avg_recall@20": 0.5367599999999999, "eval_beir-cqadupstack_ndcg@10": 0.29317750000000004, "eval_beir-cqadupstack_recall@10": 0.39285583333333335, "eval_beir-cqadupstack_recall@100": 0.6303925, "eval_beir-cqadupstack_recall@20": 0.46621000000000007, "eval_beir-fiqa_ndcg@10": 0.26548, "eval_beir-fiqa_recall@10": 0.33042, "eval_beir-fiqa_recall@100": 0.61699, "eval_beir-fiqa_recall@20": 0.41384, "eval_beir-nfcorpus_ndcg@10": 0.3279, "eval_beir-nfcorpus_recall@10": 0.16058, "eval_beir-nfcorpus_recall@100": 0.30957, "eval_beir-nfcorpus_recall@20": 0.19584, "eval_beir-nq_ndcg@10": 0.32458, "eval_beir-nq_recall@10": 0.51675, "eval_beir-nq_recall@100": 0.84166, "eval_beir-nq_recall@20": 0.64113, "eval_beir-quora_ndcg@10": 0.7938, "eval_beir-quora_recall@10": 0.89916, "eval_beir-quora_recall@100": 0.98142, "eval_beir-quora_recall@20": 0.93687, "eval_beir-scidocs_ndcg@10": 0.16549, "eval_beir-scidocs_recall@10": 0.17373, "eval_beir-scidocs_recall@100": 0.38327, "eval_beir-scidocs_recall@20": 0.23272, "eval_beir-scifact_ndcg@10": 0.67206, "eval_beir-scifact_recall@10": 0.789, "eval_beir-scifact_recall@100": 0.92322, "eval_beir-scifact_recall@20": 0.82244, "eval_beir-trec-covid_ndcg@10": 0.61492, "eval_beir-trec-covid_recall@10": 0.664, "eval_beir-trec-covid_recall@100": 0.4748, "eval_beir-trec-covid_recall@20": 0.616, "eval_beir-webis-touche2020_ndcg@10": 0.18516, "eval_beir-webis-touche2020_recall@10": 0.13739, "eval_beir-webis-touche2020_recall@100": 0.43599, "eval_beir-webis-touche2020_recall@20": 0.21751, "eval_senteval-avg_sts": 0.735992379948536, "eval_senteval-sickr_spearman": 0.7300083801833853, "eval_senteval-stsb_spearman": 0.7419763797136865, "step": 140000, "test_accuracy": 94.287109375, "test_active_queue_size": 16384.0, "test_cl_loss": 0.32711660861968994, "test_doc_norm": 2.4096617698669434, "test_inbatch_neg_score": 4.3501129150390625, "test_inbatch_pos_score": 5.376894474029541, "test_loss": 0.32711660861968994, "test_loss_align": 1.0202200412750244, "test_loss_unif": -27.722368240356445, "test_loss_unif_q@queue": -27.722368240356445, "test_norm_diff": 0.030867435038089752, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.055730819702148, "test_query_norm": 2.4405291080474854, "test_queue_k_norm": 2.4279847145080566, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.043999843299388885, "test_stdq": 0.04440036788582802, "test_stdqueue_k": 0.04948699101805687, "test_stdqueue_q": 0.0 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3791, "doc_norm": 2.4285, "encoder_q-embeddings": 4695.7773, "encoder_q-layer.0": 3221.4431, "encoder_q-layer.1": 3509.9712, "encoder_q-layer.10": 6838.1963, "encoder_q-layer.11": 15604.7529, "encoder_q-layer.2": 3927.1284, "encoder_q-layer.3": 4035.2566, "encoder_q-layer.4": 4377.061, "encoder_q-layer.5": 4507.1406, "encoder_q-layer.6": 4940.811, "encoder_q-layer.7": 5849.4082, "encoder_q-layer.8": 6772.7935, "encoder_q-layer.9": 6505.082, "epoch": 0.61, "inbatch_neg_score": 4.0567, "inbatch_pos_score": 4.668, "learning_rate": 1.5763157894736844e-05, "loss": 3.3791, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.7892, "num_token_overlap": 11.6777, "num_token_query": 31.4194, "num_token_union": 65.1618, "num_word_context": 202.0305, "num_word_doc": 49.8284, "num_word_query": 23.3494, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9930.7813, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0547, "query_norm": 2.3563, "queue_k_norm": 2.4281, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4194, "sent_len_1": 66.7892, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.83, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 140100 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.3819, "doc_norm": 2.4282, "encoder_q-embeddings": 4936.6792, "encoder_q-layer.0": 3388.6765, "encoder_q-layer.1": 3656.4082, "encoder_q-layer.10": 7058.064, "encoder_q-layer.11": 16115.2002, "encoder_q-layer.2": 4087.958, "encoder_q-layer.3": 4149.2246, "encoder_q-layer.4": 4313.9761, "encoder_q-layer.5": 4550.4302, "encoder_q-layer.6": 4945.0742, "encoder_q-layer.7": 5436.3496, "encoder_q-layer.8": 6880.0659, "encoder_q-layer.9": 6341.1704, "epoch": 0.61, "inbatch_neg_score": 4.0599, "inbatch_pos_score": 4.6641, "learning_rate": 1.573684210526316e-05, "loss": 3.3819, "norm_diff": 0.0668, "norm_loss": 0.0, "num_token_doc": 66.9173, "num_token_overlap": 11.6821, "num_token_query": 31.3011, "num_token_union": 65.1504, "num_word_context": 202.2332, "num_word_doc": 49.9118, "num_word_query": 23.2367, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10097.9555, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0586, "query_norm": 2.3614, "queue_k_norm": 2.4296, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3011, "sent_len_1": 66.9173, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.4812, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140200 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3769, "doc_norm": 2.4316, "encoder_q-embeddings": 5047.6914, "encoder_q-layer.0": 3431.4763, "encoder_q-layer.1": 3773.6833, "encoder_q-layer.10": 6946.022, "encoder_q-layer.11": 16134.1143, "encoder_q-layer.2": 4198.9873, "encoder_q-layer.3": 4297.6914, "encoder_q-layer.4": 4478.7388, "encoder_q-layer.5": 4636.7642, "encoder_q-layer.6": 5166.4971, "encoder_q-layer.7": 5926.7676, "encoder_q-layer.8": 6982.9331, "encoder_q-layer.9": 6311.4697, "epoch": 0.61, "inbatch_neg_score": 4.0659, "inbatch_pos_score": 4.6953, "learning_rate": 1.5710526315789474e-05, "loss": 3.3769, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.7046, "num_token_overlap": 11.6876, "num_token_query": 31.4045, "num_token_union": 65.0815, "num_word_context": 202.3798, "num_word_doc": 49.7675, "num_word_query": 23.3378, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10329.6788, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0625, "query_norm": 2.3657, "queue_k_norm": 2.4307, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4045, "sent_len_1": 66.7046, "sent_len_max_0": 127.985, "sent_len_max_1": 209.55, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140300 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3725, "doc_norm": 2.4337, "encoder_q-embeddings": 4725.5767, "encoder_q-layer.0": 3307.9417, "encoder_q-layer.1": 3572.4023, "encoder_q-layer.10": 6595.0156, "encoder_q-layer.11": 16367.6738, "encoder_q-layer.2": 3947.0808, "encoder_q-layer.3": 3978.9741, "encoder_q-layer.4": 4314.3247, "encoder_q-layer.5": 4301.8066, "encoder_q-layer.6": 4938.9966, "encoder_q-layer.7": 5905.1455, "encoder_q-layer.8": 6792.3892, "encoder_q-layer.9": 6435.0293, "epoch": 0.61, "inbatch_neg_score": 4.0681, "inbatch_pos_score": 4.6758, "learning_rate": 1.568421052631579e-05, "loss": 3.3725, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.8425, "num_token_overlap": 11.7081, "num_token_query": 31.4381, "num_token_union": 65.1669, "num_word_context": 202.3793, "num_word_doc": 49.8822, "num_word_query": 23.3578, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10302.1448, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0703, "query_norm": 2.3595, "queue_k_norm": 2.4314, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4381, "sent_len_1": 66.8425, "sent_len_max_0": 127.9988, "sent_len_max_1": 208.3812, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140400 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.3879, "doc_norm": 2.4319, "encoder_q-embeddings": 7749.8408, "encoder_q-layer.0": 5626.1733, "encoder_q-layer.1": 6546.9209, "encoder_q-layer.10": 6618.3135, "encoder_q-layer.11": 16482.0625, "encoder_q-layer.2": 8002.0947, "encoder_q-layer.3": 8191.8623, "encoder_q-layer.4": 8839.9785, "encoder_q-layer.5": 9206.5742, "encoder_q-layer.6": 11249.3564, "encoder_q-layer.7": 11604.1016, "encoder_q-layer.8": 11871.8135, "encoder_q-layer.9": 7727.2085, "epoch": 0.61, "inbatch_neg_score": 4.0723, "inbatch_pos_score": 4.6797, "learning_rate": 1.5657894736842104e-05, "loss": 3.3879, "norm_diff": 0.0629, "norm_loss": 0.0, "num_token_doc": 66.6904, "num_token_overlap": 11.6444, "num_token_query": 31.3295, "num_token_union": 65.0722, "num_word_context": 202.1168, "num_word_doc": 49.7854, "num_word_query": 23.2632, "postclip_grad_norm": 1.0, "preclip_grad_norm": 14518.7861, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0703, "query_norm": 2.3691, "queue_k_norm": 2.4316, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3295, "sent_len_1": 66.6904, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.4263, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 140500 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3777, "doc_norm": 2.4332, "encoder_q-embeddings": 5416.8647, "encoder_q-layer.0": 3779.8364, "encoder_q-layer.1": 3999.6731, "encoder_q-layer.10": 6809.5879, "encoder_q-layer.11": 16056.9219, "encoder_q-layer.2": 4555.7241, "encoder_q-layer.3": 4565.6338, "encoder_q-layer.4": 4839.9756, "encoder_q-layer.5": 4891.167, "encoder_q-layer.6": 5399.7729, "encoder_q-layer.7": 5956.6924, "encoder_q-layer.8": 7174.4976, "encoder_q-layer.9": 6295.7549, "epoch": 0.61, "inbatch_neg_score": 4.0752, "inbatch_pos_score": 4.6797, "learning_rate": 1.5631578947368422e-05, "loss": 3.3777, "norm_diff": 0.0741, "norm_loss": 0.0, "num_token_doc": 66.8949, "num_token_overlap": 11.6535, "num_token_query": 31.2772, "num_token_union": 65.1643, "num_word_context": 202.2743, "num_word_doc": 49.8985, "num_word_query": 23.243, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10414.4715, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0742, "query_norm": 2.359, "queue_k_norm": 2.434, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2772, "sent_len_1": 66.8949, "sent_len_max_0": 127.97, "sent_len_max_1": 208.2337, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140600 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3707, "doc_norm": 2.4348, "encoder_q-embeddings": 4727.0381, "encoder_q-layer.0": 3266.905, "encoder_q-layer.1": 3507.1018, "encoder_q-layer.10": 6792.0112, "encoder_q-layer.11": 16550.8262, "encoder_q-layer.2": 3843.2603, "encoder_q-layer.3": 3896.5481, "encoder_q-layer.4": 4335.1968, "encoder_q-layer.5": 4409.7007, "encoder_q-layer.6": 5120.5498, "encoder_q-layer.7": 6086.8501, "encoder_q-layer.8": 6941.8218, "encoder_q-layer.9": 6248.7051, "epoch": 0.61, "inbatch_neg_score": 4.0813, "inbatch_pos_score": 4.6875, "learning_rate": 1.5605263157894737e-05, "loss": 3.3707, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.6626, "num_token_overlap": 11.6951, "num_token_query": 31.3608, "num_token_union": 65.003, "num_word_context": 201.8746, "num_word_doc": 49.7228, "num_word_query": 23.2862, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10347.744, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.082, "query_norm": 2.3678, "queue_k_norm": 2.4339, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3608, "sent_len_1": 66.6626, "sent_len_max_0": 128.0, "sent_len_max_1": 209.95, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140700 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3856, "doc_norm": 2.4321, "encoder_q-embeddings": 4842.6191, "encoder_q-layer.0": 3382.8938, "encoder_q-layer.1": 3660.4775, "encoder_q-layer.10": 6424.6494, "encoder_q-layer.11": 16226.6514, "encoder_q-layer.2": 4074.2905, "encoder_q-layer.3": 4225.4712, "encoder_q-layer.4": 4415.688, "encoder_q-layer.5": 4399.8833, "encoder_q-layer.6": 4841.7451, "encoder_q-layer.7": 5441.2197, "encoder_q-layer.8": 6731.1704, "encoder_q-layer.9": 6465.4424, "epoch": 0.61, "inbatch_neg_score": 4.0847, "inbatch_pos_score": 4.6875, "learning_rate": 1.5578947368421056e-05, "loss": 3.3856, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.6053, "num_token_overlap": 11.6649, "num_token_query": 31.3012, "num_token_union": 64.9816, "num_word_context": 202.0615, "num_word_doc": 49.6884, "num_word_query": 23.2378, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10197.9856, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0859, "query_norm": 2.3651, "queue_k_norm": 2.4345, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3012, "sent_len_1": 66.6053, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.3663, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140800 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3733, "doc_norm": 2.4358, "encoder_q-embeddings": 5345.626, "encoder_q-layer.0": 3795.6907, "encoder_q-layer.1": 4076.3721, "encoder_q-layer.10": 6473.0239, "encoder_q-layer.11": 15902.6152, "encoder_q-layer.2": 4556.8555, "encoder_q-layer.3": 4652.6851, "encoder_q-layer.4": 5074.1851, "encoder_q-layer.5": 5211.7217, "encoder_q-layer.6": 5656.0342, "encoder_q-layer.7": 6275.5977, "encoder_q-layer.8": 7256.1582, "encoder_q-layer.9": 6360.6665, "epoch": 0.61, "inbatch_neg_score": 4.0893, "inbatch_pos_score": 4.7109, "learning_rate": 1.5552631578947367e-05, "loss": 3.3733, "norm_diff": 0.0658, "norm_loss": 0.0, "num_token_doc": 66.8047, "num_token_overlap": 11.6734, "num_token_query": 31.3675, "num_token_union": 65.1006, "num_word_context": 202.1126, "num_word_doc": 49.8224, "num_word_query": 23.2826, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10489.6569, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0898, "query_norm": 2.3699, "queue_k_norm": 2.4352, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3675, "sent_len_1": 66.8047, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.6975, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 140900 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3774, "doc_norm": 2.4322, "encoder_q-embeddings": 4991.7935, "encoder_q-layer.0": 3451.1689, "encoder_q-layer.1": 3551.2109, "encoder_q-layer.10": 6977.1704, "encoder_q-layer.11": 16342.4248, "encoder_q-layer.2": 3949.8076, "encoder_q-layer.3": 4040.6526, "encoder_q-layer.4": 4260.2715, "encoder_q-layer.5": 4450.0381, "encoder_q-layer.6": 5333.6729, "encoder_q-layer.7": 5843.3203, "encoder_q-layer.8": 6917.3726, "encoder_q-layer.9": 6417.7637, "epoch": 0.61, "inbatch_neg_score": 4.0924, "inbatch_pos_score": 4.6992, "learning_rate": 1.5526315789473686e-05, "loss": 3.3774, "norm_diff": 0.0628, "norm_loss": 0.0, "num_token_doc": 66.8454, "num_token_overlap": 11.6671, "num_token_query": 31.3565, "num_token_union": 65.1134, "num_word_context": 202.1088, "num_word_doc": 49.8347, "num_word_query": 23.2847, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10193.1007, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0938, "query_norm": 2.3694, "queue_k_norm": 2.4367, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3565, "sent_len_1": 66.8454, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.4512, "stdk": 0.0492, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141000 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3639, "doc_norm": 2.4373, "encoder_q-embeddings": 4725.0542, "encoder_q-layer.0": 3281.3567, "encoder_q-layer.1": 3486.1951, "encoder_q-layer.10": 6883.8418, "encoder_q-layer.11": 15771.542, "encoder_q-layer.2": 3928.0515, "encoder_q-layer.3": 4007.9719, "encoder_q-layer.4": 4343.4136, "encoder_q-layer.5": 4415.2314, "encoder_q-layer.6": 5135.1519, "encoder_q-layer.7": 5762.3975, "encoder_q-layer.8": 6928.5713, "encoder_q-layer.9": 6056.6328, "epoch": 0.61, "inbatch_neg_score": 4.097, "inbatch_pos_score": 4.7148, "learning_rate": 1.55e-05, "loss": 3.3639, "norm_diff": 0.0663, "norm_loss": 0.0, "num_token_doc": 66.7761, "num_token_overlap": 11.7056, "num_token_query": 31.4937, "num_token_union": 65.1548, "num_word_context": 202.2041, "num_word_doc": 49.8409, "num_word_query": 23.3926, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9870.8731, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.0977, "query_norm": 2.371, "queue_k_norm": 2.4367, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4937, "sent_len_1": 66.7761, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.1637, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 141100 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3727, "doc_norm": 2.4374, "encoder_q-embeddings": 9664.1602, "encoder_q-layer.0": 6825.0215, "encoder_q-layer.1": 7265.3735, "encoder_q-layer.10": 11993.373, "encoder_q-layer.11": 31216.4297, "encoder_q-layer.2": 7842.8149, "encoder_q-layer.3": 8077.8613, "encoder_q-layer.4": 8642.2295, "encoder_q-layer.5": 8838.3486, "encoder_q-layer.6": 9857.2432, "encoder_q-layer.7": 11379.1816, "encoder_q-layer.8": 12954.8301, "encoder_q-layer.9": 11974.1904, "epoch": 0.61, "inbatch_neg_score": 4.1027, "inbatch_pos_score": 4.707, "learning_rate": 1.5473684210526316e-05, "loss": 3.3727, "norm_diff": 0.07, "norm_loss": 0.0, "num_token_doc": 66.91, "num_token_overlap": 11.6869, "num_token_query": 31.2898, "num_token_union": 65.144, "num_word_context": 202.072, "num_word_doc": 49.9113, "num_word_query": 23.2212, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19694.7654, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.1016, "query_norm": 2.3675, "queue_k_norm": 2.4383, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2898, "sent_len_1": 66.91, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.19, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141200 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.369, "doc_norm": 2.438, "encoder_q-embeddings": 10152.2637, "encoder_q-layer.0": 6856.2358, "encoder_q-layer.1": 7129.415, "encoder_q-layer.10": 13509.0869, "encoder_q-layer.11": 31250.6152, "encoder_q-layer.2": 7677.8799, "encoder_q-layer.3": 8100.7188, "encoder_q-layer.4": 8586.4473, "encoder_q-layer.5": 8926.7959, "encoder_q-layer.6": 9816.7803, "encoder_q-layer.7": 11383.6748, "encoder_q-layer.8": 13880.9619, "encoder_q-layer.9": 13267.9629, "epoch": 0.61, "inbatch_neg_score": 4.1059, "inbatch_pos_score": 4.7188, "learning_rate": 1.5447368421052634e-05, "loss": 3.369, "norm_diff": 0.0642, "norm_loss": 0.0, "num_token_doc": 66.7683, "num_token_overlap": 11.6953, "num_token_query": 31.4259, "num_token_union": 65.1394, "num_word_context": 202.4559, "num_word_doc": 49.8305, "num_word_query": 23.3441, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20045.7191, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.1055, "query_norm": 2.3738, "queue_k_norm": 2.439, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4259, "sent_len_1": 66.7683, "sent_len_max_0": 127.9925, "sent_len_max_1": 206.94, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141300 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.3894, "doc_norm": 2.4375, "encoder_q-embeddings": 10104.7783, "encoder_q-layer.0": 7007.6655, "encoder_q-layer.1": 7556.6782, "encoder_q-layer.10": 13050.3809, "encoder_q-layer.11": 30996.8223, "encoder_q-layer.2": 8331.6895, "encoder_q-layer.3": 8472.0518, "encoder_q-layer.4": 9218.1123, "encoder_q-layer.5": 9467.7617, "encoder_q-layer.6": 10359.1328, "encoder_q-layer.7": 12004.2109, "encoder_q-layer.8": 13620.6504, "encoder_q-layer.9": 12910.2881, "epoch": 0.61, "inbatch_neg_score": 4.1149, "inbatch_pos_score": 4.7227, "learning_rate": 1.5421052631578946e-05, "loss": 3.3894, "norm_diff": 0.0649, "norm_loss": 0.0, "num_token_doc": 66.7392, "num_token_overlap": 11.6553, "num_token_query": 31.3923, "num_token_union": 65.1244, "num_word_context": 202.2655, "num_word_doc": 49.7858, "num_word_query": 23.3036, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20156.0413, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.1133, "query_norm": 2.3726, "queue_k_norm": 2.4401, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3923, "sent_len_1": 66.7392, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.1513, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141400 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3764, "doc_norm": 2.4411, "encoder_q-embeddings": 4787.0132, "encoder_q-layer.0": 3403.353, "encoder_q-layer.1": 3589.782, "encoder_q-layer.10": 6710.6572, "encoder_q-layer.11": 15256.5088, "encoder_q-layer.2": 3957.3491, "encoder_q-layer.3": 4115.1064, "encoder_q-layer.4": 4382.8701, "encoder_q-layer.5": 4402.1792, "encoder_q-layer.6": 5158.3359, "encoder_q-layer.7": 5991.6826, "encoder_q-layer.8": 6661.5615, "encoder_q-layer.9": 6428.7344, "epoch": 0.61, "inbatch_neg_score": 4.1204, "inbatch_pos_score": 4.7305, "learning_rate": 1.5394736842105264e-05, "loss": 3.3764, "norm_diff": 0.0616, "norm_loss": 0.0, "num_token_doc": 66.7521, "num_token_overlap": 11.6411, "num_token_query": 31.3211, "num_token_union": 65.083, "num_word_context": 202.2927, "num_word_doc": 49.8108, "num_word_query": 23.2478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9820.4151, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1211, "query_norm": 2.3795, "queue_k_norm": 2.4395, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3211, "sent_len_1": 66.7521, "sent_len_max_0": 127.98, "sent_len_max_1": 207.62, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 141500 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.3698, "doc_norm": 2.4417, "encoder_q-embeddings": 4794.7476, "encoder_q-layer.0": 3402.4121, "encoder_q-layer.1": 3507.0864, "encoder_q-layer.10": 6694.9463, "encoder_q-layer.11": 16255.2266, "encoder_q-layer.2": 3821.5364, "encoder_q-layer.3": 3801.1924, "encoder_q-layer.4": 4292.4731, "encoder_q-layer.5": 4639.3164, "encoder_q-layer.6": 4958.8994, "encoder_q-layer.7": 5921.2861, "encoder_q-layer.8": 7124.646, "encoder_q-layer.9": 6304.1919, "epoch": 0.61, "inbatch_neg_score": 4.1236, "inbatch_pos_score": 4.7578, "learning_rate": 1.536842105263158e-05, "loss": 3.3698, "norm_diff": 0.0532, "norm_loss": 0.0, "num_token_doc": 66.7885, "num_token_overlap": 11.6866, "num_token_query": 31.326, "num_token_union": 65.0707, "num_word_context": 202.2671, "num_word_doc": 49.8046, "num_word_query": 23.2721, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10091.7924, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.125, "query_norm": 2.3885, "queue_k_norm": 2.4419, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.326, "sent_len_1": 66.7885, "sent_len_max_0": 127.98, "sent_len_max_1": 209.9512, "stdk": 0.0494, "stdq": 0.0439, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141600 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3741, "doc_norm": 2.4429, "encoder_q-embeddings": 4577.0264, "encoder_q-layer.0": 3244.0181, "encoder_q-layer.1": 3449.2754, "encoder_q-layer.10": 6430.3418, "encoder_q-layer.11": 15191.9766, "encoder_q-layer.2": 3761.4746, "encoder_q-layer.3": 3813.0964, "encoder_q-layer.4": 4023.4133, "encoder_q-layer.5": 4249.2402, "encoder_q-layer.6": 4570.5435, "encoder_q-layer.7": 5176.9556, "encoder_q-layer.8": 6347.0088, "encoder_q-layer.9": 6182.6201, "epoch": 0.61, "inbatch_neg_score": 4.1259, "inbatch_pos_score": 4.7422, "learning_rate": 1.5342105263157897e-05, "loss": 3.3741, "norm_diff": 0.065, "norm_loss": 0.0, "num_token_doc": 66.8304, "num_token_overlap": 11.6813, "num_token_query": 31.3586, "num_token_union": 65.1575, "num_word_context": 202.343, "num_word_doc": 49.892, "num_word_query": 23.286, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9407.5383, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.125, "query_norm": 2.3779, "queue_k_norm": 2.4424, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3586, "sent_len_1": 66.8304, "sent_len_max_0": 127.99, "sent_len_max_1": 207.25, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141700 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3801, "doc_norm": 2.4446, "encoder_q-embeddings": 4940.0249, "encoder_q-layer.0": 3322.6479, "encoder_q-layer.1": 3477.8743, "encoder_q-layer.10": 6954.5, "encoder_q-layer.11": 16241.8779, "encoder_q-layer.2": 3948.3535, "encoder_q-layer.3": 3971.374, "encoder_q-layer.4": 4239.6851, "encoder_q-layer.5": 4368.4629, "encoder_q-layer.6": 5186.9014, "encoder_q-layer.7": 5918.8374, "encoder_q-layer.8": 6901.915, "encoder_q-layer.9": 6606.7329, "epoch": 0.62, "inbatch_neg_score": 4.1312, "inbatch_pos_score": 4.7383, "learning_rate": 1.5315789473684212e-05, "loss": 3.3801, "norm_diff": 0.0623, "norm_loss": 0.0, "num_token_doc": 66.752, "num_token_overlap": 11.6835, "num_token_query": 31.3453, "num_token_union": 65.0827, "num_word_context": 202.4439, "num_word_doc": 49.8272, "num_word_query": 23.2834, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10132.1189, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1328, "query_norm": 2.3823, "queue_k_norm": 2.4442, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3453, "sent_len_1": 66.752, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.3713, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141800 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3772, "doc_norm": 2.4446, "encoder_q-embeddings": 4638.9805, "encoder_q-layer.0": 3311.4685, "encoder_q-layer.1": 3501.7988, "encoder_q-layer.10": 6349.1719, "encoder_q-layer.11": 16269.8398, "encoder_q-layer.2": 3807.4341, "encoder_q-layer.3": 3905.6934, "encoder_q-layer.4": 4397.0674, "encoder_q-layer.5": 4413.2056, "encoder_q-layer.6": 4776.4805, "encoder_q-layer.7": 5898.4346, "encoder_q-layer.8": 6982.3843, "encoder_q-layer.9": 6214.6182, "epoch": 0.62, "inbatch_neg_score": 4.1409, "inbatch_pos_score": 4.7578, "learning_rate": 1.5289473684210527e-05, "loss": 3.3772, "norm_diff": 0.0547, "norm_loss": 0.0, "num_token_doc": 66.8591, "num_token_overlap": 11.6791, "num_token_query": 31.3364, "num_token_union": 65.1366, "num_word_context": 202.4885, "num_word_doc": 49.9108, "num_word_query": 23.2835, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10154.5296, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1406, "query_norm": 2.3899, "queue_k_norm": 2.4455, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3364, "sent_len_1": 66.8591, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.4187, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 141900 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3627, "doc_norm": 2.4448, "encoder_q-embeddings": 4734.1377, "encoder_q-layer.0": 3273.207, "encoder_q-layer.1": 3499.3652, "encoder_q-layer.10": 6407.0654, "encoder_q-layer.11": 15753.6201, "encoder_q-layer.2": 3922.2952, "encoder_q-layer.3": 3938.1948, "encoder_q-layer.4": 4216.4297, "encoder_q-layer.5": 4249.313, "encoder_q-layer.6": 4844.415, "encoder_q-layer.7": 5724.2725, "encoder_q-layer.8": 6959.0137, "encoder_q-layer.9": 6379.0488, "epoch": 0.62, "inbatch_neg_score": 4.149, "inbatch_pos_score": 4.7656, "learning_rate": 1.5263157894736842e-05, "loss": 3.3627, "norm_diff": 0.0484, "norm_loss": 0.0, "num_token_doc": 66.8536, "num_token_overlap": 11.688, "num_token_query": 31.3863, "num_token_union": 65.1517, "num_word_context": 202.4678, "num_word_doc": 49.887, "num_word_query": 23.3087, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9855.6282, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1484, "query_norm": 2.3964, "queue_k_norm": 2.4465, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3863, "sent_len_1": 66.8536, "sent_len_max_0": 127.985, "sent_len_max_1": 208.3487, "stdk": 0.0494, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142000 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.3735, "doc_norm": 2.4503, "encoder_q-embeddings": 4795.7754, "encoder_q-layer.0": 3260.3088, "encoder_q-layer.1": 3520.7644, "encoder_q-layer.10": 6597.4565, "encoder_q-layer.11": 16639.1152, "encoder_q-layer.2": 3842.6289, "encoder_q-layer.3": 3978.2979, "encoder_q-layer.4": 4425.9868, "encoder_q-layer.5": 4714.9048, "encoder_q-layer.6": 4991.3589, "encoder_q-layer.7": 5627.7021, "encoder_q-layer.8": 6802.4629, "encoder_q-layer.9": 6438.7271, "epoch": 0.62, "inbatch_neg_score": 4.1555, "inbatch_pos_score": 4.7695, "learning_rate": 1.5236842105263157e-05, "loss": 3.3735, "norm_diff": 0.0558, "norm_loss": 0.0, "num_token_doc": 66.6492, "num_token_overlap": 11.6567, "num_token_query": 31.3511, "num_token_union": 65.0755, "num_word_context": 202.2864, "num_word_doc": 49.7447, "num_word_query": 23.2748, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10198.886, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1562, "query_norm": 2.3945, "queue_k_norm": 2.4471, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3511, "sent_len_1": 66.6492, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.29, "stdk": 0.0497, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142100 }, { "accuracy": 48.6816, "active_queue_size": 16384.0, "cl_loss": 3.3652, "doc_norm": 2.446, "encoder_q-embeddings": 4657.4932, "encoder_q-layer.0": 3338.7505, "encoder_q-layer.1": 3637.0718, "encoder_q-layer.10": 7162.7236, "encoder_q-layer.11": 16610.9141, "encoder_q-layer.2": 3988.552, "encoder_q-layer.3": 4078.3027, "encoder_q-layer.4": 4387.9355, "encoder_q-layer.5": 4541.4663, "encoder_q-layer.6": 5145.6846, "encoder_q-layer.7": 6332.9937, "encoder_q-layer.8": 6794.1357, "encoder_q-layer.9": 6257.6689, "epoch": 0.62, "inbatch_neg_score": 4.1646, "inbatch_pos_score": 4.7578, "learning_rate": 1.5210526315789476e-05, "loss": 3.3652, "norm_diff": 0.0549, "norm_loss": 0.0, "num_token_doc": 66.994, "num_token_overlap": 11.7492, "num_token_query": 31.5615, "num_token_union": 65.3092, "num_word_context": 202.3663, "num_word_doc": 49.9934, "num_word_query": 23.4442, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10281.2998, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1641, "query_norm": 2.391, "queue_k_norm": 2.4484, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5615, "sent_len_1": 66.994, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.81, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142200 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3636, "doc_norm": 2.4461, "encoder_q-embeddings": 4899.4053, "encoder_q-layer.0": 3441.7129, "encoder_q-layer.1": 3681.6294, "encoder_q-layer.10": 8909.1914, "encoder_q-layer.11": 18215.0996, "encoder_q-layer.2": 4220.6758, "encoder_q-layer.3": 4353.7021, "encoder_q-layer.4": 4725.2827, "encoder_q-layer.5": 4787.2617, "encoder_q-layer.6": 5436.8945, "encoder_q-layer.7": 6692.2324, "encoder_q-layer.8": 7670.6851, "encoder_q-layer.9": 7755.9404, "epoch": 0.62, "inbatch_neg_score": 4.1727, "inbatch_pos_score": 4.793, "learning_rate": 1.518421052631579e-05, "loss": 3.3636, "norm_diff": 0.0427, "norm_loss": 0.0, "num_token_doc": 66.8792, "num_token_overlap": 11.676, "num_token_query": 31.2548, "num_token_union": 65.1264, "num_word_context": 202.3398, "num_word_doc": 49.9243, "num_word_query": 23.2149, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11055.512, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1719, "query_norm": 2.4034, "queue_k_norm": 2.4494, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2548, "sent_len_1": 66.8792, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.6662, "stdk": 0.0492, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142300 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3776, "doc_norm": 2.4515, "encoder_q-embeddings": 5232.3125, "encoder_q-layer.0": 3648.793, "encoder_q-layer.1": 3927.45, "encoder_q-layer.10": 6510.3491, "encoder_q-layer.11": 15275.1055, "encoder_q-layer.2": 4477.3379, "encoder_q-layer.3": 4589.7432, "encoder_q-layer.4": 4833.1577, "encoder_q-layer.5": 4924.2266, "encoder_q-layer.6": 5524.4849, "encoder_q-layer.7": 6233.1558, "encoder_q-layer.8": 7063.7622, "encoder_q-layer.9": 6647.7671, "epoch": 0.62, "inbatch_neg_score": 4.1809, "inbatch_pos_score": 4.7969, "learning_rate": 1.5157894736842107e-05, "loss": 3.3776, "norm_diff": 0.0512, "norm_loss": 0.0, "num_token_doc": 66.6447, "num_token_overlap": 11.6773, "num_token_query": 31.3615, "num_token_union": 65.058, "num_word_context": 202.185, "num_word_doc": 49.7581, "num_word_query": 23.2966, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10301.3832, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1797, "query_norm": 2.4003, "queue_k_norm": 2.4503, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3615, "sent_len_1": 66.6447, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.3275, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142400 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3705, "doc_norm": 2.451, "encoder_q-embeddings": 4703.4595, "encoder_q-layer.0": 3363.4907, "encoder_q-layer.1": 3520.9482, "encoder_q-layer.10": 7899.7036, "encoder_q-layer.11": 16855.1816, "encoder_q-layer.2": 4002.9741, "encoder_q-layer.3": 4043.7593, "encoder_q-layer.4": 4323.3438, "encoder_q-layer.5": 4382.9883, "encoder_q-layer.6": 4942.7153, "encoder_q-layer.7": 5983.7373, "encoder_q-layer.8": 6767.3936, "encoder_q-layer.9": 6467.2681, "epoch": 0.62, "inbatch_neg_score": 4.1894, "inbatch_pos_score": 4.8008, "learning_rate": 1.5131578947368422e-05, "loss": 3.3705, "norm_diff": 0.0528, "norm_loss": 0.0, "num_token_doc": 66.763, "num_token_overlap": 11.6879, "num_token_query": 31.341, "num_token_union": 65.0929, "num_word_context": 202.4899, "num_word_doc": 49.8422, "num_word_query": 23.295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10275.4485, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1875, "query_norm": 2.3983, "queue_k_norm": 2.4525, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.341, "sent_len_1": 66.763, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.4888, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 142500 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3893, "doc_norm": 2.4568, "encoder_q-embeddings": 5103.4551, "encoder_q-layer.0": 3526.0127, "encoder_q-layer.1": 3660.4927, "encoder_q-layer.10": 6747.2417, "encoder_q-layer.11": 15589.4609, "encoder_q-layer.2": 4159.9614, "encoder_q-layer.3": 4436.7583, "encoder_q-layer.4": 4526.3018, "encoder_q-layer.5": 4664.4365, "encoder_q-layer.6": 5245.0771, "encoder_q-layer.7": 5954.0693, "encoder_q-layer.8": 6823.3999, "encoder_q-layer.9": 6207.791, "epoch": 0.62, "inbatch_neg_score": 4.1957, "inbatch_pos_score": 4.8125, "learning_rate": 1.5105263157894735e-05, "loss": 3.3893, "norm_diff": 0.0594, "norm_loss": 0.0, "num_token_doc": 66.6877, "num_token_overlap": 11.6284, "num_token_query": 31.2505, "num_token_union": 65.0499, "num_word_context": 202.3216, "num_word_doc": 49.7867, "num_word_query": 23.1963, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10068.658, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.1953, "query_norm": 2.3974, "queue_k_norm": 2.4527, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2505, "sent_len_1": 66.6877, "sent_len_max_0": 127.9862, "sent_len_max_1": 205.7713, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142600 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3812, "doc_norm": 2.4571, "encoder_q-embeddings": 4917.981, "encoder_q-layer.0": 3428.2041, "encoder_q-layer.1": 3731.196, "encoder_q-layer.10": 6422.9922, "encoder_q-layer.11": 15736.5391, "encoder_q-layer.2": 4115.0918, "encoder_q-layer.3": 4264.6191, "encoder_q-layer.4": 4661.6265, "encoder_q-layer.5": 4795.522, "encoder_q-layer.6": 5501.9575, "encoder_q-layer.7": 5923.0977, "encoder_q-layer.8": 6716.605, "encoder_q-layer.9": 6168.3594, "epoch": 0.62, "inbatch_neg_score": 4.2018, "inbatch_pos_score": 4.8242, "learning_rate": 1.5078947368421054e-05, "loss": 3.3812, "norm_diff": 0.0556, "norm_loss": 0.0, "num_token_doc": 66.7769, "num_token_overlap": 11.6349, "num_token_query": 31.2525, "num_token_union": 65.0902, "num_word_context": 202.4084, "num_word_doc": 49.8141, "num_word_query": 23.2043, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10045.2019, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2031, "query_norm": 2.4016, "queue_k_norm": 2.4548, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2525, "sent_len_1": 66.7769, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.5675, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142700 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3829, "doc_norm": 2.4545, "encoder_q-embeddings": 4963.0371, "encoder_q-layer.0": 3450.5271, "encoder_q-layer.1": 3790.7004, "encoder_q-layer.10": 6666.5386, "encoder_q-layer.11": 16492.3359, "encoder_q-layer.2": 4214.3081, "encoder_q-layer.3": 4272.5732, "encoder_q-layer.4": 4859.1294, "encoder_q-layer.5": 4931.1816, "encoder_q-layer.6": 5283.2754, "encoder_q-layer.7": 6090.4019, "encoder_q-layer.8": 6957.1245, "encoder_q-layer.9": 6220.7959, "epoch": 0.62, "inbatch_neg_score": 4.2065, "inbatch_pos_score": 4.8125, "learning_rate": 1.5052631578947369e-05, "loss": 3.3829, "norm_diff": 0.0649, "norm_loss": 0.0, "num_token_doc": 66.7935, "num_token_overlap": 11.7253, "num_token_query": 31.434, "num_token_union": 65.1246, "num_word_context": 202.2051, "num_word_doc": 49.8398, "num_word_query": 23.3445, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10370.4307, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.207, "query_norm": 2.3896, "queue_k_norm": 2.4578, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.434, "sent_len_1": 66.7935, "sent_len_max_0": 127.985, "sent_len_max_1": 209.0838, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 142800 }, { "accuracy": 52.3926, "active_queue_size": 16384.0, "cl_loss": 3.3769, "doc_norm": 2.4583, "encoder_q-embeddings": 4448.4399, "encoder_q-layer.0": 3150.564, "encoder_q-layer.1": 3375.0671, "encoder_q-layer.10": 6571.1514, "encoder_q-layer.11": 15982.0459, "encoder_q-layer.2": 3856.5291, "encoder_q-layer.3": 4030.9058, "encoder_q-layer.4": 4407.8701, "encoder_q-layer.5": 4196.2583, "encoder_q-layer.6": 4750.6055, "encoder_q-layer.7": 5314.4307, "encoder_q-layer.8": 6868.2773, "encoder_q-layer.9": 6307.2881, "epoch": 0.62, "inbatch_neg_score": 4.2133, "inbatch_pos_score": 4.8242, "learning_rate": 1.5026315789473685e-05, "loss": 3.3769, "norm_diff": 0.0615, "norm_loss": 0.0, "num_token_doc": 66.7958, "num_token_overlap": 11.697, "num_token_query": 31.431, "num_token_union": 65.1426, "num_word_context": 202.3059, "num_word_doc": 49.8383, "num_word_query": 23.3493, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9984.294, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2148, "query_norm": 2.3969, "queue_k_norm": 2.4574, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.431, "sent_len_1": 66.7958, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.7725, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 142900 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3793, "doc_norm": 2.4589, "encoder_q-embeddings": 5329.9556, "encoder_q-layer.0": 3823.8975, "encoder_q-layer.1": 4112.4023, "encoder_q-layer.10": 6386.9619, "encoder_q-layer.11": 15850.0469, "encoder_q-layer.2": 4950.3638, "encoder_q-layer.3": 5245.4429, "encoder_q-layer.4": 5778.3877, "encoder_q-layer.5": 5376.5166, "encoder_q-layer.6": 5524.1797, "encoder_q-layer.7": 5736.3018, "encoder_q-layer.8": 6634.1929, "encoder_q-layer.9": 6065.0249, "epoch": 0.62, "inbatch_neg_score": 4.2169, "inbatch_pos_score": 4.832, "learning_rate": 1.5e-05, "loss": 3.3793, "norm_diff": 0.063, "norm_loss": 0.0, "num_token_doc": 66.8138, "num_token_overlap": 11.6724, "num_token_query": 31.3702, "num_token_union": 65.1492, "num_word_context": 202.5555, "num_word_doc": 49.852, "num_word_query": 23.3089, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10332.1761, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2148, "query_norm": 2.3959, "queue_k_norm": 2.458, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3702, "sent_len_1": 66.8138, "sent_len_max_0": 127.98, "sent_len_max_1": 209.5137, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 143000 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3821, "doc_norm": 2.4603, "encoder_q-embeddings": 4660.0269, "encoder_q-layer.0": 3313.0725, "encoder_q-layer.1": 3568.5757, "encoder_q-layer.10": 7358.6909, "encoder_q-layer.11": 17188.8633, "encoder_q-layer.2": 4001.8738, "encoder_q-layer.3": 4128.4795, "encoder_q-layer.4": 4301.9116, "encoder_q-layer.5": 4351.2466, "encoder_q-layer.6": 4861.9946, "encoder_q-layer.7": 5807.3975, "encoder_q-layer.8": 6940.709, "encoder_q-layer.9": 6551.9683, "epoch": 0.62, "inbatch_neg_score": 4.2201, "inbatch_pos_score": 4.832, "learning_rate": 1.4973684210526317e-05, "loss": 3.3821, "norm_diff": 0.0667, "norm_loss": 0.0, "num_token_doc": 66.7165, "num_token_overlap": 11.6364, "num_token_query": 31.265, "num_token_union": 65.0713, "num_word_context": 202.2598, "num_word_doc": 49.7991, "num_word_query": 23.2178, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10284.776, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2188, "query_norm": 2.3936, "queue_k_norm": 2.4602, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.265, "sent_len_1": 66.7165, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.56, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 143100 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3765, "doc_norm": 2.4589, "encoder_q-embeddings": 5509.3892, "encoder_q-layer.0": 3905.7798, "encoder_q-layer.1": 4241.79, "encoder_q-layer.10": 6180.6265, "encoder_q-layer.11": 15850.6123, "encoder_q-layer.2": 4794.4502, "encoder_q-layer.3": 4758.0503, "encoder_q-layer.4": 5100.7515, "encoder_q-layer.5": 5156.2793, "encoder_q-layer.6": 5641.5215, "encoder_q-layer.7": 6337.978, "encoder_q-layer.8": 7274.5737, "encoder_q-layer.9": 6170.7397, "epoch": 0.62, "inbatch_neg_score": 4.2209, "inbatch_pos_score": 4.8242, "learning_rate": 1.4947368421052632e-05, "loss": 3.3765, "norm_diff": 0.0627, "norm_loss": 0.0, "num_token_doc": 66.7306, "num_token_overlap": 11.6741, "num_token_query": 31.3397, "num_token_union": 65.0902, "num_word_context": 202.3135, "num_word_doc": 49.7926, "num_word_query": 23.2745, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10501.3216, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2227, "query_norm": 2.3963, "queue_k_norm": 2.4597, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3397, "sent_len_1": 66.7306, "sent_len_max_0": 127.98, "sent_len_max_1": 207.8663, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 143200 }, { "accuracy": 53.0762, "active_queue_size": 16384.0, "cl_loss": 3.375, "doc_norm": 2.4657, "encoder_q-embeddings": 5040.8804, "encoder_q-layer.0": 3785.1917, "encoder_q-layer.1": 3968.6094, "encoder_q-layer.10": 6982.9697, "encoder_q-layer.11": 15856.5049, "encoder_q-layer.2": 4404.8643, "encoder_q-layer.3": 4672.5239, "encoder_q-layer.4": 5070.8672, "encoder_q-layer.5": 5487.2827, "encoder_q-layer.6": 5754.4575, "encoder_q-layer.7": 6025.7598, "encoder_q-layer.8": 6973.1855, "encoder_q-layer.9": 6371.4189, "epoch": 0.62, "inbatch_neg_score": 4.2231, "inbatch_pos_score": 4.8555, "learning_rate": 1.4921052631578947e-05, "loss": 3.375, "norm_diff": 0.0647, "norm_loss": 0.0, "num_token_doc": 66.8011, "num_token_overlap": 11.681, "num_token_query": 31.3015, "num_token_union": 65.0952, "num_word_context": 202.3479, "num_word_doc": 49.8402, "num_word_query": 23.2527, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10450.1529, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2227, "query_norm": 2.4009, "queue_k_norm": 2.4624, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3015, "sent_len_1": 66.8011, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.1163, "stdk": 0.0497, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 143300 }, { "accuracy": 53.0273, "active_queue_size": 16384.0, "cl_loss": 3.3633, "doc_norm": 2.4656, "encoder_q-embeddings": 4831.6597, "encoder_q-layer.0": 3373.2688, "encoder_q-layer.1": 3616.5571, "encoder_q-layer.10": 6460.8115, "encoder_q-layer.11": 15949.0146, "encoder_q-layer.2": 4028.4888, "encoder_q-layer.3": 4144.6309, "encoder_q-layer.4": 4535.0815, "encoder_q-layer.5": 4467.98, "encoder_q-layer.6": 5003.7773, "encoder_q-layer.7": 5676.332, "encoder_q-layer.8": 6921.4253, "encoder_q-layer.9": 6238.3989, "epoch": 0.62, "inbatch_neg_score": 4.2259, "inbatch_pos_score": 4.8477, "learning_rate": 1.4894736842105264e-05, "loss": 3.3633, "norm_diff": 0.072, "norm_loss": 0.0, "num_token_doc": 66.811, "num_token_overlap": 11.7, "num_token_query": 31.3806, "num_token_union": 65.1516, "num_word_context": 202.2036, "num_word_doc": 49.8347, "num_word_query": 23.3013, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10060.2469, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.2266, "query_norm": 2.3936, "queue_k_norm": 2.4634, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3806, "sent_len_1": 66.811, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.2237, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 143400 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3848, "doc_norm": 2.4636, "encoder_q-embeddings": 9432.0957, "encoder_q-layer.0": 6518.1328, "encoder_q-layer.1": 6954.5771, "encoder_q-layer.10": 12897.2168, "encoder_q-layer.11": 31433.9883, "encoder_q-layer.2": 7729.3066, "encoder_q-layer.3": 8048.9888, "encoder_q-layer.4": 8743.8848, "encoder_q-layer.5": 8830.2539, "encoder_q-layer.6": 9701.4014, "encoder_q-layer.7": 11008.3838, "encoder_q-layer.8": 13454.3564, "encoder_q-layer.9": 12512.1387, "epoch": 0.62, "inbatch_neg_score": 4.225, "inbatch_pos_score": 4.8359, "learning_rate": 1.4868421052631579e-05, "loss": 3.3848, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.7032, "num_token_overlap": 11.6284, "num_token_query": 31.2233, "num_token_union": 65.0064, "num_word_context": 202.1539, "num_word_doc": 49.7746, "num_word_query": 23.1722, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19824.6956, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2266, "query_norm": 2.3917, "queue_k_norm": 2.4637, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2233, "sent_len_1": 66.7032, "sent_len_max_0": 127.95, "sent_len_max_1": 209.3675, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 143500 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.368, "doc_norm": 2.4638, "encoder_q-embeddings": 9998.7314, "encoder_q-layer.0": 7093.7012, "encoder_q-layer.1": 7194.2075, "encoder_q-layer.10": 13050.5303, "encoder_q-layer.11": 31442.1191, "encoder_q-layer.2": 8439.9053, "encoder_q-layer.3": 8840.2881, "encoder_q-layer.4": 9276.9189, "encoder_q-layer.5": 9256.7061, "encoder_q-layer.6": 10486.8301, "encoder_q-layer.7": 11262.9365, "encoder_q-layer.8": 13551.3838, "encoder_q-layer.9": 12405.4766, "epoch": 0.62, "inbatch_neg_score": 4.228, "inbatch_pos_score": 4.8359, "learning_rate": 1.4842105263157895e-05, "loss": 3.368, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.9444, "num_token_overlap": 11.6922, "num_token_query": 31.4627, "num_token_union": 65.2758, "num_word_context": 202.6242, "num_word_doc": 49.9741, "num_word_query": 23.374, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20292.1698, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2305, "query_norm": 2.3881, "queue_k_norm": 2.4657, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4627, "sent_len_1": 66.9444, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.015, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 143600 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3728, "doc_norm": 2.4664, "encoder_q-embeddings": 9186.3242, "encoder_q-layer.0": 6559.6787, "encoder_q-layer.1": 6958.229, "encoder_q-layer.10": 14943.6152, "encoder_q-layer.11": 32769.0391, "encoder_q-layer.2": 7812.4873, "encoder_q-layer.3": 8075.5728, "encoder_q-layer.4": 8711.3574, "encoder_q-layer.5": 8866.6084, "encoder_q-layer.6": 9687.2783, "encoder_q-layer.7": 10973.1543, "encoder_q-layer.8": 13657.8564, "encoder_q-layer.9": 12557.2646, "epoch": 0.62, "inbatch_neg_score": 4.2336, "inbatch_pos_score": 4.8477, "learning_rate": 1.481578947368421e-05, "loss": 3.3728, "norm_diff": 0.0737, "norm_loss": 0.0, "num_token_doc": 66.9089, "num_token_overlap": 11.7059, "num_token_query": 31.3782, "num_token_union": 65.1942, "num_word_context": 202.4528, "num_word_doc": 49.932, "num_word_query": 23.3077, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20532.6012, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2344, "query_norm": 2.3927, "queue_k_norm": 2.4657, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3782, "sent_len_1": 66.9089, "sent_len_max_0": 127.99, "sent_len_max_1": 208.5762, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 143700 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3751, "doc_norm": 2.4684, "encoder_q-embeddings": 9895.2227, "encoder_q-layer.0": 7066.6973, "encoder_q-layer.1": 7203.584, "encoder_q-layer.10": 12746.4805, "encoder_q-layer.11": 31332.0449, "encoder_q-layer.2": 7638.0542, "encoder_q-layer.3": 8061.8398, "encoder_q-layer.4": 8382.5098, "encoder_q-layer.5": 8619.459, "encoder_q-layer.6": 9565.957, "encoder_q-layer.7": 11401.0566, "encoder_q-layer.8": 13121.6084, "encoder_q-layer.9": 12449.4014, "epoch": 0.62, "inbatch_neg_score": 4.234, "inbatch_pos_score": 4.8516, "learning_rate": 1.4789473684210529e-05, "loss": 3.3751, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.8827, "num_token_overlap": 11.6836, "num_token_query": 31.3093, "num_token_union": 65.1346, "num_word_context": 202.4211, "num_word_doc": 49.8844, "num_word_query": 23.2546, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20064.2259, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2344, "query_norm": 2.3927, "queue_k_norm": 2.4659, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3093, "sent_len_1": 66.8827, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.9675, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 143800 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3701, "doc_norm": 2.4672, "encoder_q-embeddings": 9997.6523, "encoder_q-layer.0": 7037.8872, "encoder_q-layer.1": 7592.7612, "encoder_q-layer.10": 13615.5459, "encoder_q-layer.11": 32537.5977, "encoder_q-layer.2": 8842.2061, "encoder_q-layer.3": 9525.4922, "encoder_q-layer.4": 9836.4248, "encoder_q-layer.5": 10101.6875, "encoder_q-layer.6": 10661.6768, "encoder_q-layer.7": 12125.8945, "encoder_q-layer.8": 14324.0488, "encoder_q-layer.9": 12404.4561, "epoch": 0.62, "inbatch_neg_score": 4.2341, "inbatch_pos_score": 4.8477, "learning_rate": 1.4763157894736842e-05, "loss": 3.3701, "norm_diff": 0.0753, "norm_loss": 0.0, "num_token_doc": 66.7465, "num_token_overlap": 11.6959, "num_token_query": 31.4128, "num_token_union": 65.1076, "num_word_context": 202.1049, "num_word_doc": 49.8233, "num_word_query": 23.3497, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21005.4125, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2344, "query_norm": 2.3919, "queue_k_norm": 2.4682, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4128, "sent_len_1": 66.7465, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.0975, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 143900 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3702, "doc_norm": 2.4676, "encoder_q-embeddings": 9750.666, "encoder_q-layer.0": 6754.4473, "encoder_q-layer.1": 7073.5913, "encoder_q-layer.10": 14112.0322, "encoder_q-layer.11": 33519.5547, "encoder_q-layer.2": 7799.8018, "encoder_q-layer.3": 7951.665, "encoder_q-layer.4": 8448.3008, "encoder_q-layer.5": 9054.5283, "encoder_q-layer.6": 9594.2451, "encoder_q-layer.7": 11141.8555, "encoder_q-layer.8": 13472.2852, "encoder_q-layer.9": 12838.2012, "epoch": 0.62, "inbatch_neg_score": 4.238, "inbatch_pos_score": 4.8359, "learning_rate": 1.4736842105263157e-05, "loss": 3.3702, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.8821, "num_token_overlap": 11.6655, "num_token_query": 31.3975, "num_token_union": 65.2411, "num_word_context": 202.3363, "num_word_doc": 49.9073, "num_word_query": 23.3123, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20475.5726, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2383, "query_norm": 2.3874, "queue_k_norm": 2.4687, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3975, "sent_len_1": 66.8821, "sent_len_max_0": 127.98, "sent_len_max_1": 207.9487, "stdk": 0.0494, "stdq": 0.0427, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 144000 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.3636, "doc_norm": 2.471, "encoder_q-embeddings": 9477.3389, "encoder_q-layer.0": 6621.2354, "encoder_q-layer.1": 7151.7256, "encoder_q-layer.10": 13265.2383, "encoder_q-layer.11": 32139.1699, "encoder_q-layer.2": 8132.8032, "encoder_q-layer.3": 8500.127, "encoder_q-layer.4": 9244.5049, "encoder_q-layer.5": 9562.1592, "encoder_q-layer.6": 10795.6436, "encoder_q-layer.7": 12349.4648, "encoder_q-layer.8": 14291.748, "encoder_q-layer.9": 12747.6221, "epoch": 0.63, "inbatch_neg_score": 4.2422, "inbatch_pos_score": 4.8555, "learning_rate": 1.4710526315789475e-05, "loss": 3.3636, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.7422, "num_token_overlap": 11.6658, "num_token_query": 31.3994, "num_token_union": 65.1357, "num_word_context": 202.1437, "num_word_doc": 49.7978, "num_word_query": 23.3172, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20531.9593, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2422, "query_norm": 2.4007, "queue_k_norm": 2.4692, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3994, "sent_len_1": 66.7422, "sent_len_max_0": 127.98, "sent_len_max_1": 209.125, "stdk": 0.0496, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144100 }, { "accuracy": 48.4375, "active_queue_size": 16384.0, "cl_loss": 3.3744, "doc_norm": 2.4707, "encoder_q-embeddings": 9855.2891, "encoder_q-layer.0": 6791.7759, "encoder_q-layer.1": 7246.125, "encoder_q-layer.10": 12935.1299, "encoder_q-layer.11": 31872.6152, "encoder_q-layer.2": 8053.5728, "encoder_q-layer.3": 8070.3711, "encoder_q-layer.4": 8435.4414, "encoder_q-layer.5": 8660.54, "encoder_q-layer.6": 10162.9229, "encoder_q-layer.7": 11379.7832, "encoder_q-layer.8": 13747.6055, "encoder_q-layer.9": 12853.708, "epoch": 0.63, "inbatch_neg_score": 4.2431, "inbatch_pos_score": 4.8359, "learning_rate": 1.4684210526315789e-05, "loss": 3.3744, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.7568, "num_token_overlap": 11.6919, "num_token_query": 31.4199, "num_token_union": 65.129, "num_word_context": 202.2637, "num_word_doc": 49.8262, "num_word_query": 23.3403, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20107.1039, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2422, "query_norm": 2.3909, "queue_k_norm": 2.469, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4199, "sent_len_1": 66.7568, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.8313, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144200 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.3775, "doc_norm": 2.4703, "encoder_q-embeddings": 9205.1465, "encoder_q-layer.0": 6615.7993, "encoder_q-layer.1": 7154.7324, "encoder_q-layer.10": 14412.2598, "encoder_q-layer.11": 33187.4219, "encoder_q-layer.2": 8112.0039, "encoder_q-layer.3": 8642.4707, "encoder_q-layer.4": 8743.1416, "encoder_q-layer.5": 9210.8477, "encoder_q-layer.6": 11073.3691, "encoder_q-layer.7": 12560.8115, "encoder_q-layer.8": 14746.6562, "encoder_q-layer.9": 12773.248, "epoch": 0.63, "inbatch_neg_score": 4.2468, "inbatch_pos_score": 4.8477, "learning_rate": 1.4657894736842107e-05, "loss": 3.3775, "norm_diff": 0.0763, "norm_loss": 0.0, "num_token_doc": 66.7283, "num_token_overlap": 11.6715, "num_token_query": 31.4212, "num_token_union": 65.1303, "num_word_context": 202.3255, "num_word_doc": 49.7868, "num_word_query": 23.3532, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21181.2153, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2461, "query_norm": 2.394, "queue_k_norm": 2.4698, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4212, "sent_len_1": 66.7283, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.5762, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144300 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.3759, "doc_norm": 2.4701, "encoder_q-embeddings": 9571.9707, "encoder_q-layer.0": 6794.9009, "encoder_q-layer.1": 7161.5713, "encoder_q-layer.10": 14106.2041, "encoder_q-layer.11": 33504.0391, "encoder_q-layer.2": 7916.5083, "encoder_q-layer.3": 8119.2827, "encoder_q-layer.4": 8540.0039, "encoder_q-layer.5": 8898.793, "encoder_q-layer.6": 9817.333, "encoder_q-layer.7": 11573.9043, "encoder_q-layer.8": 14157.4199, "encoder_q-layer.9": 13378.9922, "epoch": 0.63, "inbatch_neg_score": 4.2524, "inbatch_pos_score": 4.8672, "learning_rate": 1.4631578947368422e-05, "loss": 3.3759, "norm_diff": 0.067, "norm_loss": 0.0, "num_token_doc": 66.6942, "num_token_overlap": 11.6954, "num_token_query": 31.484, "num_token_union": 65.1224, "num_word_context": 202.1204, "num_word_doc": 49.7541, "num_word_query": 23.393, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20763.0638, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2539, "query_norm": 2.4031, "queue_k_norm": 2.4705, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.484, "sent_len_1": 66.6942, "sent_len_max_0": 127.985, "sent_len_max_1": 209.5712, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144400 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.377, "doc_norm": 2.4703, "encoder_q-embeddings": 9781.7549, "encoder_q-layer.0": 6640.3838, "encoder_q-layer.1": 6993.8408, "encoder_q-layer.10": 14490.1143, "encoder_q-layer.11": 33113.4492, "encoder_q-layer.2": 7876.3477, "encoder_q-layer.3": 7842.6875, "encoder_q-layer.4": 8532.0439, "encoder_q-layer.5": 9104.3135, "encoder_q-layer.6": 10086.3076, "encoder_q-layer.7": 12039.4355, "encoder_q-layer.8": 13980.6562, "encoder_q-layer.9": 13266.4619, "epoch": 0.63, "inbatch_neg_score": 4.2538, "inbatch_pos_score": 4.8711, "learning_rate": 1.4605263157894739e-05, "loss": 3.377, "norm_diff": 0.0688, "norm_loss": 0.0, "num_token_doc": 66.7579, "num_token_overlap": 11.7158, "num_token_query": 31.4644, "num_token_union": 65.1286, "num_word_context": 202.4813, "num_word_doc": 49.8135, "num_word_query": 23.3766, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20846.5518, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2539, "query_norm": 2.4015, "queue_k_norm": 2.4711, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4644, "sent_len_1": 66.7579, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.1887, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144500 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.3799, "doc_norm": 2.476, "encoder_q-embeddings": 9666.7969, "encoder_q-layer.0": 7021.8296, "encoder_q-layer.1": 7399.4517, "encoder_q-layer.10": 14583.9883, "encoder_q-layer.11": 32686.6211, "encoder_q-layer.2": 8129.8159, "encoder_q-layer.3": 8650.4961, "encoder_q-layer.4": 9237.8174, "encoder_q-layer.5": 9385.3672, "encoder_q-layer.6": 10558.9951, "encoder_q-layer.7": 11862.6172, "encoder_q-layer.8": 14558.7129, "encoder_q-layer.9": 13167.5957, "epoch": 0.63, "inbatch_neg_score": 4.2588, "inbatch_pos_score": 4.8828, "learning_rate": 1.4578947368421053e-05, "loss": 3.3799, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.9662, "num_token_overlap": 11.7072, "num_token_query": 31.4902, "num_token_union": 65.2622, "num_word_context": 202.4018, "num_word_doc": 49.959, "num_word_query": 23.3848, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20764.8343, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2578, "query_norm": 2.4068, "queue_k_norm": 2.4724, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4902, "sent_len_1": 66.9662, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.1438, "stdk": 0.0497, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144600 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3753, "doc_norm": 2.4728, "encoder_q-embeddings": 15140.4648, "encoder_q-layer.0": 11857.4102, "encoder_q-layer.1": 13378.4355, "encoder_q-layer.10": 13073.876, "encoder_q-layer.11": 31873.6953, "encoder_q-layer.2": 14624.5859, "encoder_q-layer.3": 15827.2207, "encoder_q-layer.4": 16415.1211, "encoder_q-layer.5": 16599.0371, "encoder_q-layer.6": 15621.9023, "encoder_q-layer.7": 14517.501, "encoder_q-layer.8": 15842.1377, "encoder_q-layer.9": 13855.8076, "epoch": 0.63, "inbatch_neg_score": 4.2588, "inbatch_pos_score": 4.8828, "learning_rate": 1.455263157894737e-05, "loss": 3.3753, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.6705, "num_token_overlap": 11.6803, "num_token_query": 31.4311, "num_token_union": 65.0842, "num_word_context": 202.1564, "num_word_doc": 49.7568, "num_word_query": 23.3599, "postclip_grad_norm": 1.0, "preclip_grad_norm": 25782.6281, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2578, "query_norm": 2.4013, "queue_k_norm": 2.4734, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4311, "sent_len_1": 66.6705, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.7612, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 144700 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.372, "doc_norm": 2.4733, "encoder_q-embeddings": 10105.8682, "encoder_q-layer.0": 6858.7012, "encoder_q-layer.1": 7092.4795, "encoder_q-layer.10": 13725.7314, "encoder_q-layer.11": 32733.9883, "encoder_q-layer.2": 7819.6934, "encoder_q-layer.3": 7885.7231, "encoder_q-layer.4": 8490.0391, "encoder_q-layer.5": 8730.334, "encoder_q-layer.6": 9688.0293, "encoder_q-layer.7": 10981.5117, "encoder_q-layer.8": 13757.1631, "encoder_q-layer.9": 12540.7646, "epoch": 0.63, "inbatch_neg_score": 4.2637, "inbatch_pos_score": 4.875, "learning_rate": 1.4526315789473685e-05, "loss": 3.372, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.7582, "num_token_overlap": 11.6554, "num_token_query": 31.3689, "num_token_union": 65.1231, "num_word_context": 202.5589, "num_word_doc": 49.8319, "num_word_query": 23.3009, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20373.951, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2617, "query_norm": 2.3991, "queue_k_norm": 2.4734, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3689, "sent_len_1": 66.7582, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.2512, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144800 }, { "accuracy": 49.2676, "active_queue_size": 16384.0, "cl_loss": 3.3747, "doc_norm": 2.4748, "encoder_q-embeddings": 21344.4453, "encoder_q-layer.0": 15591.9395, "encoder_q-layer.1": 18639.291, "encoder_q-layer.10": 13958.1602, "encoder_q-layer.11": 31481.4141, "encoder_q-layer.2": 20803.8359, "encoder_q-layer.3": 22055.1953, "encoder_q-layer.4": 23722.6855, "encoder_q-layer.5": 22979.0156, "encoder_q-layer.6": 23738.502, "encoder_q-layer.7": 21140.5645, "encoder_q-layer.8": 18083.9512, "encoder_q-layer.9": 13816.1426, "epoch": 0.63, "inbatch_neg_score": 4.2657, "inbatch_pos_score": 4.8633, "learning_rate": 1.45e-05, "loss": 3.3747, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.8299, "num_token_overlap": 11.6737, "num_token_query": 31.349, "num_token_union": 65.1454, "num_word_context": 202.466, "num_word_doc": 49.8805, "num_word_query": 23.2758, "postclip_grad_norm": 1.0, "preclip_grad_norm": 31703.1291, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 4.2656, "query_norm": 2.3902, "queue_k_norm": 2.4738, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.349, "sent_len_1": 66.8299, "sent_len_max_0": 127.98, "sent_len_max_1": 208.0087, "stdk": 0.0495, "stdq": 0.0424, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 144900 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3764, "doc_norm": 2.4757, "encoder_q-embeddings": 9392.2207, "encoder_q-layer.0": 6652.2236, "encoder_q-layer.1": 7050.3623, "encoder_q-layer.10": 12713.3008, "encoder_q-layer.11": 31467.1953, "encoder_q-layer.2": 8058.5312, "encoder_q-layer.3": 8044.3203, "encoder_q-layer.4": 8681.3916, "encoder_q-layer.5": 8831.417, "encoder_q-layer.6": 10199.1377, "encoder_q-layer.7": 11391.7334, "encoder_q-layer.8": 13204.6562, "encoder_q-layer.9": 11713.5117, "epoch": 0.63, "inbatch_neg_score": 4.2712, "inbatch_pos_score": 4.8789, "learning_rate": 1.4473684210526317e-05, "loss": 3.3764, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.7605, "num_token_overlap": 11.7107, "num_token_query": 31.4405, "num_token_union": 65.1229, "num_word_context": 202.2804, "num_word_doc": 49.8382, "num_word_query": 23.3605, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19867.0048, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2695, "query_norm": 2.3972, "queue_k_norm": 2.4755, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4405, "sent_len_1": 66.7605, "sent_len_max_0": 127.975, "sent_len_max_1": 208.435, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 145000 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.3748, "doc_norm": 2.4727, "encoder_q-embeddings": 11941.6621, "encoder_q-layer.0": 8747.9961, "encoder_q-layer.1": 9125.1973, "encoder_q-layer.10": 14107.6953, "encoder_q-layer.11": 32949.0508, "encoder_q-layer.2": 10693.8877, "encoder_q-layer.3": 10788.1328, "encoder_q-layer.4": 11459.415, "encoder_q-layer.5": 11591.2959, "encoder_q-layer.6": 12231.1035, "encoder_q-layer.7": 12528.8857, "encoder_q-layer.8": 14345.8984, "encoder_q-layer.9": 12560.5635, "epoch": 0.63, "inbatch_neg_score": 4.2727, "inbatch_pos_score": 4.8789, "learning_rate": 1.4447368421052632e-05, "loss": 3.3748, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.8282, "num_token_overlap": 11.676, "num_token_query": 31.3812, "num_token_union": 65.1359, "num_word_context": 202.3906, "num_word_doc": 49.8792, "num_word_query": 23.3001, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22103.3116, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2734, "query_norm": 2.4036, "queue_k_norm": 2.4754, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3812, "sent_len_1": 66.8282, "sent_len_max_0": 127.9813, "sent_len_max_1": 211.37, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145100 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.3745, "doc_norm": 2.4763, "encoder_q-embeddings": 9278.873, "encoder_q-layer.0": 6498.1108, "encoder_q-layer.1": 7091.2632, "encoder_q-layer.10": 12570.6543, "encoder_q-layer.11": 31899.418, "encoder_q-layer.2": 7647.6533, "encoder_q-layer.3": 7941.9414, "encoder_q-layer.4": 8306.8486, "encoder_q-layer.5": 8721.3213, "encoder_q-layer.6": 9714.7451, "encoder_q-layer.7": 10895.5879, "encoder_q-layer.8": 13178.9385, "encoder_q-layer.9": 12245.8379, "epoch": 0.63, "inbatch_neg_score": 4.2768, "inbatch_pos_score": 4.875, "learning_rate": 1.4421052631578948e-05, "loss": 3.3745, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.8285, "num_token_overlap": 11.7186, "num_token_query": 31.4769, "num_token_union": 65.2023, "num_word_context": 202.3786, "num_word_doc": 49.897, "num_word_query": 23.3862, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19809.1922, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2734, "query_norm": 2.4007, "queue_k_norm": 2.4756, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4769, "sent_len_1": 66.8285, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.83, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145200 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3836, "doc_norm": 2.4793, "encoder_q-embeddings": 9259.0684, "encoder_q-layer.0": 6390.7871, "encoder_q-layer.1": 6804.7686, "encoder_q-layer.10": 13314.0635, "encoder_q-layer.11": 33047.457, "encoder_q-layer.2": 7816.147, "encoder_q-layer.3": 8221.9072, "encoder_q-layer.4": 8426.5029, "encoder_q-layer.5": 8505.082, "encoder_q-layer.6": 9450.3496, "encoder_q-layer.7": 10960.5205, "encoder_q-layer.8": 13070.6523, "encoder_q-layer.9": 12433.5352, "epoch": 0.63, "inbatch_neg_score": 4.2755, "inbatch_pos_score": 4.8867, "learning_rate": 1.4394736842105263e-05, "loss": 3.3836, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.7817, "num_token_overlap": 11.6379, "num_token_query": 31.3537, "num_token_union": 65.1351, "num_word_context": 202.2876, "num_word_doc": 49.8172, "num_word_query": 23.2791, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20371.0547, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2773, "query_norm": 2.402, "queue_k_norm": 2.4754, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3537, "sent_len_1": 66.7817, "sent_len_max_0": 127.995, "sent_len_max_1": 209.1525, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 145300 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3776, "doc_norm": 2.4739, "encoder_q-embeddings": 9038.6348, "encoder_q-layer.0": 6359.6807, "encoder_q-layer.1": 6914.0474, "encoder_q-layer.10": 12676.748, "encoder_q-layer.11": 30952.3633, "encoder_q-layer.2": 7832.8174, "encoder_q-layer.3": 8099.4883, "encoder_q-layer.4": 8580.3223, "encoder_q-layer.5": 8756.8223, "encoder_q-layer.6": 9767.7998, "encoder_q-layer.7": 11176.2627, "encoder_q-layer.8": 13148.4834, "encoder_q-layer.9": 12237.2451, "epoch": 0.63, "inbatch_neg_score": 4.2805, "inbatch_pos_score": 4.8867, "learning_rate": 1.4368421052631582e-05, "loss": 3.3776, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.7705, "num_token_overlap": 11.6825, "num_token_query": 31.3875, "num_token_union": 65.1289, "num_word_context": 202.2661, "num_word_doc": 49.8073, "num_word_query": 23.301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19268.5362, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2812, "query_norm": 2.3983, "queue_k_norm": 2.4764, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3875, "sent_len_1": 66.7705, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.4112, "stdk": 0.0492, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145400 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3763, "doc_norm": 2.4789, "encoder_q-embeddings": 19435.0, "encoder_q-layer.0": 13786.6602, "encoder_q-layer.1": 15220.8789, "encoder_q-layer.10": 26248.6621, "encoder_q-layer.11": 62478.6445, "encoder_q-layer.2": 16780.873, "encoder_q-layer.3": 16781.3906, "encoder_q-layer.4": 16865.627, "encoder_q-layer.5": 17212.5449, "encoder_q-layer.6": 19418.1797, "encoder_q-layer.7": 21897.8555, "encoder_q-layer.8": 26132.748, "encoder_q-layer.9": 24448.043, "epoch": 0.63, "inbatch_neg_score": 4.2861, "inbatch_pos_score": 4.8984, "learning_rate": 1.4342105263157895e-05, "loss": 3.3763, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.6357, "num_token_overlap": 11.6405, "num_token_query": 31.2855, "num_token_union": 65.0283, "num_word_context": 202.057, "num_word_doc": 49.7334, "num_word_query": 23.2201, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39281.8507, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.2852, "query_norm": 2.4061, "queue_k_norm": 2.4773, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2855, "sent_len_1": 66.6357, "sent_len_max_0": 127.98, "sent_len_max_1": 208.1262, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145500 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3746, "doc_norm": 2.4773, "encoder_q-embeddings": 11315.6865, "encoder_q-layer.0": 7827.9751, "encoder_q-layer.1": 8544.6963, "encoder_q-layer.10": 13039.8975, "encoder_q-layer.11": 32118.9199, "encoder_q-layer.2": 9528.0898, "encoder_q-layer.3": 10154.3047, "encoder_q-layer.4": 11249.4072, "encoder_q-layer.5": 11658.8418, "encoder_q-layer.6": 12341.2588, "encoder_q-layer.7": 13296.2461, "encoder_q-layer.8": 13998.5664, "encoder_q-layer.9": 12312.9053, "epoch": 0.63, "inbatch_neg_score": 4.2901, "inbatch_pos_score": 4.9023, "learning_rate": 1.431578947368421e-05, "loss": 3.3746, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.828, "num_token_overlap": 11.6858, "num_token_query": 31.4045, "num_token_union": 65.1533, "num_word_context": 202.4476, "num_word_doc": 49.8249, "num_word_query": 23.3218, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21618.2661, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2891, "query_norm": 2.406, "queue_k_norm": 2.4781, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4045, "sent_len_1": 66.828, "sent_len_max_0": 127.985, "sent_len_max_1": 210.7225, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145600 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3648, "doc_norm": 2.4814, "encoder_q-embeddings": 9735.916, "encoder_q-layer.0": 6724.1235, "encoder_q-layer.1": 7240.9834, "encoder_q-layer.10": 12765.3125, "encoder_q-layer.11": 32382.9395, "encoder_q-layer.2": 7897.5132, "encoder_q-layer.3": 8006.8779, "encoder_q-layer.4": 8520.1748, "encoder_q-layer.5": 8933.6006, "encoder_q-layer.6": 9832.0938, "encoder_q-layer.7": 11138.8535, "encoder_q-layer.8": 12912.8477, "encoder_q-layer.9": 11958.6963, "epoch": 0.63, "inbatch_neg_score": 4.2933, "inbatch_pos_score": 4.8945, "learning_rate": 1.4289473684210528e-05, "loss": 3.3648, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.7112, "num_token_overlap": 11.6987, "num_token_query": 31.3868, "num_token_union": 65.0383, "num_word_context": 202.3179, "num_word_doc": 49.7984, "num_word_query": 23.3146, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20192.4592, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.293, "query_norm": 2.4075, "queue_k_norm": 2.4788, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3868, "sent_len_1": 66.7112, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.0288, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 145700 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3688, "doc_norm": 2.4818, "encoder_q-embeddings": 8933.5938, "encoder_q-layer.0": 6421.2505, "encoder_q-layer.1": 6894.3491, "encoder_q-layer.10": 14129.3086, "encoder_q-layer.11": 33067.043, "encoder_q-layer.2": 7696.9199, "encoder_q-layer.3": 8027.9854, "encoder_q-layer.4": 8448.2227, "encoder_q-layer.5": 8901.9238, "encoder_q-layer.6": 10027.5361, "encoder_q-layer.7": 11261.6562, "encoder_q-layer.8": 14247.3057, "encoder_q-layer.9": 12606.4727, "epoch": 0.63, "inbatch_neg_score": 4.2962, "inbatch_pos_score": 4.9102, "learning_rate": 1.4263157894736842e-05, "loss": 3.3688, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.7255, "num_token_overlap": 11.6503, "num_token_query": 31.2971, "num_token_union": 65.0625, "num_word_context": 202.2087, "num_word_doc": 49.7998, "num_word_query": 23.2403, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20398.1967, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.2969, "query_norm": 2.4042, "queue_k_norm": 2.4784, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2971, "sent_len_1": 66.7255, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.2763, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 145800 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.368, "doc_norm": 2.4767, "encoder_q-embeddings": 9953.3818, "encoder_q-layer.0": 6878.1895, "encoder_q-layer.1": 7541.4341, "encoder_q-layer.10": 13823.3838, "encoder_q-layer.11": 33391.6562, "encoder_q-layer.2": 8368.3457, "encoder_q-layer.3": 8860.4854, "encoder_q-layer.4": 9525.5986, "encoder_q-layer.5": 9606.1865, "encoder_q-layer.6": 10793.4062, "encoder_q-layer.7": 12100.8926, "encoder_q-layer.8": 14865.7588, "encoder_q-layer.9": 13751.2764, "epoch": 0.63, "inbatch_neg_score": 4.2929, "inbatch_pos_score": 4.9062, "learning_rate": 1.423684210526316e-05, "loss": 3.368, "norm_diff": 0.0674, "norm_loss": 0.0, "num_token_doc": 66.8311, "num_token_overlap": 11.6621, "num_token_query": 31.3889, "num_token_union": 65.1912, "num_word_context": 202.2876, "num_word_doc": 49.8577, "num_word_query": 23.3107, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21499.8315, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.293, "query_norm": 2.4092, "queue_k_norm": 2.4815, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3889, "sent_len_1": 66.8311, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.0125, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 145900 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3715, "doc_norm": 2.483, "encoder_q-embeddings": 9516.0117, "encoder_q-layer.0": 6913.7021, "encoder_q-layer.1": 7233.999, "encoder_q-layer.10": 12375.3564, "encoder_q-layer.11": 30892.4336, "encoder_q-layer.2": 7922.439, "encoder_q-layer.3": 7809.002, "encoder_q-layer.4": 8343.1084, "encoder_q-layer.5": 8498.4551, "encoder_q-layer.6": 9647.5684, "encoder_q-layer.7": 10324.9375, "encoder_q-layer.8": 12868.6582, "encoder_q-layer.9": 11917.6357, "epoch": 0.63, "inbatch_neg_score": 4.3028, "inbatch_pos_score": 4.9219, "learning_rate": 1.4210526315789475e-05, "loss": 3.3715, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.7783, "num_token_overlap": 11.6291, "num_token_query": 31.2372, "num_token_union": 65.106, "num_word_context": 202.4113, "num_word_doc": 49.8726, "num_word_query": 23.209, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19676.0104, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3008, "query_norm": 2.4079, "queue_k_norm": 2.4802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2372, "sent_len_1": 66.7783, "sent_len_max_0": 127.985, "sent_len_max_1": 207.485, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 146000 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3752, "doc_norm": 2.4821, "encoder_q-embeddings": 10005.9307, "encoder_q-layer.0": 7023.8618, "encoder_q-layer.1": 7550.2612, "encoder_q-layer.10": 14192.873, "encoder_q-layer.11": 33342.2773, "encoder_q-layer.2": 8692.2979, "encoder_q-layer.3": 8860.5176, "encoder_q-layer.4": 10054.5293, "encoder_q-layer.5": 10341.6787, "encoder_q-layer.6": 10545.2715, "encoder_q-layer.7": 11747.8232, "encoder_q-layer.8": 14112.6826, "encoder_q-layer.9": 12816.4873, "epoch": 0.63, "inbatch_neg_score": 4.306, "inbatch_pos_score": 4.9062, "learning_rate": 1.4184210526315792e-05, "loss": 3.3752, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.7423, "num_token_overlap": 11.6761, "num_token_query": 31.4489, "num_token_union": 65.1689, "num_word_context": 202.3921, "num_word_doc": 49.7895, "num_word_query": 23.3551, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21193.7483, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3047, "query_norm": 2.4053, "queue_k_norm": 2.4817, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4489, "sent_len_1": 66.7423, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.985, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146100 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3858, "doc_norm": 2.4787, "encoder_q-embeddings": 9786.1797, "encoder_q-layer.0": 6842.5327, "encoder_q-layer.1": 7661.3086, "encoder_q-layer.10": 13445.625, "encoder_q-layer.11": 31330.8711, "encoder_q-layer.2": 8298.1406, "encoder_q-layer.3": 8518.7666, "encoder_q-layer.4": 8969.1875, "encoder_q-layer.5": 8834.334, "encoder_q-layer.6": 9942.7471, "encoder_q-layer.7": 11253.1016, "encoder_q-layer.8": 13537.5029, "encoder_q-layer.9": 12336.5566, "epoch": 0.63, "inbatch_neg_score": 4.3087, "inbatch_pos_score": 4.9219, "learning_rate": 1.4157894736842107e-05, "loss": 3.3858, "norm_diff": 0.065, "norm_loss": 0.0, "num_token_doc": 66.6433, "num_token_overlap": 11.6259, "num_token_query": 31.296, "num_token_union": 65.0435, "num_word_context": 202.4154, "num_word_doc": 49.7549, "num_word_query": 23.2568, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20032.8607, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3086, "query_norm": 2.4137, "queue_k_norm": 2.4822, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.296, "sent_len_1": 66.6433, "sent_len_max_0": 127.935, "sent_len_max_1": 209.3738, "stdk": 0.0492, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146200 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3702, "doc_norm": 2.4832, "encoder_q-embeddings": 9387.1445, "encoder_q-layer.0": 6520.4619, "encoder_q-layer.1": 6899.1011, "encoder_q-layer.10": 13051.2637, "encoder_q-layer.11": 33389.7305, "encoder_q-layer.2": 7619.208, "encoder_q-layer.3": 7887.1587, "encoder_q-layer.4": 8494.417, "encoder_q-layer.5": 8515.4639, "encoder_q-layer.6": 9686.3779, "encoder_q-layer.7": 10899.54, "encoder_q-layer.8": 12867.3525, "encoder_q-layer.9": 12414.2734, "epoch": 0.63, "inbatch_neg_score": 4.3121, "inbatch_pos_score": 4.9258, "learning_rate": 1.413157894736842e-05, "loss": 3.3702, "norm_diff": 0.0737, "norm_loss": 0.0, "num_token_doc": 66.842, "num_token_overlap": 11.671, "num_token_query": 31.3695, "num_token_union": 65.1437, "num_word_context": 202.2824, "num_word_doc": 49.8475, "num_word_query": 23.2881, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20576.3896, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3125, "query_norm": 2.4095, "queue_k_norm": 2.4824, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3695, "sent_len_1": 66.842, "sent_len_max_0": 127.9887, "sent_len_max_1": 211.2625, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 146300 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3631, "doc_norm": 2.4857, "encoder_q-embeddings": 19484.543, "encoder_q-layer.0": 14310.0342, "encoder_q-layer.1": 17898.3457, "encoder_q-layer.10": 14724.2012, "encoder_q-layer.11": 33809.8477, "encoder_q-layer.2": 21360.5742, "encoder_q-layer.3": 23189.8164, "encoder_q-layer.4": 23657.582, "encoder_q-layer.5": 25861.7422, "encoder_q-layer.6": 26610.9258, "encoder_q-layer.7": 23373.1172, "encoder_q-layer.8": 19922.7988, "encoder_q-layer.9": 13885.4238, "epoch": 0.64, "inbatch_neg_score": 4.3169, "inbatch_pos_score": 4.9336, "learning_rate": 1.4105263157894738e-05, "loss": 3.3631, "norm_diff": 0.0686, "norm_loss": 0.0, "num_token_doc": 66.7028, "num_token_overlap": 11.7039, "num_token_query": 31.5059, "num_token_union": 65.1163, "num_word_context": 202.3169, "num_word_doc": 49.7588, "num_word_query": 23.4071, "postclip_grad_norm": 1.0, "preclip_grad_norm": 33347.0301, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 4.3164, "query_norm": 2.417, "queue_k_norm": 2.4844, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5059, "sent_len_1": 66.7028, "sent_len_max_0": 127.9737, "sent_len_max_1": 211.1037, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146400 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3712, "doc_norm": 2.4824, "encoder_q-embeddings": 9369.3105, "encoder_q-layer.0": 6487.2363, "encoder_q-layer.1": 6920.0938, "encoder_q-layer.10": 12592.2832, "encoder_q-layer.11": 30464.5977, "encoder_q-layer.2": 7958.7246, "encoder_q-layer.3": 8264.6875, "encoder_q-layer.4": 8729.5469, "encoder_q-layer.5": 8823.9102, "encoder_q-layer.6": 9913.834, "encoder_q-layer.7": 11671.4219, "encoder_q-layer.8": 12848.9932, "encoder_q-layer.9": 11911.3203, "epoch": 0.64, "inbatch_neg_score": 4.3206, "inbatch_pos_score": 4.9414, "learning_rate": 1.4078947368421053e-05, "loss": 3.3712, "norm_diff": 0.0661, "norm_loss": 0.0, "num_token_doc": 66.7527, "num_token_overlap": 11.6694, "num_token_query": 31.4382, "num_token_union": 65.1338, "num_word_context": 202.3063, "num_word_doc": 49.8129, "num_word_query": 23.3497, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19240.3855, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3203, "query_norm": 2.4163, "queue_k_norm": 2.4844, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4382, "sent_len_1": 66.7527, "sent_len_max_0": 127.98, "sent_len_max_1": 208.3175, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146500 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3686, "doc_norm": 2.4826, "encoder_q-embeddings": 31701.4648, "encoder_q-layer.0": 24597.0391, "encoder_q-layer.1": 28215.1836, "encoder_q-layer.10": 13090.0342, "encoder_q-layer.11": 31448.3945, "encoder_q-layer.2": 34872.8867, "encoder_q-layer.3": 36594.5742, "encoder_q-layer.4": 39582.1719, "encoder_q-layer.5": 37979.7695, "encoder_q-layer.6": 39338.1289, "encoder_q-layer.7": 36419.8438, "encoder_q-layer.8": 25437.1914, "encoder_q-layer.9": 13860.6084, "epoch": 0.64, "inbatch_neg_score": 4.3238, "inbatch_pos_score": 4.9375, "learning_rate": 1.405263157894737e-05, "loss": 3.3686, "norm_diff": 0.0635, "norm_loss": 0.0, "num_token_doc": 66.8644, "num_token_overlap": 11.708, "num_token_query": 31.4644, "num_token_union": 65.2467, "num_word_context": 202.4409, "num_word_doc": 49.9102, "num_word_query": 23.383, "postclip_grad_norm": 1.0, "preclip_grad_norm": 47654.1029, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.3242, "query_norm": 2.4191, "queue_k_norm": 2.4854, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4644, "sent_len_1": 66.8644, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.4613, "stdk": 0.0492, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146600 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3643, "doc_norm": 2.4821, "encoder_q-embeddings": 10180.0547, "encoder_q-layer.0": 7069.686, "encoder_q-layer.1": 7443.1279, "encoder_q-layer.10": 12757.6602, "encoder_q-layer.11": 31599.1738, "encoder_q-layer.2": 8110.4829, "encoder_q-layer.3": 8234.4658, "encoder_q-layer.4": 8731.8115, "encoder_q-layer.5": 9160.6797, "encoder_q-layer.6": 10111.6367, "encoder_q-layer.7": 11584.8389, "encoder_q-layer.8": 13386.0986, "encoder_q-layer.9": 12346.6484, "epoch": 0.64, "inbatch_neg_score": 4.3248, "inbatch_pos_score": 4.9453, "learning_rate": 1.4026315789473685e-05, "loss": 3.3643, "norm_diff": 0.0598, "norm_loss": 0.0, "num_token_doc": 66.754, "num_token_overlap": 11.6452, "num_token_query": 31.3413, "num_token_union": 65.1139, "num_word_context": 202.1687, "num_word_doc": 49.8436, "num_word_query": 23.2858, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20244.6158, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3242, "query_norm": 2.4223, "queue_k_norm": 2.485, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3413, "sent_len_1": 66.754, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.8212, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146700 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3683, "doc_norm": 2.4854, "encoder_q-embeddings": 9635.71, "encoder_q-layer.0": 6649.6987, "encoder_q-layer.1": 7171.271, "encoder_q-layer.10": 13249.0732, "encoder_q-layer.11": 32936.9336, "encoder_q-layer.2": 8013.5493, "encoder_q-layer.3": 8133.7993, "encoder_q-layer.4": 8745.1953, "encoder_q-layer.5": 8793.8301, "encoder_q-layer.6": 9865.6309, "encoder_q-layer.7": 11234.4404, "encoder_q-layer.8": 13541.2939, "encoder_q-layer.9": 12414.9863, "epoch": 0.64, "inbatch_neg_score": 4.3323, "inbatch_pos_score": 4.9375, "learning_rate": 1.4000000000000001e-05, "loss": 3.3683, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.7962, "num_token_overlap": 11.6833, "num_token_query": 31.4383, "num_token_union": 65.1912, "num_word_context": 202.2514, "num_word_doc": 49.821, "num_word_query": 23.345, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20474.5478, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.332, "query_norm": 2.4137, "queue_k_norm": 2.4848, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4383, "sent_len_1": 66.7962, "sent_len_max_0": 127.97, "sent_len_max_1": 209.5188, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 146800 }, { "accuracy": 48.9258, "active_queue_size": 16384.0, "cl_loss": 3.3737, "doc_norm": 2.4853, "encoder_q-embeddings": 9265.0088, "encoder_q-layer.0": 6484.2637, "encoder_q-layer.1": 6937.2319, "encoder_q-layer.10": 13249.71, "encoder_q-layer.11": 33425.4492, "encoder_q-layer.2": 7822.814, "encoder_q-layer.3": 7950.6982, "encoder_q-layer.4": 8557.7256, "encoder_q-layer.5": 8802.3174, "encoder_q-layer.6": 9857.7363, "encoder_q-layer.7": 11469.3047, "encoder_q-layer.8": 14314.5664, "encoder_q-layer.9": 12940.1016, "epoch": 0.64, "inbatch_neg_score": 4.3388, "inbatch_pos_score": 4.9414, "learning_rate": 1.3973684210526316e-05, "loss": 3.3737, "norm_diff": 0.0667, "norm_loss": 0.0, "num_token_doc": 66.7268, "num_token_overlap": 11.6632, "num_token_query": 31.4059, "num_token_union": 65.1241, "num_word_context": 202.4894, "num_word_doc": 49.8289, "num_word_query": 23.3443, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21030.2852, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3398, "query_norm": 2.4186, "queue_k_norm": 2.4875, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4059, "sent_len_1": 66.7268, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.3137, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 146900 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.376, "doc_norm": 2.4897, "encoder_q-embeddings": 9502.5488, "encoder_q-layer.0": 6728.9746, "encoder_q-layer.1": 7023.2695, "encoder_q-layer.10": 13032.043, "encoder_q-layer.11": 30966.2168, "encoder_q-layer.2": 7640.7803, "encoder_q-layer.3": 7849.3936, "encoder_q-layer.4": 8094.5244, "encoder_q-layer.5": 8178.9346, "encoder_q-layer.6": 9452.8379, "encoder_q-layer.7": 11422.6309, "encoder_q-layer.8": 13048.5547, "encoder_q-layer.9": 11841.6367, "epoch": 0.64, "inbatch_neg_score": 4.3433, "inbatch_pos_score": 4.9492, "learning_rate": 1.3947368421052631e-05, "loss": 3.376, "norm_diff": 0.0686, "norm_loss": 0.0, "num_token_doc": 66.6984, "num_token_overlap": 11.6829, "num_token_query": 31.3939, "num_token_union": 65.086, "num_word_context": 202.3514, "num_word_doc": 49.7797, "num_word_query": 23.3236, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19649.1553, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3438, "query_norm": 2.4211, "queue_k_norm": 2.4863, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3939, "sent_len_1": 66.6984, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.8975, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147000 }, { "accuracy": 53.5156, "active_queue_size": 16384.0, "cl_loss": 3.3634, "doc_norm": 2.4857, "encoder_q-embeddings": 10136.1816, "encoder_q-layer.0": 7004.7183, "encoder_q-layer.1": 7528.1299, "encoder_q-layer.10": 15209.3408, "encoder_q-layer.11": 32405.9199, "encoder_q-layer.2": 8181.6328, "encoder_q-layer.3": 8532.96, "encoder_q-layer.4": 9505.9414, "encoder_q-layer.5": 9774.0928, "encoder_q-layer.6": 10746.874, "encoder_q-layer.7": 12034.1064, "encoder_q-layer.8": 14208.8379, "encoder_q-layer.9": 13944.54, "epoch": 0.64, "inbatch_neg_score": 4.3493, "inbatch_pos_score": 4.9727, "learning_rate": 1.3921052631578948e-05, "loss": 3.3634, "norm_diff": 0.0549, "norm_loss": 0.0, "num_token_doc": 66.7037, "num_token_overlap": 11.7135, "num_token_query": 31.4061, "num_token_union": 65.0488, "num_word_context": 202.4065, "num_word_doc": 49.7632, "num_word_query": 23.3302, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20912.5347, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3516, "query_norm": 2.4308, "queue_k_norm": 2.4872, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4061, "sent_len_1": 66.7037, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.6438, "stdk": 0.0492, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147100 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3681, "doc_norm": 2.4897, "encoder_q-embeddings": 10402.2061, "encoder_q-layer.0": 7291.9805, "encoder_q-layer.1": 7680.8843, "encoder_q-layer.10": 13816.0928, "encoder_q-layer.11": 31099.3359, "encoder_q-layer.2": 8648.5049, "encoder_q-layer.3": 8635.0547, "encoder_q-layer.4": 8921.6758, "encoder_q-layer.5": 9118.8643, "encoder_q-layer.6": 9689.751, "encoder_q-layer.7": 10889.8252, "encoder_q-layer.8": 12837.5918, "encoder_q-layer.9": 12188.1875, "epoch": 0.64, "inbatch_neg_score": 4.3588, "inbatch_pos_score": 4.9648, "learning_rate": 1.3894736842105263e-05, "loss": 3.3681, "norm_diff": 0.0615, "norm_loss": 0.0, "num_token_doc": 66.8943, "num_token_overlap": 11.6908, "num_token_query": 31.3688, "num_token_union": 65.17, "num_word_context": 202.5671, "num_word_doc": 49.9534, "num_word_query": 23.3015, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20003.1324, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3594, "query_norm": 2.4281, "queue_k_norm": 2.4891, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3688, "sent_len_1": 66.8943, "sent_len_max_0": 127.985, "sent_len_max_1": 207.3413, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 147200 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3589, "doc_norm": 2.4891, "encoder_q-embeddings": 9486.4492, "encoder_q-layer.0": 6825.9229, "encoder_q-layer.1": 7211.1431, "encoder_q-layer.10": 12120.6846, "encoder_q-layer.11": 30477.4727, "encoder_q-layer.2": 8062.4468, "encoder_q-layer.3": 8225.0898, "encoder_q-layer.4": 8828.0049, "encoder_q-layer.5": 9024.1973, "encoder_q-layer.6": 9987.5186, "encoder_q-layer.7": 12047.707, "encoder_q-layer.8": 13548.9844, "encoder_q-layer.9": 12339.0049, "epoch": 0.64, "inbatch_neg_score": 4.3664, "inbatch_pos_score": 4.9844, "learning_rate": 1.386842105263158e-05, "loss": 3.3589, "norm_diff": 0.0569, "norm_loss": 0.0, "num_token_doc": 66.7683, "num_token_overlap": 11.6796, "num_token_query": 31.305, "num_token_union": 65.0662, "num_word_context": 202.3281, "num_word_doc": 49.816, "num_word_query": 23.2517, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19824.043, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3672, "query_norm": 2.4322, "queue_k_norm": 2.4899, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.305, "sent_len_1": 66.7683, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.9625, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 147300 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.3815, "doc_norm": 2.4886, "encoder_q-embeddings": 9743.5518, "encoder_q-layer.0": 7145.377, "encoder_q-layer.1": 7574.604, "encoder_q-layer.10": 13189.5557, "encoder_q-layer.11": 32067.9961, "encoder_q-layer.2": 8618.3779, "encoder_q-layer.3": 8534.3486, "encoder_q-layer.4": 8672.9785, "encoder_q-layer.5": 8975.2832, "encoder_q-layer.6": 10083.1777, "encoder_q-layer.7": 11518.1191, "encoder_q-layer.8": 14546.9434, "encoder_q-layer.9": 13089.2207, "epoch": 0.64, "inbatch_neg_score": 4.3744, "inbatch_pos_score": 4.9727, "learning_rate": 1.3842105263157895e-05, "loss": 3.3815, "norm_diff": 0.0591, "norm_loss": 0.0, "num_token_doc": 66.6267, "num_token_overlap": 11.6908, "num_token_query": 31.4511, "num_token_union": 65.0833, "num_word_context": 202.3147, "num_word_doc": 49.7063, "num_word_query": 23.3728, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20260.9723, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.375, "query_norm": 2.4295, "queue_k_norm": 2.4905, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4511, "sent_len_1": 66.6267, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.0337, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147400 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3748, "doc_norm": 2.4883, "encoder_q-embeddings": 10296.5547, "encoder_q-layer.0": 7364.7231, "encoder_q-layer.1": 7953.3291, "encoder_q-layer.10": 14912.0127, "encoder_q-layer.11": 34016.2227, "encoder_q-layer.2": 8842.6475, "encoder_q-layer.3": 8888.1641, "encoder_q-layer.4": 9941.9668, "encoder_q-layer.5": 10254.9141, "encoder_q-layer.6": 11644.0703, "encoder_q-layer.7": 13415.8135, "encoder_q-layer.8": 15580.2627, "encoder_q-layer.9": 13971.8633, "epoch": 0.64, "inbatch_neg_score": 4.381, "inbatch_pos_score": 4.9883, "learning_rate": 1.3815789473684213e-05, "loss": 3.3748, "norm_diff": 0.0565, "norm_loss": 0.0, "num_token_doc": 66.7656, "num_token_overlap": 11.6631, "num_token_query": 31.3538, "num_token_union": 65.0814, "num_word_context": 202.2918, "num_word_doc": 49.7977, "num_word_query": 23.2924, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21984.9992, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.3828, "query_norm": 2.4318, "queue_k_norm": 2.4908, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3538, "sent_len_1": 66.7656, "sent_len_max_0": 127.9762, "sent_len_max_1": 210.82, "stdk": 0.0492, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147500 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.364, "doc_norm": 2.4928, "encoder_q-embeddings": 19125.8789, "encoder_q-layer.0": 14196.0518, "encoder_q-layer.1": 15323.6436, "encoder_q-layer.10": 26010.1699, "encoder_q-layer.11": 61407.543, "encoder_q-layer.2": 16578.0254, "encoder_q-layer.3": 16524.6699, "encoder_q-layer.4": 17401.1836, "encoder_q-layer.5": 18586.6973, "encoder_q-layer.6": 20890.9043, "encoder_q-layer.7": 22859.1348, "encoder_q-layer.8": 26657.2266, "encoder_q-layer.9": 24626.8906, "epoch": 0.64, "inbatch_neg_score": 4.389, "inbatch_pos_score": 5.0039, "learning_rate": 1.3789473684210526e-05, "loss": 3.364, "norm_diff": 0.0532, "norm_loss": 0.0, "num_token_doc": 66.7227, "num_token_overlap": 11.6816, "num_token_query": 31.3922, "num_token_union": 65.064, "num_word_context": 202.1282, "num_word_doc": 49.7583, "num_word_query": 23.3127, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39471.8226, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.3906, "query_norm": 2.4396, "queue_k_norm": 2.4922, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3922, "sent_len_1": 66.7227, "sent_len_max_0": 127.9363, "sent_len_max_1": 209.41, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147600 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3632, "doc_norm": 2.4932, "encoder_q-embeddings": 20592.791, "encoder_q-layer.0": 13489.2393, "encoder_q-layer.1": 14953.751, "encoder_q-layer.10": 28147.3867, "encoder_q-layer.11": 66865.5234, "encoder_q-layer.2": 16995.5547, "encoder_q-layer.3": 17490.625, "encoder_q-layer.4": 18550.3926, "encoder_q-layer.5": 19826.3457, "encoder_q-layer.6": 22255.1289, "encoder_q-layer.7": 24311.1094, "encoder_q-layer.8": 29519.0117, "encoder_q-layer.9": 27485.5059, "epoch": 0.64, "inbatch_neg_score": 4.3956, "inbatch_pos_score": 4.9922, "learning_rate": 1.3763157894736841e-05, "loss": 3.3632, "norm_diff": 0.064, "norm_loss": 0.0, "num_token_doc": 66.975, "num_token_overlap": 11.6774, "num_token_query": 31.3738, "num_token_union": 65.2496, "num_word_context": 202.593, "num_word_doc": 49.9585, "num_word_query": 23.2893, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43460.4188, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.3945, "query_norm": 2.4291, "queue_k_norm": 2.4936, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3738, "sent_len_1": 66.975, "sent_len_max_0": 127.99, "sent_len_max_1": 209.705, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 147700 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3637, "doc_norm": 2.4945, "encoder_q-embeddings": 19953.3438, "encoder_q-layer.0": 13981.0049, "encoder_q-layer.1": 15138.9717, "encoder_q-layer.10": 26958.7285, "encoder_q-layer.11": 63755.7539, "encoder_q-layer.2": 17547.4492, "encoder_q-layer.3": 17810.5391, "encoder_q-layer.4": 19408.2891, "encoder_q-layer.5": 19390.6777, "encoder_q-layer.6": 21927.707, "encoder_q-layer.7": 24276.0859, "encoder_q-layer.8": 27540.209, "encoder_q-layer.9": 24892.3027, "epoch": 0.64, "inbatch_neg_score": 4.4006, "inbatch_pos_score": 5.0078, "learning_rate": 1.373684210526316e-05, "loss": 3.3637, "norm_diff": 0.0608, "norm_loss": 0.0, "num_token_doc": 66.8425, "num_token_overlap": 11.698, "num_token_query": 31.3803, "num_token_union": 65.1645, "num_word_context": 202.2488, "num_word_doc": 49.8599, "num_word_query": 23.3042, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41490.4351, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4023, "query_norm": 2.4337, "queue_k_norm": 2.4955, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3803, "sent_len_1": 66.8425, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.535, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 147800 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.3596, "doc_norm": 2.4966, "encoder_q-embeddings": 23356.0547, "encoder_q-layer.0": 17771.6445, "encoder_q-layer.1": 20238.8223, "encoder_q-layer.10": 25568.582, "encoder_q-layer.11": 64367.3867, "encoder_q-layer.2": 24713.0234, "encoder_q-layer.3": 24506.6992, "encoder_q-layer.4": 23864.623, "encoder_q-layer.5": 22630.4375, "encoder_q-layer.6": 23181.125, "encoder_q-layer.7": 24867.8203, "encoder_q-layer.8": 28304.2871, "encoder_q-layer.9": 24753.7148, "epoch": 0.64, "inbatch_neg_score": 4.408, "inbatch_pos_score": 5.0117, "learning_rate": 1.3710526315789473e-05, "loss": 3.3596, "norm_diff": 0.0637, "norm_loss": 0.0, "num_token_doc": 66.8148, "num_token_overlap": 11.6799, "num_token_query": 31.3653, "num_token_union": 65.146, "num_word_context": 202.1926, "num_word_doc": 49.8439, "num_word_query": 23.2989, "postclip_grad_norm": 1.0, "preclip_grad_norm": 44210.5469, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4062, "query_norm": 2.4329, "queue_k_norm": 2.4968, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3653, "sent_len_1": 66.8148, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.7962, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 147900 }, { "accuracy": 49.0234, "active_queue_size": 16384.0, "cl_loss": 3.3585, "doc_norm": 2.4964, "encoder_q-embeddings": 18856.9121, "encoder_q-layer.0": 13792.8896, "encoder_q-layer.1": 14425.1709, "encoder_q-layer.10": 27745.8066, "encoder_q-layer.11": 67566.3125, "encoder_q-layer.2": 16250.4082, "encoder_q-layer.3": 16629.9102, "encoder_q-layer.4": 17755.1309, "encoder_q-layer.5": 18418.5684, "encoder_q-layer.6": 20523.4727, "encoder_q-layer.7": 24487.2285, "encoder_q-layer.8": 29721.0059, "encoder_q-layer.9": 26223.5137, "epoch": 0.64, "inbatch_neg_score": 4.412, "inbatch_pos_score": 5.0156, "learning_rate": 1.3684210526315791e-05, "loss": 3.3585, "norm_diff": 0.0617, "norm_loss": 0.0, "num_token_doc": 66.8147, "num_token_overlap": 11.6991, "num_token_query": 31.4059, "num_token_union": 65.1882, "num_word_context": 202.4708, "num_word_doc": 49.8716, "num_word_query": 23.3292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41888.9279, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4141, "query_norm": 2.4347, "queue_k_norm": 2.4976, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4059, "sent_len_1": 66.8147, "sent_len_max_0": 127.98, "sent_len_max_1": 209.5012, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 148000 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3711, "doc_norm": 2.5002, "encoder_q-embeddings": 19846.7363, "encoder_q-layer.0": 13845.7324, "encoder_q-layer.1": 14781.6045, "encoder_q-layer.10": 31338.4844, "encoder_q-layer.11": 70850.0312, "encoder_q-layer.2": 16804.6484, "encoder_q-layer.3": 17687.9453, "encoder_q-layer.4": 18980.2637, "encoder_q-layer.5": 19972.1973, "encoder_q-layer.6": 22359.4551, "encoder_q-layer.7": 24009.6797, "encoder_q-layer.8": 30921.5273, "encoder_q-layer.9": 29938.166, "epoch": 0.64, "inbatch_neg_score": 4.4114, "inbatch_pos_score": 5.0195, "learning_rate": 1.3657894736842106e-05, "loss": 3.3711, "norm_diff": 0.0724, "norm_loss": 0.0, "num_token_doc": 66.7868, "num_token_overlap": 11.6409, "num_token_query": 31.3195, "num_token_union": 65.1391, "num_word_context": 202.3971, "num_word_doc": 49.8481, "num_word_query": 23.2606, "postclip_grad_norm": 1.0, "preclip_grad_norm": 44210.2546, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4141, "query_norm": 2.4278, "queue_k_norm": 2.4983, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3195, "sent_len_1": 66.7868, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.9475, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 148100 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3581, "doc_norm": 2.4995, "encoder_q-embeddings": 20345.8203, "encoder_q-layer.0": 13941.791, "encoder_q-layer.1": 15144.3594, "encoder_q-layer.10": 27215.5273, "encoder_q-layer.11": 67260.5312, "encoder_q-layer.2": 16788.3633, "encoder_q-layer.3": 17636.7441, "encoder_q-layer.4": 18593.5176, "encoder_q-layer.5": 19145.5703, "encoder_q-layer.6": 21232.4082, "encoder_q-layer.7": 23719.4336, "encoder_q-layer.8": 29527.4746, "encoder_q-layer.9": 25822.6094, "epoch": 0.64, "inbatch_neg_score": 4.4147, "inbatch_pos_score": 5.0312, "learning_rate": 1.3631578947368423e-05, "loss": 3.3581, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.7095, "num_token_overlap": 11.6544, "num_token_query": 31.3084, "num_token_union": 65.0601, "num_word_context": 202.1211, "num_word_doc": 49.8134, "num_word_query": 23.2454, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42460.1389, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4141, "query_norm": 2.4339, "queue_k_norm": 2.4998, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3084, "sent_len_1": 66.7095, "sent_len_max_0": 127.985, "sent_len_max_1": 208.0412, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148200 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3551, "doc_norm": 2.4998, "encoder_q-embeddings": 21316.2734, "encoder_q-layer.0": 14728.001, "encoder_q-layer.1": 15869.2852, "encoder_q-layer.10": 24448.084, "encoder_q-layer.11": 62885.4453, "encoder_q-layer.2": 18229.9941, "encoder_q-layer.3": 18021.957, "encoder_q-layer.4": 19124.7812, "encoder_q-layer.5": 19363.873, "encoder_q-layer.6": 20844.9297, "encoder_q-layer.7": 22565.6836, "encoder_q-layer.8": 26220.873, "encoder_q-layer.9": 24162.2285, "epoch": 0.64, "inbatch_neg_score": 4.4194, "inbatch_pos_score": 5.0234, "learning_rate": 1.3605263157894738e-05, "loss": 3.3551, "norm_diff": 0.0636, "norm_loss": 0.0, "num_token_doc": 66.7657, "num_token_overlap": 11.6431, "num_token_query": 31.3461, "num_token_union": 65.1537, "num_word_context": 202.2824, "num_word_doc": 49.8352, "num_word_query": 23.2754, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41147.7754, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.418, "query_norm": 2.4361, "queue_k_norm": 2.5017, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3461, "sent_len_1": 66.7657, "sent_len_max_0": 127.9425, "sent_len_max_1": 207.5962, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148300 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3631, "doc_norm": 2.5044, "encoder_q-embeddings": 10662.4678, "encoder_q-layer.0": 8007.0713, "encoder_q-layer.1": 8600.2139, "encoder_q-layer.10": 13431.7803, "encoder_q-layer.11": 33263.4688, "encoder_q-layer.2": 9427.8525, "encoder_q-layer.3": 9468.0322, "encoder_q-layer.4": 9853.4443, "encoder_q-layer.5": 9556.3711, "encoder_q-layer.6": 10331.1543, "encoder_q-layer.7": 11483.042, "encoder_q-layer.8": 14076.0215, "encoder_q-layer.9": 12620.0078, "epoch": 0.64, "inbatch_neg_score": 4.4223, "inbatch_pos_score": 5.0273, "learning_rate": 1.3578947368421053e-05, "loss": 3.3631, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.937, "num_token_overlap": 11.7023, "num_token_query": 31.4432, "num_token_union": 65.222, "num_word_context": 202.3987, "num_word_doc": 49.9049, "num_word_query": 23.3577, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20993.5896, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4219, "query_norm": 2.4289, "queue_k_norm": 2.5022, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4432, "sent_len_1": 66.937, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.6562, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148400 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.368, "doc_norm": 2.5034, "encoder_q-embeddings": 10934.915, "encoder_q-layer.0": 7888.0015, "encoder_q-layer.1": 8424.0117, "encoder_q-layer.10": 12901.8838, "encoder_q-layer.11": 31753.2539, "encoder_q-layer.2": 9787.8975, "encoder_q-layer.3": 10118.0381, "encoder_q-layer.4": 10257.6748, "encoder_q-layer.5": 11083.8262, "encoder_q-layer.6": 11314.1641, "encoder_q-layer.7": 12712.998, "encoder_q-layer.8": 14034.6865, "encoder_q-layer.9": 12294.7246, "epoch": 0.64, "inbatch_neg_score": 4.4234, "inbatch_pos_score": 5.0391, "learning_rate": 1.355263157894737e-05, "loss": 3.368, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.7744, "num_token_overlap": 11.6654, "num_token_query": 31.3397, "num_token_union": 65.088, "num_word_context": 202.2184, "num_word_doc": 49.7638, "num_word_query": 23.2613, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21561.4216, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4258, "query_norm": 2.4352, "queue_k_norm": 2.504, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3397, "sent_len_1": 66.7744, "sent_len_max_0": 127.9688, "sent_len_max_1": 212.2075, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148500 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3649, "doc_norm": 2.5061, "encoder_q-embeddings": 9905.1113, "encoder_q-layer.0": 6884.0327, "encoder_q-layer.1": 7454.644, "encoder_q-layer.10": 14732.1709, "encoder_q-layer.11": 32672.2246, "encoder_q-layer.2": 8307.7168, "encoder_q-layer.3": 8455.2949, "encoder_q-layer.4": 9036.0303, "encoder_q-layer.5": 9426.917, "encoder_q-layer.6": 9798.8438, "encoder_q-layer.7": 11626.9844, "encoder_q-layer.8": 13475.623, "encoder_q-layer.9": 12478.0732, "epoch": 0.64, "inbatch_neg_score": 4.4231, "inbatch_pos_score": 5.0352, "learning_rate": 1.3526315789473685e-05, "loss": 3.3649, "norm_diff": 0.075, "norm_loss": 0.0, "num_token_doc": 66.7092, "num_token_overlap": 11.6966, "num_token_query": 31.4574, "num_token_union": 65.1314, "num_word_context": 202.3751, "num_word_doc": 49.7768, "num_word_query": 23.3778, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20515.9654, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4258, "query_norm": 2.4311, "queue_k_norm": 2.5035, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4574, "sent_len_1": 66.7092, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.3325, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148600 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3512, "doc_norm": 2.504, "encoder_q-embeddings": 9777.8857, "encoder_q-layer.0": 6735.855, "encoder_q-layer.1": 7249.0957, "encoder_q-layer.10": 14041.0615, "encoder_q-layer.11": 32254.3223, "encoder_q-layer.2": 8125.4526, "encoder_q-layer.3": 8082.1367, "encoder_q-layer.4": 8676.0234, "encoder_q-layer.5": 8684.9531, "encoder_q-layer.6": 9740.8105, "encoder_q-layer.7": 11190.3535, "encoder_q-layer.8": 13601.6328, "encoder_q-layer.9": 12480.457, "epoch": 0.65, "inbatch_neg_score": 4.4299, "inbatch_pos_score": 5.0469, "learning_rate": 1.3500000000000001e-05, "loss": 3.3512, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.8757, "num_token_overlap": 11.6967, "num_token_query": 31.357, "num_token_union": 65.1705, "num_word_context": 202.2598, "num_word_doc": 49.9238, "num_word_query": 23.2811, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20362.7739, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4297, "query_norm": 2.4345, "queue_k_norm": 2.5056, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.357, "sent_len_1": 66.8757, "sent_len_max_0": 127.9562, "sent_len_max_1": 209.1587, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148700 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3729, "doc_norm": 2.5066, "encoder_q-embeddings": 10041.5029, "encoder_q-layer.0": 6991.9092, "encoder_q-layer.1": 7644.8228, "encoder_q-layer.10": 13194.5352, "encoder_q-layer.11": 32104.4453, "encoder_q-layer.2": 8692.1807, "encoder_q-layer.3": 8996.1406, "encoder_q-layer.4": 9708.8418, "encoder_q-layer.5": 9878.9043, "encoder_q-layer.6": 11188.7549, "encoder_q-layer.7": 11843.6865, "encoder_q-layer.8": 13842.3408, "encoder_q-layer.9": 12495.0771, "epoch": 0.65, "inbatch_neg_score": 4.4268, "inbatch_pos_score": 5.0391, "learning_rate": 1.3473684210526316e-05, "loss": 3.3729, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.7342, "num_token_overlap": 11.6281, "num_token_query": 31.2284, "num_token_union": 65.0629, "num_word_context": 202.0339, "num_word_doc": 49.8155, "num_word_query": 23.1729, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20879.2102, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4258, "query_norm": 2.4286, "queue_k_norm": 2.5063, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2284, "sent_len_1": 66.7342, "sent_len_max_0": 127.96, "sent_len_max_1": 207.23, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148800 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3792, "doc_norm": 2.508, "encoder_q-embeddings": 10064.9082, "encoder_q-layer.0": 6911.4878, "encoder_q-layer.1": 7489.1621, "encoder_q-layer.10": 13820.8887, "encoder_q-layer.11": 32558.8066, "encoder_q-layer.2": 8727.1299, "encoder_q-layer.3": 9209.4043, "encoder_q-layer.4": 9951.248, "encoder_q-layer.5": 10496.168, "encoder_q-layer.6": 11278.627, "encoder_q-layer.7": 12764.9746, "encoder_q-layer.8": 14430.4199, "encoder_q-layer.9": 12519.6777, "epoch": 0.65, "inbatch_neg_score": 4.4281, "inbatch_pos_score": 5.0273, "learning_rate": 1.3447368421052633e-05, "loss": 3.3792, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.9434, "num_token_overlap": 11.6497, "num_token_query": 31.3008, "num_token_union": 65.21, "num_word_context": 202.4486, "num_word_doc": 49.9518, "num_word_query": 23.2373, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21527.8959, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4297, "query_norm": 2.425, "queue_k_norm": 2.5068, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3008, "sent_len_1": 66.9434, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.3663, "stdk": 0.0496, "stdq": 0.0426, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 148900 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3705, "doc_norm": 2.5074, "encoder_q-embeddings": 9695.9766, "encoder_q-layer.0": 6858.2964, "encoder_q-layer.1": 7312.1226, "encoder_q-layer.10": 15628.3311, "encoder_q-layer.11": 34364.7617, "encoder_q-layer.2": 8189.978, "encoder_q-layer.3": 8520.9727, "encoder_q-layer.4": 9349.2549, "encoder_q-layer.5": 9406.0312, "encoder_q-layer.6": 10657.4062, "encoder_q-layer.7": 12227.2402, "encoder_q-layer.8": 15584.459, "encoder_q-layer.9": 14776.9346, "epoch": 0.65, "inbatch_neg_score": 4.4299, "inbatch_pos_score": 5.0391, "learning_rate": 1.3421052631578948e-05, "loss": 3.3705, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.5825, "num_token_overlap": 11.6643, "num_token_query": 31.316, "num_token_union": 64.9777, "num_word_context": 202.0029, "num_word_doc": 49.6706, "num_word_query": 23.2568, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21476.2925, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4297, "query_norm": 2.428, "queue_k_norm": 2.5065, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.316, "sent_len_1": 66.5825, "sent_len_max_0": 127.9663, "sent_len_max_1": 211.2337, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149000 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3824, "doc_norm": 2.5057, "encoder_q-embeddings": 9764.6865, "encoder_q-layer.0": 7021.0967, "encoder_q-layer.1": 7379.6499, "encoder_q-layer.10": 12754.2734, "encoder_q-layer.11": 31546.416, "encoder_q-layer.2": 8181.3687, "encoder_q-layer.3": 8682.918, "encoder_q-layer.4": 9177.4297, "encoder_q-layer.5": 9495.4023, "encoder_q-layer.6": 10220.9639, "encoder_q-layer.7": 11429.2227, "encoder_q-layer.8": 13503.8867, "encoder_q-layer.9": 12251.1299, "epoch": 0.65, "inbatch_neg_score": 4.4328, "inbatch_pos_score": 5.0391, "learning_rate": 1.3394736842105263e-05, "loss": 3.3824, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.7554, "num_token_overlap": 11.6378, "num_token_query": 31.2931, "num_token_union": 65.0573, "num_word_context": 202.3635, "num_word_doc": 49.765, "num_word_query": 23.2189, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20194.504, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4336, "query_norm": 2.4299, "queue_k_norm": 2.5076, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2931, "sent_len_1": 66.7554, "sent_len_max_0": 127.93, "sent_len_max_1": 212.09, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149100 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.355, "doc_norm": 2.5072, "encoder_q-embeddings": 10527.1514, "encoder_q-layer.0": 7598.2134, "encoder_q-layer.1": 8092.6597, "encoder_q-layer.10": 12846.0479, "encoder_q-layer.11": 30939.9395, "encoder_q-layer.2": 9321.1816, "encoder_q-layer.3": 9547.7979, "encoder_q-layer.4": 10506.4775, "encoder_q-layer.5": 11084.543, "encoder_q-layer.6": 11276.665, "encoder_q-layer.7": 12540.8994, "encoder_q-layer.8": 13294.749, "encoder_q-layer.9": 11907.1016, "epoch": 0.65, "inbatch_neg_score": 4.435, "inbatch_pos_score": 5.0508, "learning_rate": 1.336842105263158e-05, "loss": 3.355, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.8428, "num_token_overlap": 11.7559, "num_token_query": 31.5397, "num_token_union": 65.162, "num_word_context": 202.2766, "num_word_doc": 49.8495, "num_word_query": 23.4336, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20705.615, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4336, "query_norm": 2.4333, "queue_k_norm": 2.5084, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5397, "sent_len_1": 66.8428, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.0875, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149200 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3598, "doc_norm": 2.5107, "encoder_q-embeddings": 9886.7529, "encoder_q-layer.0": 6955.5986, "encoder_q-layer.1": 7224.9355, "encoder_q-layer.10": 12811.6553, "encoder_q-layer.11": 32462.9141, "encoder_q-layer.2": 8245.1221, "encoder_q-layer.3": 8612.2188, "encoder_q-layer.4": 9439.1895, "encoder_q-layer.5": 9706.6729, "encoder_q-layer.6": 11067.7881, "encoder_q-layer.7": 12830.999, "encoder_q-layer.8": 13493.4883, "encoder_q-layer.9": 12221.083, "epoch": 0.65, "inbatch_neg_score": 4.438, "inbatch_pos_score": 5.0547, "learning_rate": 1.3342105263157894e-05, "loss": 3.3598, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.7439, "num_token_overlap": 11.6656, "num_token_query": 31.3477, "num_token_union": 65.1135, "num_word_context": 202.2493, "num_word_doc": 49.8049, "num_word_query": 23.2838, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20856.1396, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4375, "query_norm": 2.4322, "queue_k_norm": 2.5084, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3477, "sent_len_1": 66.7439, "sent_len_max_0": 127.975, "sent_len_max_1": 209.42, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149300 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3688, "doc_norm": 2.5079, "encoder_q-embeddings": 18288.582, "encoder_q-layer.0": 13466.4268, "encoder_q-layer.1": 16030.9912, "encoder_q-layer.10": 14097.1377, "encoder_q-layer.11": 34092.6367, "encoder_q-layer.2": 17611.6211, "encoder_q-layer.3": 19400.0059, "encoder_q-layer.4": 21523.6758, "encoder_q-layer.5": 24545.7949, "encoder_q-layer.6": 24371.9512, "encoder_q-layer.7": 23370.6641, "encoder_q-layer.8": 21050.8418, "encoder_q-layer.9": 13774.2373, "epoch": 0.65, "inbatch_neg_score": 4.4392, "inbatch_pos_score": 5.0508, "learning_rate": 1.3315789473684213e-05, "loss": 3.3688, "norm_diff": 0.0815, "norm_loss": 0.0, "num_token_doc": 66.935, "num_token_overlap": 11.6415, "num_token_query": 31.2599, "num_token_union": 65.1933, "num_word_context": 202.5352, "num_word_doc": 49.9539, "num_word_query": 23.2061, "postclip_grad_norm": 1.0, "preclip_grad_norm": 31575.5023, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 4.4375, "query_norm": 2.4264, "queue_k_norm": 2.5109, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2599, "sent_len_1": 66.935, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.28, "stdk": 0.0493, "stdq": 0.0426, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 149400 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.3779, "doc_norm": 2.5077, "encoder_q-embeddings": 9982.8984, "encoder_q-layer.0": 7058.9277, "encoder_q-layer.1": 7351.2192, "encoder_q-layer.10": 15320.9492, "encoder_q-layer.11": 35121.2031, "encoder_q-layer.2": 8223.9902, "encoder_q-layer.3": 8534.4209, "encoder_q-layer.4": 9040.3828, "encoder_q-layer.5": 9438.2002, "encoder_q-layer.6": 10625.457, "encoder_q-layer.7": 12565.7139, "encoder_q-layer.8": 14732.5273, "encoder_q-layer.9": 13934.8174, "epoch": 0.65, "inbatch_neg_score": 4.4403, "inbatch_pos_score": 5.043, "learning_rate": 1.3289473684210526e-05, "loss": 3.3779, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.6053, "num_token_overlap": 11.643, "num_token_query": 31.3735, "num_token_union": 65.054, "num_word_context": 202.3819, "num_word_doc": 49.6948, "num_word_query": 23.3141, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21824.8087, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4414, "query_norm": 2.4355, "queue_k_norm": 2.5109, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3735, "sent_len_1": 66.6053, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.6037, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 149500 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3599, "doc_norm": 2.5071, "encoder_q-embeddings": 10194.3379, "encoder_q-layer.0": 7174.7754, "encoder_q-layer.1": 7664.0063, "encoder_q-layer.10": 13030.1396, "encoder_q-layer.11": 33389.625, "encoder_q-layer.2": 8832.9199, "encoder_q-layer.3": 8967.748, "encoder_q-layer.4": 9807.2285, "encoder_q-layer.5": 10143.8828, "encoder_q-layer.6": 10999.5215, "encoder_q-layer.7": 12662.3203, "encoder_q-layer.8": 13966.8926, "encoder_q-layer.9": 12500.9219, "epoch": 0.65, "inbatch_neg_score": 4.4413, "inbatch_pos_score": 5.0391, "learning_rate": 1.3263157894736844e-05, "loss": 3.3599, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.6019, "num_token_overlap": 11.6822, "num_token_query": 31.3637, "num_token_union": 64.9943, "num_word_context": 201.8912, "num_word_doc": 49.6811, "num_word_query": 23.2851, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21073.4069, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4414, "query_norm": 2.4257, "queue_k_norm": 2.5098, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3637, "sent_len_1": 66.6019, "sent_len_max_0": 127.9737, "sent_len_max_1": 206.6687, "stdk": 0.0492, "stdq": 0.0425, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149600 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3568, "doc_norm": 2.5083, "encoder_q-embeddings": 9488.749, "encoder_q-layer.0": 6558.5962, "encoder_q-layer.1": 7126.605, "encoder_q-layer.10": 12856.8867, "encoder_q-layer.11": 31397.9141, "encoder_q-layer.2": 7640.4561, "encoder_q-layer.3": 7792.75, "encoder_q-layer.4": 8478.4678, "encoder_q-layer.5": 8496.0166, "encoder_q-layer.6": 9588.4736, "encoder_q-layer.7": 10776.8311, "encoder_q-layer.8": 13648.1641, "encoder_q-layer.9": 12294.3789, "epoch": 0.65, "inbatch_neg_score": 4.4457, "inbatch_pos_score": 5.0625, "learning_rate": 1.323684210526316e-05, "loss": 3.3568, "norm_diff": 0.0689, "norm_loss": 0.0, "num_token_doc": 66.6669, "num_token_overlap": 11.6596, "num_token_query": 31.3646, "num_token_union": 65.047, "num_word_context": 202.1146, "num_word_doc": 49.7307, "num_word_query": 23.2862, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19649.1776, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4453, "query_norm": 2.4394, "queue_k_norm": 2.5105, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3646, "sent_len_1": 66.6669, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.525, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149700 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.3773, "doc_norm": 2.5101, "encoder_q-embeddings": 10025.4688, "encoder_q-layer.0": 6908.3564, "encoder_q-layer.1": 7234.7734, "encoder_q-layer.10": 14453.0107, "encoder_q-layer.11": 32738.2715, "encoder_q-layer.2": 8039.3896, "encoder_q-layer.3": 8200.0977, "encoder_q-layer.4": 9042.8721, "encoder_q-layer.5": 9321.6572, "encoder_q-layer.6": 10379.6309, "encoder_q-layer.7": 11804.3984, "encoder_q-layer.8": 14708.0449, "encoder_q-layer.9": 13776.2031, "epoch": 0.65, "inbatch_neg_score": 4.4503, "inbatch_pos_score": 5.0508, "learning_rate": 1.3210526315789473e-05, "loss": 3.3773, "norm_diff": 0.0786, "norm_loss": 0.0, "num_token_doc": 66.9167, "num_token_overlap": 11.6432, "num_token_query": 31.3038, "num_token_union": 65.1555, "num_word_context": 202.4563, "num_word_doc": 49.9457, "num_word_query": 23.2408, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20585.0845, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4492, "query_norm": 2.4315, "queue_k_norm": 2.5109, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3038, "sent_len_1": 66.9167, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.5725, "stdk": 0.0493, "stdq": 0.0427, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149800 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.3658, "doc_norm": 2.5133, "encoder_q-embeddings": 9426.4824, "encoder_q-layer.0": 6561.1343, "encoder_q-layer.1": 6988.9556, "encoder_q-layer.10": 12873.3789, "encoder_q-layer.11": 32327.9141, "encoder_q-layer.2": 7830.6924, "encoder_q-layer.3": 7735.313, "encoder_q-layer.4": 8284.4482, "encoder_q-layer.5": 8585.8799, "encoder_q-layer.6": 9900.085, "encoder_q-layer.7": 11718.3281, "encoder_q-layer.8": 14185.0898, "encoder_q-layer.9": 13115.9404, "epoch": 0.65, "inbatch_neg_score": 4.4537, "inbatch_pos_score": 5.0664, "learning_rate": 1.3184210526315791e-05, "loss": 3.3658, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.7403, "num_token_overlap": 11.6811, "num_token_query": 31.3484, "num_token_union": 65.0965, "num_word_context": 202.1485, "num_word_doc": 49.7914, "num_word_query": 23.289, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20557.7403, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4531, "query_norm": 2.4381, "queue_k_norm": 2.5125, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3484, "sent_len_1": 66.7403, "sent_len_max_0": 127.98, "sent_len_max_1": 207.7575, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 149900 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3688, "doc_norm": 2.5158, "encoder_q-embeddings": 9738.873, "encoder_q-layer.0": 6788.7422, "encoder_q-layer.1": 7161.2183, "encoder_q-layer.10": 14551.2852, "encoder_q-layer.11": 33294.0195, "encoder_q-layer.2": 8045.4385, "encoder_q-layer.3": 8071.6641, "encoder_q-layer.4": 8908.1895, "encoder_q-layer.5": 9469.4209, "encoder_q-layer.6": 10407.252, "encoder_q-layer.7": 12810.3867, "encoder_q-layer.8": 15900.3027, "encoder_q-layer.9": 13742.4199, "epoch": 0.65, "inbatch_neg_score": 4.4576, "inbatch_pos_score": 5.0586, "learning_rate": 1.3157894736842106e-05, "loss": 3.3688, "norm_diff": 0.0738, "norm_loss": 0.0, "num_token_doc": 66.7546, "num_token_overlap": 11.6459, "num_token_query": 31.2979, "num_token_union": 65.0964, "num_word_context": 202.1874, "num_word_doc": 49.7863, "num_word_query": 23.2389, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21264.3065, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.457, "query_norm": 2.442, "queue_k_norm": 2.5135, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2979, "sent_len_1": 66.7546, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.9387, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150000 }, { "dev_runtime": 29.419, "dev_samples_per_second": 1.088, "dev_steps_per_second": 0.034, "epoch": 0.65, "step": 150000, "test_accuracy": 94.49462890625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.31351497769355774, "test_doc_norm": 2.500676155090332, "test_inbatch_neg_score": 4.75770378112793, "test_inbatch_pos_score": 5.7862548828125, "test_loss": 0.31351497769355774, "test_loss_align": 1.0287288427352905, "test_loss_unif": -34.364994049072266, "test_loss_unif_q@queue": -34.364994049072266, "test_norm_diff": 0.016643665730953217, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.458407402038574, "test_query_norm": 2.517320156097412, "test_queue_k_norm": 2.513869285583496, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04433240368962288, "test_stdq": 0.044126659631729126, "test_stdqueue_k": 0.04957287758588791, "test_stdqueue_q": 0.0 }, { "dev_runtime": 29.419, "dev_samples_per_second": 1.088, "dev_steps_per_second": 0.034, "epoch": 0.65, "eval_beir-arguana_ndcg@10": 0.40541, "eval_beir-arguana_recall@10": 0.68279, "eval_beir-arguana_recall@100": 0.94239, "eval_beir-arguana_recall@20": 0.81437, "eval_beir-avg_ndcg@10": 0.4026885, "eval_beir-avg_recall@10": 0.4725350833333333, "eval_beir-avg_recall@100": 0.6524301666666666, "eval_beir-avg_recall@20": 0.5344859166666667, "eval_beir-cqadupstack_ndcg@10": 0.29470500000000005, "eval_beir-cqadupstack_recall@10": 0.39654083333333334, "eval_beir-cqadupstack_recall@100": 0.6301516666666667, "eval_beir-cqadupstack_recall@20": 0.4645291666666666, "eval_beir-fiqa_ndcg@10": 0.26199, "eval_beir-fiqa_recall@10": 0.32875, "eval_beir-fiqa_recall@100": 0.60765, "eval_beir-fiqa_recall@20": 0.41591, "eval_beir-nfcorpus_ndcg@10": 0.32445, "eval_beir-nfcorpus_recall@10": 0.16032, "eval_beir-nfcorpus_recall@100": 0.31264, "eval_beir-nfcorpus_recall@20": 0.19522, "eval_beir-nq_ndcg@10": 0.32772, "eval_beir-nq_recall@10": 0.51922, "eval_beir-nq_recall@100": 0.84432, "eval_beir-nq_recall@20": 0.634, "eval_beir-quora_ndcg@10": 0.79409, "eval_beir-quora_recall@10": 0.89928, "eval_beir-quora_recall@100": 0.9814, "eval_beir-quora_recall@20": 0.93757, "eval_beir-scidocs_ndcg@10": 0.16627, "eval_beir-scidocs_recall@10": 0.17463, "eval_beir-scidocs_recall@100": 0.3879, "eval_beir-scidocs_recall@20": 0.23512, "eval_beir-scifact_ndcg@10": 0.67783, "eval_beir-scifact_recall@10": 0.79967, "eval_beir-scifact_recall@100": 0.92989, "eval_beir-scifact_recall@20": 0.84811, "eval_beir-trec-covid_ndcg@10": 0.60154, "eval_beir-trec-covid_recall@10": 0.642, "eval_beir-trec-covid_recall@100": 0.4576, "eval_beir-trec-covid_recall@20": 0.592, "eval_beir-webis-touche2020_ndcg@10": 0.17288, "eval_beir-webis-touche2020_recall@10": 0.12215, "eval_beir-webis-touche2020_recall@100": 0.43036, "eval_beir-webis-touche2020_recall@20": 0.20803, "eval_senteval-avg_sts": 0.7297933411047239, "eval_senteval-sickr_spearman": 0.7253875480695317, "eval_senteval-stsb_spearman": 0.7341991341399161, "step": 150000, "test_accuracy": 94.49462890625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.31351497769355774, "test_doc_norm": 2.500676155090332, "test_inbatch_neg_score": 4.75770378112793, "test_inbatch_pos_score": 5.7862548828125, "test_loss": 0.31351497769355774, "test_loss_align": 1.0287288427352905, "test_loss_unif": -34.364994049072266, "test_loss_unif_q@queue": -34.364994049072266, "test_norm_diff": 0.016643665730953217, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.458407402038574, "test_query_norm": 2.517320156097412, "test_queue_k_norm": 2.513869285583496, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04433240368962288, "test_stdq": 0.044126659631729126, "test_stdqueue_k": 0.04957287758588791, "test_stdqueue_q": 0.0 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.3559, "doc_norm": 2.5147, "encoder_q-embeddings": 9273.1807, "encoder_q-layer.0": 6677.9121, "encoder_q-layer.1": 6872.2002, "encoder_q-layer.10": 13671.6602, "encoder_q-layer.11": 32713.4316, "encoder_q-layer.2": 7485.8774, "encoder_q-layer.3": 7672.5288, "encoder_q-layer.4": 7986.5459, "encoder_q-layer.5": 8426.7256, "encoder_q-layer.6": 9645.7246, "encoder_q-layer.7": 11005.0156, "encoder_q-layer.8": 13453.1982, "encoder_q-layer.9": 12937.2617, "epoch": 0.65, "inbatch_neg_score": 4.4643, "inbatch_pos_score": 5.0781, "learning_rate": 1.3131578947368423e-05, "loss": 3.3559, "norm_diff": 0.0709, "norm_loss": 0.0, "num_token_doc": 66.8279, "num_token_overlap": 11.6453, "num_token_query": 31.4166, "num_token_union": 65.2146, "num_word_context": 202.6084, "num_word_doc": 49.8849, "num_word_query": 23.3429, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20049.2647, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4648, "query_norm": 2.4438, "queue_k_norm": 2.5142, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4166, "sent_len_1": 66.8279, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.1525, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 150100 }, { "accuracy": 52.832, "active_queue_size": 16384.0, "cl_loss": 3.3558, "doc_norm": 2.5143, "encoder_q-embeddings": 11612.8164, "encoder_q-layer.0": 8247.7119, "encoder_q-layer.1": 9538.0703, "encoder_q-layer.10": 12809.0566, "encoder_q-layer.11": 31816.2344, "encoder_q-layer.2": 11747.0312, "encoder_q-layer.3": 11573.6064, "encoder_q-layer.4": 11729.6416, "encoder_q-layer.5": 10596.0615, "encoder_q-layer.6": 11193.2412, "encoder_q-layer.7": 11593.6064, "encoder_q-layer.8": 13750.4297, "encoder_q-layer.9": 12427.3857, "epoch": 0.65, "inbatch_neg_score": 4.4694, "inbatch_pos_score": 5.0938, "learning_rate": 1.3105263157894738e-05, "loss": 3.3558, "norm_diff": 0.0694, "norm_loss": 0.0, "num_token_doc": 66.7519, "num_token_overlap": 11.6737, "num_token_query": 31.3542, "num_token_union": 65.0793, "num_word_context": 202.1509, "num_word_doc": 49.805, "num_word_query": 23.2758, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21615.6022, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4688, "query_norm": 2.445, "queue_k_norm": 2.5136, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3542, "sent_len_1": 66.7519, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.3575, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150200 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.355, "doc_norm": 2.5121, "encoder_q-embeddings": 11109.1465, "encoder_q-layer.0": 7789.2188, "encoder_q-layer.1": 8331.6396, "encoder_q-layer.10": 14612.0361, "encoder_q-layer.11": 32350.4668, "encoder_q-layer.2": 9689.1289, "encoder_q-layer.3": 9940.3154, "encoder_q-layer.4": 10505.7744, "encoder_q-layer.5": 10458.4287, "encoder_q-layer.6": 10888.6201, "encoder_q-layer.7": 12193.8994, "encoder_q-layer.8": 13390.5967, "encoder_q-layer.9": 12571.1426, "epoch": 0.65, "inbatch_neg_score": 4.4739, "inbatch_pos_score": 5.0938, "learning_rate": 1.3078947368421054e-05, "loss": 3.355, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.8436, "num_token_overlap": 11.6962, "num_token_query": 31.3702, "num_token_union": 65.1287, "num_word_context": 201.9443, "num_word_doc": 49.8725, "num_word_query": 23.2823, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21334.776, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.4727, "query_norm": 2.4463, "queue_k_norm": 2.5154, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3702, "sent_len_1": 66.8436, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.9288, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150300 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.368, "doc_norm": 2.5154, "encoder_q-embeddings": 21220.8281, "encoder_q-layer.0": 15365.6777, "encoder_q-layer.1": 16739.2207, "encoder_q-layer.10": 27753.0059, "encoder_q-layer.11": 69818.5, "encoder_q-layer.2": 19254.7109, "encoder_q-layer.3": 20594.0703, "encoder_q-layer.4": 23035.543, "encoder_q-layer.5": 23579.9414, "encoder_q-layer.6": 28663.3535, "encoder_q-layer.7": 28962.0371, "encoder_q-layer.8": 33242.1211, "encoder_q-layer.9": 27709.8887, "epoch": 0.65, "inbatch_neg_score": 4.4832, "inbatch_pos_score": 5.1016, "learning_rate": 1.305263157894737e-05, "loss": 3.368, "norm_diff": 0.0596, "norm_loss": 0.0, "num_token_doc": 66.8323, "num_token_overlap": 11.6689, "num_token_query": 31.3182, "num_token_union": 65.1324, "num_word_context": 202.5132, "num_word_doc": 49.8444, "num_word_query": 23.2726, "postclip_grad_norm": 1.0, "preclip_grad_norm": 46512.39, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4844, "query_norm": 2.4558, "queue_k_norm": 2.5163, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3182, "sent_len_1": 66.8323, "sent_len_max_0": 127.99, "sent_len_max_1": 209.6087, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 150400 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.376, "doc_norm": 2.516, "encoder_q-embeddings": 22417.5996, "encoder_q-layer.0": 16646.8242, "encoder_q-layer.1": 17407.3496, "encoder_q-layer.10": 28477.8047, "encoder_q-layer.11": 65876.4062, "encoder_q-layer.2": 19866.7637, "encoder_q-layer.3": 19623.8008, "encoder_q-layer.4": 20933.1484, "encoder_q-layer.5": 20413.043, "encoder_q-layer.6": 21778.1621, "encoder_q-layer.7": 24776.6328, "encoder_q-layer.8": 28594.1562, "encoder_q-layer.9": 25630.3223, "epoch": 0.65, "inbatch_neg_score": 4.485, "inbatch_pos_score": 5.0898, "learning_rate": 1.3026315789473684e-05, "loss": 3.376, "norm_diff": 0.0672, "norm_loss": 0.0, "num_token_doc": 66.8363, "num_token_overlap": 11.6841, "num_token_query": 31.4514, "num_token_union": 65.1845, "num_word_context": 202.4969, "num_word_doc": 49.8802, "num_word_query": 23.3549, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43372.897, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4844, "query_norm": 2.4488, "queue_k_norm": 2.5161, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4514, "sent_len_1": 66.8363, "sent_len_max_0": 127.98, "sent_len_max_1": 208.3363, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150500 }, { "accuracy": 50.0, "active_queue_size": 16384.0, "cl_loss": 3.3685, "doc_norm": 2.5178, "encoder_q-embeddings": 21010.8145, "encoder_q-layer.0": 14543.377, "encoder_q-layer.1": 15410.2227, "encoder_q-layer.10": 25761.5859, "encoder_q-layer.11": 64835.6016, "encoder_q-layer.2": 16814.8477, "encoder_q-layer.3": 17028.4277, "encoder_q-layer.4": 18355.6797, "encoder_q-layer.5": 18626.9766, "encoder_q-layer.6": 21278.8008, "encoder_q-layer.7": 23140.2266, "encoder_q-layer.8": 25301.4414, "encoder_q-layer.9": 25117.9199, "epoch": 0.65, "inbatch_neg_score": 4.495, "inbatch_pos_score": 5.0977, "learning_rate": 1.3000000000000001e-05, "loss": 3.3685, "norm_diff": 0.0679, "norm_loss": 0.0, "num_token_doc": 66.7955, "num_token_overlap": 11.7073, "num_token_query": 31.4207, "num_token_union": 65.1535, "num_word_context": 202.2817, "num_word_doc": 49.8736, "num_word_query": 23.3385, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41426.9338, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.4961, "query_norm": 2.4499, "queue_k_norm": 2.5177, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4207, "sent_len_1": 66.7955, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.0513, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150600 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3834, "doc_norm": 2.5179, "encoder_q-embeddings": 19062.2422, "encoder_q-layer.0": 13119.9102, "encoder_q-layer.1": 14548.3896, "encoder_q-layer.10": 26252.4902, "encoder_q-layer.11": 66539.5469, "encoder_q-layer.2": 16733.0078, "encoder_q-layer.3": 17558.5781, "encoder_q-layer.4": 18759.0293, "encoder_q-layer.5": 20346.6016, "encoder_q-layer.6": 22758.7266, "encoder_q-layer.7": 26463.1113, "encoder_q-layer.8": 31541.8379, "encoder_q-layer.9": 26717.873, "epoch": 0.65, "inbatch_neg_score": 4.5028, "inbatch_pos_score": 5.1133, "learning_rate": 1.2973684210526316e-05, "loss": 3.3834, "norm_diff": 0.0655, "norm_loss": 0.0, "num_token_doc": 66.6101, "num_token_overlap": 11.6388, "num_token_query": 31.3021, "num_token_union": 65.0237, "num_word_context": 202.2144, "num_word_doc": 49.7143, "num_word_query": 23.2363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43276.0964, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.5, "query_norm": 2.4523, "queue_k_norm": 2.518, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3021, "sent_len_1": 66.6101, "sent_len_max_0": 127.985, "sent_len_max_1": 206.8487, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150700 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3592, "doc_norm": 2.5181, "encoder_q-embeddings": 9577.9111, "encoder_q-layer.0": 6804.3608, "encoder_q-layer.1": 7341.4229, "encoder_q-layer.10": 13988.2051, "encoder_q-layer.11": 34457.9062, "encoder_q-layer.2": 8328.5879, "encoder_q-layer.3": 8640.085, "encoder_q-layer.4": 9362.3018, "encoder_q-layer.5": 9695.5664, "encoder_q-layer.6": 11209.5713, "encoder_q-layer.7": 12121.0205, "encoder_q-layer.8": 14731.3428, "encoder_q-layer.9": 13234.1133, "epoch": 0.65, "inbatch_neg_score": 4.5132, "inbatch_pos_score": 5.1172, "learning_rate": 1.2947368421052633e-05, "loss": 3.3592, "norm_diff": 0.0595, "norm_loss": 0.0, "num_token_doc": 66.8239, "num_token_overlap": 11.6808, "num_token_query": 31.3509, "num_token_union": 65.1546, "num_word_context": 202.4449, "num_word_doc": 49.8905, "num_word_query": 23.2803, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21752.5281, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5117, "query_norm": 2.4586, "queue_k_norm": 2.5188, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3509, "sent_len_1": 66.8239, "sent_len_max_0": 128.0, "sent_len_max_1": 208.5625, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150800 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.5192, "encoder_q-embeddings": 9517.6045, "encoder_q-layer.0": 6686.1338, "encoder_q-layer.1": 7243.5771, "encoder_q-layer.10": 14427.3525, "encoder_q-layer.11": 33981.2344, "encoder_q-layer.2": 7985.4985, "encoder_q-layer.3": 8096.46, "encoder_q-layer.4": 8688.4385, "encoder_q-layer.5": 9172.4697, "encoder_q-layer.6": 10568.7656, "encoder_q-layer.7": 12139.2529, "encoder_q-layer.8": 14718.2676, "encoder_q-layer.9": 14043.873, "epoch": 0.65, "inbatch_neg_score": 4.5148, "inbatch_pos_score": 5.1328, "learning_rate": 1.2921052631578948e-05, "loss": 3.3508, "norm_diff": 0.0582, "norm_loss": 0.0, "num_token_doc": 66.796, "num_token_overlap": 11.6703, "num_token_query": 31.4229, "num_token_union": 65.18, "num_word_context": 202.4656, "num_word_doc": 49.8706, "num_word_query": 23.3577, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21433.8978, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5156, "query_norm": 2.4611, "queue_k_norm": 2.5206, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4229, "sent_len_1": 66.796, "sent_len_max_0": 127.9862, "sent_len_max_1": 205.8762, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 150900 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.356, "doc_norm": 2.5212, "encoder_q-embeddings": 9467.0, "encoder_q-layer.0": 6798.7271, "encoder_q-layer.1": 7474.4243, "encoder_q-layer.10": 13236.4463, "encoder_q-layer.11": 34111.8945, "encoder_q-layer.2": 8509.3262, "encoder_q-layer.3": 8399.3389, "encoder_q-layer.4": 9111.4014, "encoder_q-layer.5": 9301.4756, "encoder_q-layer.6": 10421.0635, "encoder_q-layer.7": 11234.25, "encoder_q-layer.8": 14051.6729, "encoder_q-layer.9": 12723.0088, "epoch": 0.66, "inbatch_neg_score": 4.5248, "inbatch_pos_score": 5.1484, "learning_rate": 1.2894736842105264e-05, "loss": 3.356, "norm_diff": 0.0531, "norm_loss": 0.0, "num_token_doc": 66.7269, "num_token_overlap": 11.6784, "num_token_query": 31.355, "num_token_union": 65.1034, "num_word_context": 202.2496, "num_word_doc": 49.8008, "num_word_query": 23.2869, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21125.5999, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5234, "query_norm": 2.4681, "queue_k_norm": 2.5198, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.355, "sent_len_1": 66.7269, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.5825, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 151000 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3705, "doc_norm": 2.5234, "encoder_q-embeddings": 9520.2412, "encoder_q-layer.0": 6677.3706, "encoder_q-layer.1": 7124.0977, "encoder_q-layer.10": 12478.4307, "encoder_q-layer.11": 31711.0039, "encoder_q-layer.2": 8030.6802, "encoder_q-layer.3": 8274.2139, "encoder_q-layer.4": 8663.3877, "encoder_q-layer.5": 8953.4072, "encoder_q-layer.6": 9920.874, "encoder_q-layer.7": 10862.2686, "encoder_q-layer.8": 12545.9775, "encoder_q-layer.9": 11994.4824, "epoch": 0.66, "inbatch_neg_score": 4.5336, "inbatch_pos_score": 5.1523, "learning_rate": 1.2868421052631579e-05, "loss": 3.3705, "norm_diff": 0.0502, "norm_loss": 0.0, "num_token_doc": 66.7318, "num_token_overlap": 11.6233, "num_token_query": 31.1986, "num_token_union": 65.0343, "num_word_context": 202.1642, "num_word_doc": 49.8026, "num_word_query": 23.1726, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19942.6116, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5312, "query_norm": 2.4732, "queue_k_norm": 2.5225, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.1986, "sent_len_1": 66.7318, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.425, "stdk": 0.0495, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151100 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3672, "doc_norm": 2.5243, "encoder_q-embeddings": 9398.7939, "encoder_q-layer.0": 6619.3628, "encoder_q-layer.1": 6904.501, "encoder_q-layer.10": 14580.585, "encoder_q-layer.11": 33382.4844, "encoder_q-layer.2": 7671.2148, "encoder_q-layer.3": 7641.8877, "encoder_q-layer.4": 8112.2671, "encoder_q-layer.5": 8184.896, "encoder_q-layer.6": 9299.3818, "encoder_q-layer.7": 10948.6221, "encoder_q-layer.8": 14349.0469, "encoder_q-layer.9": 12588.7275, "epoch": 0.66, "inbatch_neg_score": 4.5357, "inbatch_pos_score": 5.1562, "learning_rate": 1.2842105263157894e-05, "loss": 3.3672, "norm_diff": 0.0573, "norm_loss": 0.0, "num_token_doc": 66.7949, "num_token_overlap": 11.6662, "num_token_query": 31.2996, "num_token_union": 65.0856, "num_word_context": 202.2217, "num_word_doc": 49.8498, "num_word_query": 23.2352, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20365.3748, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5352, "query_norm": 2.467, "queue_k_norm": 2.5235, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2996, "sent_len_1": 66.7949, "sent_len_max_0": 127.995, "sent_len_max_1": 208.5563, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151200 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3615, "doc_norm": 2.5234, "encoder_q-embeddings": 9797.3057, "encoder_q-layer.0": 6879.4956, "encoder_q-layer.1": 7320.0122, "encoder_q-layer.10": 13161.3203, "encoder_q-layer.11": 32431.9277, "encoder_q-layer.2": 8252.4219, "encoder_q-layer.3": 8589.0059, "encoder_q-layer.4": 9265.7285, "encoder_q-layer.5": 9540.7002, "encoder_q-layer.6": 10267.8818, "encoder_q-layer.7": 11423.7686, "encoder_q-layer.8": 13680.5645, "encoder_q-layer.9": 12469.5957, "epoch": 0.66, "inbatch_neg_score": 4.5415, "inbatch_pos_score": 5.1562, "learning_rate": 1.281578947368421e-05, "loss": 3.3615, "norm_diff": 0.0536, "norm_loss": 0.0, "num_token_doc": 66.7076, "num_token_overlap": 11.6577, "num_token_query": 31.3108, "num_token_union": 65.0141, "num_word_context": 202.083, "num_word_doc": 49.7527, "num_word_query": 23.2491, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20764.7261, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.543, "query_norm": 2.4698, "queue_k_norm": 2.5246, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3108, "sent_len_1": 66.7076, "sent_len_max_0": 127.98, "sent_len_max_1": 209.79, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151300 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3598, "doc_norm": 2.5245, "encoder_q-embeddings": 9472.1523, "encoder_q-layer.0": 6563.1514, "encoder_q-layer.1": 6817.9956, "encoder_q-layer.10": 13124.5996, "encoder_q-layer.11": 32293.8262, "encoder_q-layer.2": 7558.3472, "encoder_q-layer.3": 7771.896, "encoder_q-layer.4": 8386.4746, "encoder_q-layer.5": 8648.9639, "encoder_q-layer.6": 9960.3047, "encoder_q-layer.7": 11377.2559, "encoder_q-layer.8": 13638.3867, "encoder_q-layer.9": 12478.7646, "epoch": 0.66, "inbatch_neg_score": 4.552, "inbatch_pos_score": 5.1719, "learning_rate": 1.2789473684210526e-05, "loss": 3.3598, "norm_diff": 0.0494, "norm_loss": 0.0, "num_token_doc": 66.9031, "num_token_overlap": 11.7089, "num_token_query": 31.4029, "num_token_union": 65.1975, "num_word_context": 202.3944, "num_word_doc": 49.9065, "num_word_query": 23.3332, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20293.7007, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5508, "query_norm": 2.4751, "queue_k_norm": 2.5262, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4029, "sent_len_1": 66.9031, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.3512, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 151400 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3597, "doc_norm": 2.5259, "encoder_q-embeddings": 9469.459, "encoder_q-layer.0": 6468.8228, "encoder_q-layer.1": 7015.8101, "encoder_q-layer.10": 14724.7207, "encoder_q-layer.11": 33143.4648, "encoder_q-layer.2": 7806.9209, "encoder_q-layer.3": 8206.3877, "encoder_q-layer.4": 8390.0752, "encoder_q-layer.5": 8617.9004, "encoder_q-layer.6": 9642.0029, "encoder_q-layer.7": 10972.7871, "encoder_q-layer.8": 13952.6631, "encoder_q-layer.9": 13342.3154, "epoch": 0.66, "inbatch_neg_score": 4.5588, "inbatch_pos_score": 5.168, "learning_rate": 1.2763157894736844e-05, "loss": 3.3597, "norm_diff": 0.0541, "norm_loss": 0.0, "num_token_doc": 66.9906, "num_token_overlap": 11.6882, "num_token_query": 31.4213, "num_token_union": 65.227, "num_word_context": 202.5376, "num_word_doc": 49.9648, "num_word_query": 23.3321, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20387.5823, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.5586, "query_norm": 2.4718, "queue_k_norm": 2.5273, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4213, "sent_len_1": 66.9906, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.4263, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 151500 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3583, "doc_norm": 2.529, "encoder_q-embeddings": 4860.374, "encoder_q-layer.0": 3399.3633, "encoder_q-layer.1": 3615.3555, "encoder_q-layer.10": 6482.123, "encoder_q-layer.11": 16380.6348, "encoder_q-layer.2": 4050.4045, "encoder_q-layer.3": 4329.5146, "encoder_q-layer.4": 4538.6904, "encoder_q-layer.5": 4622.9985, "encoder_q-layer.6": 5197.166, "encoder_q-layer.7": 5810.6636, "encoder_q-layer.8": 6928.5649, "encoder_q-layer.9": 6546.2354, "epoch": 0.66, "inbatch_neg_score": 4.5627, "inbatch_pos_score": 5.1758, "learning_rate": 1.2736842105263157e-05, "loss": 3.3583, "norm_diff": 0.0549, "norm_loss": 0.0, "num_token_doc": 66.7085, "num_token_overlap": 11.6841, "num_token_query": 31.3761, "num_token_union": 65.0515, "num_word_context": 202.0336, "num_word_doc": 49.7571, "num_word_query": 23.304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10522.4688, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5625, "query_norm": 2.4741, "queue_k_norm": 2.5276, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3761, "sent_len_1": 66.7085, "sent_len_max_0": 127.98, "sent_len_max_1": 209.86, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151600 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3636, "doc_norm": 2.5305, "encoder_q-embeddings": 4764.4355, "encoder_q-layer.0": 3444.5364, "encoder_q-layer.1": 3927.4202, "encoder_q-layer.10": 6363.5537, "encoder_q-layer.11": 15729.2578, "encoder_q-layer.2": 4333.3457, "encoder_q-layer.3": 4623.4048, "encoder_q-layer.4": 5033.5083, "encoder_q-layer.5": 4985.1221, "encoder_q-layer.6": 5517.856, "encoder_q-layer.7": 6408.0913, "encoder_q-layer.8": 6883.2368, "encoder_q-layer.9": 6258.4126, "epoch": 0.66, "inbatch_neg_score": 4.5692, "inbatch_pos_score": 5.1797, "learning_rate": 1.2710526315789476e-05, "loss": 3.3636, "norm_diff": 0.0578, "norm_loss": 0.0, "num_token_doc": 66.6728, "num_token_overlap": 11.6477, "num_token_query": 31.3078, "num_token_union": 65.0675, "num_word_context": 202.4597, "num_word_doc": 49.7774, "num_word_query": 23.2642, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10303.5435, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5703, "query_norm": 2.4727, "queue_k_norm": 2.5288, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3078, "sent_len_1": 66.6728, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.4162, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151700 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3568, "doc_norm": 2.5307, "encoder_q-embeddings": 4934.1357, "encoder_q-layer.0": 3463.4651, "encoder_q-layer.1": 3853.5911, "encoder_q-layer.10": 6871.2446, "encoder_q-layer.11": 16237.6709, "encoder_q-layer.2": 4231.27, "encoder_q-layer.3": 4174.6533, "encoder_q-layer.4": 4577.481, "encoder_q-layer.5": 4754.877, "encoder_q-layer.6": 5475.2388, "encoder_q-layer.7": 6102.8301, "encoder_q-layer.8": 7073.5439, "encoder_q-layer.9": 6346.9717, "epoch": 0.66, "inbatch_neg_score": 4.5745, "inbatch_pos_score": 5.1953, "learning_rate": 1.268421052631579e-05, "loss": 3.3568, "norm_diff": 0.0572, "norm_loss": 0.0, "num_token_doc": 66.6738, "num_token_overlap": 11.6955, "num_token_query": 31.3919, "num_token_union": 65.0507, "num_word_context": 202.2544, "num_word_doc": 49.7619, "num_word_query": 23.296, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10261.0946, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5742, "query_norm": 2.4735, "queue_k_norm": 2.531, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3919, "sent_len_1": 66.6738, "sent_len_max_0": 127.97, "sent_len_max_1": 208.2525, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 151800 }, { "accuracy": 49.6094, "active_queue_size": 16384.0, "cl_loss": 3.3634, "doc_norm": 2.5331, "encoder_q-embeddings": 5098.4678, "encoder_q-layer.0": 3569.9514, "encoder_q-layer.1": 3918.7134, "encoder_q-layer.10": 6789.187, "encoder_q-layer.11": 16674.8242, "encoder_q-layer.2": 4405.9277, "encoder_q-layer.3": 4609.0913, "encoder_q-layer.4": 4831.9644, "encoder_q-layer.5": 4917.9351, "encoder_q-layer.6": 5593.0996, "encoder_q-layer.7": 6106.6562, "encoder_q-layer.8": 6990.3633, "encoder_q-layer.9": 6330.6929, "epoch": 0.66, "inbatch_neg_score": 4.5785, "inbatch_pos_score": 5.1797, "learning_rate": 1.2657894736842104e-05, "loss": 3.3634, "norm_diff": 0.0632, "norm_loss": 0.0, "num_token_doc": 66.8081, "num_token_overlap": 11.675, "num_token_query": 31.3301, "num_token_union": 65.1011, "num_word_context": 202.4089, "num_word_doc": 49.8817, "num_word_query": 23.278, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10789.7012, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5781, "query_norm": 2.4699, "queue_k_norm": 2.5309, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3301, "sent_len_1": 66.8081, "sent_len_max_0": 127.9488, "sent_len_max_1": 207.1538, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 151900 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3647, "doc_norm": 2.5339, "encoder_q-embeddings": 4918.79, "encoder_q-layer.0": 3433.3965, "encoder_q-layer.1": 3635.6079, "encoder_q-layer.10": 7028.6597, "encoder_q-layer.11": 16995.3965, "encoder_q-layer.2": 4107.8486, "encoder_q-layer.3": 4237.3481, "encoder_q-layer.4": 4609.5908, "encoder_q-layer.5": 4897.9458, "encoder_q-layer.6": 5450.6401, "encoder_q-layer.7": 6096.3428, "encoder_q-layer.8": 7030.4512, "encoder_q-layer.9": 6359.4004, "epoch": 0.66, "inbatch_neg_score": 4.5797, "inbatch_pos_score": 5.207, "learning_rate": 1.2631578947368422e-05, "loss": 3.3647, "norm_diff": 0.0619, "norm_loss": 0.0, "num_token_doc": 66.5974, "num_token_overlap": 11.6684, "num_token_query": 31.4046, "num_token_union": 65.0511, "num_word_context": 202.1612, "num_word_doc": 49.7295, "num_word_query": 23.3433, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10664.9543, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.582, "query_norm": 2.4719, "queue_k_norm": 2.533, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4046, "sent_len_1": 66.5974, "sent_len_max_0": 127.9488, "sent_len_max_1": 206.2975, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.365, "doc_norm": 2.5317, "encoder_q-embeddings": 4969.394, "encoder_q-layer.0": 3423.3472, "encoder_q-layer.1": 3671.4314, "encoder_q-layer.10": 6961.5537, "encoder_q-layer.11": 17592.082, "encoder_q-layer.2": 4151.1934, "encoder_q-layer.3": 4242.8853, "encoder_q-layer.4": 4517.8091, "encoder_q-layer.5": 4545.6978, "encoder_q-layer.6": 5311.5601, "encoder_q-layer.7": 5985.7461, "encoder_q-layer.8": 7024.583, "encoder_q-layer.9": 6578.0991, "epoch": 0.66, "inbatch_neg_score": 4.5819, "inbatch_pos_score": 5.1875, "learning_rate": 1.2605263157894737e-05, "loss": 3.365, "norm_diff": 0.0615, "norm_loss": 0.0, "num_token_doc": 66.7791, "num_token_overlap": 11.6749, "num_token_query": 31.3528, "num_token_union": 65.152, "num_word_context": 202.2757, "num_word_doc": 49.8313, "num_word_query": 23.2922, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10830.2222, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.582, "query_norm": 2.4703, "queue_k_norm": 2.5343, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3528, "sent_len_1": 66.7791, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.1225, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152100 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.362, "doc_norm": 2.5375, "encoder_q-embeddings": 5641.647, "encoder_q-layer.0": 3933.2402, "encoder_q-layer.1": 4063.3884, "encoder_q-layer.10": 6786.4648, "encoder_q-layer.11": 17064.5977, "encoder_q-layer.2": 4410.3169, "encoder_q-layer.3": 4515.249, "encoder_q-layer.4": 4669.8262, "encoder_q-layer.5": 4884.8892, "encoder_q-layer.6": 5368.5928, "encoder_q-layer.7": 5820.96, "encoder_q-layer.8": 6866.8164, "encoder_q-layer.9": 6399.4961, "epoch": 0.66, "inbatch_neg_score": 4.5882, "inbatch_pos_score": 5.1914, "learning_rate": 1.2578947368421054e-05, "loss": 3.362, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.9227, "num_token_overlap": 11.7186, "num_token_query": 31.4358, "num_token_union": 65.1986, "num_word_context": 202.662, "num_word_doc": 49.9531, "num_word_query": 23.3468, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10769.6303, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5859, "query_norm": 2.4692, "queue_k_norm": 2.5352, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4358, "sent_len_1": 66.9227, "sent_len_max_0": 127.9587, "sent_len_max_1": 209.0288, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152200 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3652, "doc_norm": 2.5341, "encoder_q-embeddings": 5195.2021, "encoder_q-layer.0": 3604.9849, "encoder_q-layer.1": 4056.8042, "encoder_q-layer.10": 6399.5464, "encoder_q-layer.11": 16738.791, "encoder_q-layer.2": 4498.3418, "encoder_q-layer.3": 4641.8706, "encoder_q-layer.4": 5015.1328, "encoder_q-layer.5": 5103.7065, "encoder_q-layer.6": 5481.937, "encoder_q-layer.7": 6293.0386, "encoder_q-layer.8": 7046.8315, "encoder_q-layer.9": 6267.7739, "epoch": 0.66, "inbatch_neg_score": 4.5906, "inbatch_pos_score": 5.1953, "learning_rate": 1.2552631578947369e-05, "loss": 3.3652, "norm_diff": 0.0697, "norm_loss": 0.0, "num_token_doc": 66.889, "num_token_overlap": 11.7176, "num_token_query": 31.4772, "num_token_union": 65.1807, "num_word_context": 202.2953, "num_word_doc": 49.8744, "num_word_query": 23.3876, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10636.9631, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5898, "query_norm": 2.4644, "queue_k_norm": 2.5353, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4772, "sent_len_1": 66.889, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.13, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 152300 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3646, "doc_norm": 2.5374, "encoder_q-embeddings": 2438.3665, "encoder_q-layer.0": 1721.4371, "encoder_q-layer.1": 1808.4718, "encoder_q-layer.10": 3186.2212, "encoder_q-layer.11": 8220.6299, "encoder_q-layer.2": 2005.8579, "encoder_q-layer.3": 2077.4524, "encoder_q-layer.4": 2244.0161, "encoder_q-layer.5": 2244.8569, "encoder_q-layer.6": 2678.8982, "encoder_q-layer.7": 3018.3855, "encoder_q-layer.8": 3676.4109, "encoder_q-layer.9": 3172.2283, "epoch": 0.66, "inbatch_neg_score": 4.5902, "inbatch_pos_score": 5.2031, "learning_rate": 1.2526315789473686e-05, "loss": 3.3646, "norm_diff": 0.0658, "norm_loss": 0.0, "num_token_doc": 66.7967, "num_token_overlap": 11.6567, "num_token_query": 31.2775, "num_token_union": 65.0705, "num_word_context": 202.1911, "num_word_doc": 49.8203, "num_word_query": 23.2359, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5196.2025, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.5898, "query_norm": 2.4715, "queue_k_norm": 2.5365, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2775, "sent_len_1": 66.7967, "sent_len_max_0": 127.945, "sent_len_max_1": 209.4775, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 152400 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.362, "doc_norm": 2.5396, "encoder_q-embeddings": 2514.9307, "encoder_q-layer.0": 1745.6093, "encoder_q-layer.1": 1912.0217, "encoder_q-layer.10": 3309.3289, "encoder_q-layer.11": 7717.0049, "encoder_q-layer.2": 2147.4221, "encoder_q-layer.3": 2158.8286, "encoder_q-layer.4": 2278.2524, "encoder_q-layer.5": 2367.2163, "encoder_q-layer.6": 2573.6997, "encoder_q-layer.7": 2899.0691, "encoder_q-layer.8": 3438.6936, "encoder_q-layer.9": 3200.0667, "epoch": 0.66, "inbatch_neg_score": 4.591, "inbatch_pos_score": 5.2188, "learning_rate": 1.25e-05, "loss": 3.362, "norm_diff": 0.0666, "norm_loss": 0.0, "num_token_doc": 66.8168, "num_token_overlap": 11.6664, "num_token_query": 31.3274, "num_token_union": 65.1147, "num_word_context": 202.0695, "num_word_doc": 49.8205, "num_word_query": 23.2601, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5010.7544, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.5898, "query_norm": 2.473, "queue_k_norm": 2.5374, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3274, "sent_len_1": 66.8168, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.7688, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 152500 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3624, "doc_norm": 2.5396, "encoder_q-embeddings": 3085.1304, "encoder_q-layer.0": 2362.4575, "encoder_q-layer.1": 2502.749, "encoder_q-layer.10": 3427.9075, "encoder_q-layer.11": 8732.5459, "encoder_q-layer.2": 2995.8296, "encoder_q-layer.3": 3124.8997, "encoder_q-layer.4": 2859.4546, "encoder_q-layer.5": 2703.2266, "encoder_q-layer.6": 2963.8003, "encoder_q-layer.7": 3455.8743, "encoder_q-layer.8": 3771.8064, "encoder_q-layer.9": 3285.4097, "epoch": 0.66, "inbatch_neg_score": 4.592, "inbatch_pos_score": 5.1992, "learning_rate": 1.2473684210526317e-05, "loss": 3.3624, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.7673, "num_token_overlap": 11.6838, "num_token_query": 31.3648, "num_token_union": 65.0731, "num_word_context": 202.0707, "num_word_doc": 49.812, "num_word_query": 23.2822, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5799.2829, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5938, "query_norm": 2.4666, "queue_k_norm": 2.5386, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3648, "sent_len_1": 66.7673, "sent_len_max_0": 127.98, "sent_len_max_1": 209.6125, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152600 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3517, "doc_norm": 2.5414, "encoder_q-embeddings": 2464.2749, "encoder_q-layer.0": 1654.2068, "encoder_q-layer.1": 1802.0325, "encoder_q-layer.10": 3264.4282, "encoder_q-layer.11": 8106.7305, "encoder_q-layer.2": 1993.6787, "encoder_q-layer.3": 2047.0789, "encoder_q-layer.4": 2229.1396, "encoder_q-layer.5": 2286.0308, "encoder_q-layer.6": 2540.5137, "encoder_q-layer.7": 2857.5483, "encoder_q-layer.8": 3450.292, "encoder_q-layer.9": 3257.2583, "epoch": 0.66, "inbatch_neg_score": 4.5947, "inbatch_pos_score": 5.2031, "learning_rate": 1.2447368421052632e-05, "loss": 3.3517, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.9509, "num_token_overlap": 11.6935, "num_token_query": 31.328, "num_token_union": 65.1608, "num_word_context": 202.3337, "num_word_doc": 49.9383, "num_word_query": 23.2531, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5194.6522, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.5938, "query_norm": 2.4678, "queue_k_norm": 2.5402, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.328, "sent_len_1": 66.9509, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.8375, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152700 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3576, "doc_norm": 2.5398, "encoder_q-embeddings": 2231.0659, "encoder_q-layer.0": 1623.386, "encoder_q-layer.1": 1760.672, "encoder_q-layer.10": 3355.707, "encoder_q-layer.11": 7856.7285, "encoder_q-layer.2": 1986.6898, "encoder_q-layer.3": 2023.6423, "encoder_q-layer.4": 2156.468, "encoder_q-layer.5": 2207.8494, "encoder_q-layer.6": 2470.3147, "encoder_q-layer.7": 2798.5005, "encoder_q-layer.8": 3336.9998, "encoder_q-layer.9": 3230.5176, "epoch": 0.66, "inbatch_neg_score": 4.5962, "inbatch_pos_score": 5.2266, "learning_rate": 1.2421052631578949e-05, "loss": 3.3576, "norm_diff": 0.0653, "norm_loss": 0.0, "num_token_doc": 66.7585, "num_token_overlap": 11.6848, "num_token_query": 31.3801, "num_token_union": 65.1195, "num_word_context": 202.0913, "num_word_doc": 49.788, "num_word_query": 23.3063, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4990.6849, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.5977, "query_norm": 2.4745, "queue_k_norm": 2.5401, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3801, "sent_len_1": 66.7585, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.8313, "stdk": 0.0495, "stdq": 0.0438, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152800 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3632, "doc_norm": 2.5395, "encoder_q-embeddings": 2420.0774, "encoder_q-layer.0": 1682.0496, "encoder_q-layer.1": 1720.854, "encoder_q-layer.10": 3142.5642, "encoder_q-layer.11": 7916.4111, "encoder_q-layer.2": 1904.7244, "encoder_q-layer.3": 1971.7716, "encoder_q-layer.4": 2116.8154, "encoder_q-layer.5": 2199.1841, "encoder_q-layer.6": 2461.0535, "encoder_q-layer.7": 2776.0933, "encoder_q-layer.8": 3319.6575, "encoder_q-layer.9": 3046.459, "epoch": 0.66, "inbatch_neg_score": 4.5964, "inbatch_pos_score": 5.207, "learning_rate": 1.2394736842105264e-05, "loss": 3.3632, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.7612, "num_token_overlap": 11.6462, "num_token_query": 31.2584, "num_token_union": 65.0776, "num_word_context": 202.3608, "num_word_doc": 49.8624, "num_word_query": 23.2208, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5016.3043, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.5977, "query_norm": 2.4624, "queue_k_norm": 2.541, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2584, "sent_len_1": 66.7612, "sent_len_max_0": 127.9412, "sent_len_max_1": 208.2088, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 152900 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3582, "doc_norm": 2.542, "encoder_q-embeddings": 2711.1172, "encoder_q-layer.0": 1902.8787, "encoder_q-layer.1": 2073.2151, "encoder_q-layer.10": 3285.6416, "encoder_q-layer.11": 8394.8984, "encoder_q-layer.2": 2437.3103, "encoder_q-layer.3": 2525.9048, "encoder_q-layer.4": 2791.6843, "encoder_q-layer.5": 2892.2678, "encoder_q-layer.6": 3085.9094, "encoder_q-layer.7": 3348.2415, "encoder_q-layer.8": 3587.2991, "encoder_q-layer.9": 3183.9421, "epoch": 0.66, "inbatch_neg_score": 4.5984, "inbatch_pos_score": 5.207, "learning_rate": 1.2368421052631579e-05, "loss": 3.3582, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.8329, "num_token_overlap": 11.675, "num_token_query": 31.3363, "num_token_union": 65.1406, "num_word_context": 202.1746, "num_word_doc": 49.8815, "num_word_query": 23.2668, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5625.9928, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.5977, "query_norm": 2.464, "queue_k_norm": 2.5412, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3363, "sent_len_1": 66.8329, "sent_len_max_0": 128.0, "sent_len_max_1": 210.1912, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153000 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3722, "doc_norm": 2.5407, "encoder_q-embeddings": 2249.0142, "encoder_q-layer.0": 1594.668, "encoder_q-layer.1": 1735.8174, "encoder_q-layer.10": 3223.7837, "encoder_q-layer.11": 8226.667, "encoder_q-layer.2": 1896.9408, "encoder_q-layer.3": 1929.644, "encoder_q-layer.4": 2094.5549, "encoder_q-layer.5": 2143.8091, "encoder_q-layer.6": 2443.2576, "encoder_q-layer.7": 2871.9629, "encoder_q-layer.8": 3432.4834, "encoder_q-layer.9": 3192.8604, "epoch": 0.66, "inbatch_neg_score": 4.5996, "inbatch_pos_score": 5.2148, "learning_rate": 1.2342105263157896e-05, "loss": 3.3722, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.7809, "num_token_overlap": 11.6583, "num_token_query": 31.4303, "num_token_union": 65.2036, "num_word_context": 202.347, "num_word_doc": 49.8583, "num_word_query": 23.3607, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5057.3503, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6016, "query_norm": 2.4656, "queue_k_norm": 2.5416, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4303, "sent_len_1": 66.7809, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.1962, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153100 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.344, "doc_norm": 2.5425, "encoder_q-embeddings": 2384.6189, "encoder_q-layer.0": 1673.5985, "encoder_q-layer.1": 1761.4393, "encoder_q-layer.10": 3292.3369, "encoder_q-layer.11": 8454.792, "encoder_q-layer.2": 1991.9053, "encoder_q-layer.3": 1998.2365, "encoder_q-layer.4": 2082.4524, "encoder_q-layer.5": 2183.2705, "encoder_q-layer.6": 2570.9265, "encoder_q-layer.7": 2831.9587, "encoder_q-layer.8": 3430.594, "encoder_q-layer.9": 3307.1221, "epoch": 0.66, "inbatch_neg_score": 4.605, "inbatch_pos_score": 5.2227, "learning_rate": 1.231578947368421e-05, "loss": 3.344, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.8384, "num_token_overlap": 11.6823, "num_token_query": 31.4536, "num_token_union": 65.199, "num_word_context": 202.1082, "num_word_doc": 49.8417, "num_word_query": 23.3629, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5116.8111, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6055, "query_norm": 2.468, "queue_k_norm": 2.543, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4536, "sent_len_1": 66.8384, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.9375, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 153200 }, { "accuracy": 49.3652, "active_queue_size": 16384.0, "cl_loss": 3.351, "doc_norm": 2.5453, "encoder_q-embeddings": 2432.1895, "encoder_q-layer.0": 1724.1895, "encoder_q-layer.1": 1812.698, "encoder_q-layer.10": 3294.0095, "encoder_q-layer.11": 8312.5264, "encoder_q-layer.2": 2041.7736, "encoder_q-layer.3": 2119.9531, "encoder_q-layer.4": 2333.2856, "encoder_q-layer.5": 2336.5339, "encoder_q-layer.6": 2657.7407, "encoder_q-layer.7": 2884.9124, "encoder_q-layer.8": 3375.2053, "encoder_q-layer.9": 3176.7097, "epoch": 0.67, "inbatch_neg_score": 4.6084, "inbatch_pos_score": 5.2188, "learning_rate": 1.2289473684210527e-05, "loss": 3.351, "norm_diff": 0.072, "norm_loss": 0.0, "num_token_doc": 66.8008, "num_token_overlap": 11.6519, "num_token_query": 31.3017, "num_token_union": 65.0912, "num_word_context": 202.4014, "num_word_doc": 49.8235, "num_word_query": 23.2335, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5294.7762, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6094, "query_norm": 2.4734, "queue_k_norm": 2.5443, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3017, "sent_len_1": 66.8008, "sent_len_max_0": 127.9425, "sent_len_max_1": 212.055, "stdk": 0.0497, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 153300 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.3633, "doc_norm": 2.5449, "encoder_q-embeddings": 2624.7734, "encoder_q-layer.0": 1814.5898, "encoder_q-layer.1": 2029.1295, "encoder_q-layer.10": 3637.261, "encoder_q-layer.11": 8891.2227, "encoder_q-layer.2": 2298.4141, "encoder_q-layer.3": 2337.6313, "encoder_q-layer.4": 2529.5632, "encoder_q-layer.5": 2605.3865, "encoder_q-layer.6": 2865.4597, "encoder_q-layer.7": 3383.6638, "encoder_q-layer.8": 3768.0581, "encoder_q-layer.9": 3572.2483, "epoch": 0.67, "inbatch_neg_score": 4.6122, "inbatch_pos_score": 5.2266, "learning_rate": 1.2263157894736844e-05, "loss": 3.3633, "norm_diff": 0.068, "norm_loss": 0.0, "num_token_doc": 66.7299, "num_token_overlap": 11.6735, "num_token_query": 31.3326, "num_token_union": 65.0619, "num_word_context": 202.1201, "num_word_doc": 49.7768, "num_word_query": 23.2716, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5507.4576, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6094, "query_norm": 2.4769, "queue_k_norm": 2.5443, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3326, "sent_len_1": 66.7299, "sent_len_max_0": 127.9775, "sent_len_max_1": 211.2212, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 153400 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.3654, "doc_norm": 2.5436, "encoder_q-embeddings": 2488.7639, "encoder_q-layer.0": 1773.7859, "encoder_q-layer.1": 1892.0613, "encoder_q-layer.10": 3602.5222, "encoder_q-layer.11": 8638.8447, "encoder_q-layer.2": 2143.678, "encoder_q-layer.3": 2141.313, "encoder_q-layer.4": 2300.6028, "encoder_q-layer.5": 2352.2192, "encoder_q-layer.6": 2612.6582, "encoder_q-layer.7": 2789.093, "encoder_q-layer.8": 3340.7593, "encoder_q-layer.9": 3179.0908, "epoch": 0.67, "inbatch_neg_score": 4.6135, "inbatch_pos_score": 5.2422, "learning_rate": 1.2236842105263159e-05, "loss": 3.3654, "norm_diff": 0.0711, "norm_loss": 0.0, "num_token_doc": 66.813, "num_token_overlap": 11.6954, "num_token_query": 31.4172, "num_token_union": 65.1505, "num_word_context": 202.4193, "num_word_doc": 49.8304, "num_word_query": 23.3298, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5307.6403, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6133, "query_norm": 2.4726, "queue_k_norm": 2.5439, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4172, "sent_len_1": 66.813, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.9575, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153500 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3656, "doc_norm": 2.544, "encoder_q-embeddings": 2732.6587, "encoder_q-layer.0": 1946.1157, "encoder_q-layer.1": 2107.4268, "encoder_q-layer.10": 3688.7168, "encoder_q-layer.11": 8515.1094, "encoder_q-layer.2": 2336.1763, "encoder_q-layer.3": 2461.2705, "encoder_q-layer.4": 2672.5762, "encoder_q-layer.5": 2864.9622, "encoder_q-layer.6": 3086.6721, "encoder_q-layer.7": 3418.2048, "encoder_q-layer.8": 3688.2534, "encoder_q-layer.9": 3352.2969, "epoch": 0.67, "inbatch_neg_score": 4.6186, "inbatch_pos_score": 5.2383, "learning_rate": 1.2210526315789474e-05, "loss": 3.3656, "norm_diff": 0.0639, "norm_loss": 0.0, "num_token_doc": 66.7825, "num_token_overlap": 11.6522, "num_token_query": 31.303, "num_token_union": 65.1057, "num_word_context": 202.3189, "num_word_doc": 49.8269, "num_word_query": 23.2379, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5558.6896, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6172, "query_norm": 2.4801, "queue_k_norm": 2.5444, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.303, "sent_len_1": 66.7825, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.5712, "stdk": 0.0494, "stdq": 0.0438, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153600 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3707, "doc_norm": 2.545, "encoder_q-embeddings": 2348.4768, "encoder_q-layer.0": 1654.1588, "encoder_q-layer.1": 1723.3094, "encoder_q-layer.10": 3244.5549, "encoder_q-layer.11": 8322.5488, "encoder_q-layer.2": 1931.9702, "encoder_q-layer.3": 2018.9095, "encoder_q-layer.4": 2143.6382, "encoder_q-layer.5": 2259.9724, "encoder_q-layer.6": 2578.926, "encoder_q-layer.7": 2944.7266, "encoder_q-layer.8": 3360.8608, "encoder_q-layer.9": 3126.6499, "epoch": 0.67, "inbatch_neg_score": 4.6243, "inbatch_pos_score": 5.2422, "learning_rate": 1.218421052631579e-05, "loss": 3.3707, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.7823, "num_token_overlap": 11.6872, "num_token_query": 31.3688, "num_token_union": 65.1153, "num_word_context": 202.131, "num_word_doc": 49.8077, "num_word_query": 23.2898, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5111.1281, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.625, "query_norm": 2.4733, "queue_k_norm": 2.5454, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3688, "sent_len_1": 66.7823, "sent_len_max_0": 127.9925, "sent_len_max_1": 211.4225, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153700 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3632, "doc_norm": 2.5446, "encoder_q-embeddings": 2338.9351, "encoder_q-layer.0": 1675.958, "encoder_q-layer.1": 1825.4161, "encoder_q-layer.10": 3244.7356, "encoder_q-layer.11": 8018.4639, "encoder_q-layer.2": 2038.2283, "encoder_q-layer.3": 2056.6707, "encoder_q-layer.4": 2164.7354, "encoder_q-layer.5": 2323.3708, "encoder_q-layer.6": 2580.8372, "encoder_q-layer.7": 2867.8518, "encoder_q-layer.8": 3370.0161, "encoder_q-layer.9": 3068.3528, "epoch": 0.67, "inbatch_neg_score": 4.6252, "inbatch_pos_score": 5.2539, "learning_rate": 1.2157894736842105e-05, "loss": 3.3632, "norm_diff": 0.0639, "norm_loss": 0.0, "num_token_doc": 66.7523, "num_token_overlap": 11.6473, "num_token_query": 31.2828, "num_token_union": 65.0684, "num_word_context": 202.5301, "num_word_doc": 49.801, "num_word_query": 23.2311, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5103.3734, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.625, "query_norm": 2.4807, "queue_k_norm": 2.5449, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2828, "sent_len_1": 66.7523, "sent_len_max_0": 127.97, "sent_len_max_1": 210.9288, "stdk": 0.0494, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153800 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3552, "doc_norm": 2.5491, "encoder_q-embeddings": 2397.1472, "encoder_q-layer.0": 1652.7783, "encoder_q-layer.1": 1724.9227, "encoder_q-layer.10": 3049.4866, "encoder_q-layer.11": 7757.29, "encoder_q-layer.2": 1931.0411, "encoder_q-layer.3": 1988.4574, "encoder_q-layer.4": 2083.0347, "encoder_q-layer.5": 2044.4532, "encoder_q-layer.6": 2431.1436, "encoder_q-layer.7": 2668.8391, "encoder_q-layer.8": 3222.2957, "encoder_q-layer.9": 2978.2173, "epoch": 0.67, "inbatch_neg_score": 4.6284, "inbatch_pos_score": 5.2422, "learning_rate": 1.2131578947368422e-05, "loss": 3.3552, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.8369, "num_token_overlap": 11.6893, "num_token_query": 31.3873, "num_token_union": 65.1556, "num_word_context": 202.3121, "num_word_doc": 49.8657, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 4940.923, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6289, "query_norm": 2.4763, "queue_k_norm": 2.5454, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3873, "sent_len_1": 66.8369, "sent_len_max_0": 127.985, "sent_len_max_1": 209.5637, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 153900 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.3606, "doc_norm": 2.5468, "encoder_q-embeddings": 2568.3838, "encoder_q-layer.0": 1764.0415, "encoder_q-layer.1": 1900.0189, "encoder_q-layer.10": 3620.4021, "encoder_q-layer.11": 8294.0586, "encoder_q-layer.2": 2164.3555, "encoder_q-layer.3": 2183.7126, "encoder_q-layer.4": 2307.9006, "encoder_q-layer.5": 2321.8296, "encoder_q-layer.6": 2602.0298, "encoder_q-layer.7": 2913.853, "encoder_q-layer.8": 3630.561, "encoder_q-layer.9": 3282.262, "epoch": 0.67, "inbatch_neg_score": 4.6308, "inbatch_pos_score": 5.2305, "learning_rate": 1.2105263157894737e-05, "loss": 3.3606, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.8163, "num_token_overlap": 11.692, "num_token_query": 31.4156, "num_token_union": 65.1623, "num_word_context": 202.4039, "num_word_doc": 49.832, "num_word_query": 23.3402, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5301.8242, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6289, "query_norm": 2.4741, "queue_k_norm": 2.5484, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4156, "sent_len_1": 66.8163, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.955, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 154000 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3566, "doc_norm": 2.5497, "encoder_q-embeddings": 2728.377, "encoder_q-layer.0": 1840.8379, "encoder_q-layer.1": 1925.0266, "encoder_q-layer.10": 3454.7864, "encoder_q-layer.11": 8358.6934, "encoder_q-layer.2": 2168.5391, "encoder_q-layer.3": 2144.4639, "encoder_q-layer.4": 2256.4529, "encoder_q-layer.5": 2349.6509, "encoder_q-layer.6": 2697.7209, "encoder_q-layer.7": 3003.2034, "encoder_q-layer.8": 3576.199, "encoder_q-layer.9": 3258.334, "epoch": 0.67, "inbatch_neg_score": 4.6332, "inbatch_pos_score": 5.2617, "learning_rate": 1.2078947368421054e-05, "loss": 3.3566, "norm_diff": 0.0672, "norm_loss": 0.0, "num_token_doc": 66.8409, "num_token_overlap": 11.6818, "num_token_query": 31.4055, "num_token_union": 65.1748, "num_word_context": 202.3968, "num_word_doc": 49.8729, "num_word_query": 23.3239, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5333.73, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6328, "query_norm": 2.4825, "queue_k_norm": 2.549, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4055, "sent_len_1": 66.8409, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.2612, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 154100 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.365, "doc_norm": 2.5484, "encoder_q-embeddings": 2390.9104, "encoder_q-layer.0": 1640.5632, "encoder_q-layer.1": 1767.8695, "encoder_q-layer.10": 3329.8862, "encoder_q-layer.11": 8153.0415, "encoder_q-layer.2": 1973.1014, "encoder_q-layer.3": 1976.1527, "encoder_q-layer.4": 2140.3486, "encoder_q-layer.5": 2140.5891, "encoder_q-layer.6": 2417.3743, "encoder_q-layer.7": 2761.1118, "encoder_q-layer.8": 3386.894, "encoder_q-layer.9": 3104.3491, "epoch": 0.67, "inbatch_neg_score": 4.631, "inbatch_pos_score": 5.2383, "learning_rate": 1.205263157894737e-05, "loss": 3.365, "norm_diff": 0.0778, "norm_loss": 0.0, "num_token_doc": 66.7967, "num_token_overlap": 11.659, "num_token_query": 31.3666, "num_token_union": 65.1391, "num_word_context": 202.0925, "num_word_doc": 49.8465, "num_word_query": 23.2869, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5100.1825, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 4.6328, "query_norm": 2.4706, "queue_k_norm": 2.5481, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3666, "sent_len_1": 66.7967, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.4712, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154200 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3606, "doc_norm": 2.5499, "encoder_q-embeddings": 2699.4045, "encoder_q-layer.0": 1990.1539, "encoder_q-layer.1": 2023.0959, "encoder_q-layer.10": 3307.9827, "encoder_q-layer.11": 8291.1494, "encoder_q-layer.2": 2270.7422, "encoder_q-layer.3": 2377.7212, "encoder_q-layer.4": 2682.3015, "encoder_q-layer.5": 2870.1577, "encoder_q-layer.6": 2993.7886, "encoder_q-layer.7": 3515.9766, "encoder_q-layer.8": 3650.0159, "encoder_q-layer.9": 3189.4126, "epoch": 0.67, "inbatch_neg_score": 4.634, "inbatch_pos_score": 5.25, "learning_rate": 1.2026315789473684e-05, "loss": 3.3606, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.7743, "num_token_overlap": 11.6764, "num_token_query": 31.5084, "num_token_union": 65.2059, "num_word_context": 202.5146, "num_word_doc": 49.8048, "num_word_query": 23.3889, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5523.8799, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.475, "queue_k_norm": 2.5488, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5084, "sent_len_1": 66.7743, "sent_len_max_0": 128.0, "sent_len_max_1": 207.3963, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154300 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3527, "doc_norm": 2.5511, "encoder_q-embeddings": 4854.0815, "encoder_q-layer.0": 3455.6499, "encoder_q-layer.1": 3632.7761, "encoder_q-layer.10": 6647.8765, "encoder_q-layer.11": 15767.9834, "encoder_q-layer.2": 4002.4133, "encoder_q-layer.3": 4150.2446, "encoder_q-layer.4": 4353.7007, "encoder_q-layer.5": 4500.8975, "encoder_q-layer.6": 5127.2578, "encoder_q-layer.7": 5701.5322, "encoder_q-layer.8": 7070.5635, "encoder_q-layer.9": 6247.8003, "epoch": 0.67, "inbatch_neg_score": 4.6343, "inbatch_pos_score": 5.2578, "learning_rate": 1.2e-05, "loss": 3.3527, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.5799, "num_token_overlap": 11.6574, "num_token_query": 31.3899, "num_token_union": 65.0215, "num_word_context": 202.0848, "num_word_doc": 49.6825, "num_word_query": 23.3126, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10036.567, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4745, "queue_k_norm": 2.5493, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3899, "sent_len_1": 66.5799, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.7262, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154400 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3616, "doc_norm": 2.5476, "encoder_q-embeddings": 4737.4131, "encoder_q-layer.0": 3295.9297, "encoder_q-layer.1": 3576.1934, "encoder_q-layer.10": 6806.8457, "encoder_q-layer.11": 17019.957, "encoder_q-layer.2": 3959.8374, "encoder_q-layer.3": 4153.54, "encoder_q-layer.4": 4376.5171, "encoder_q-layer.5": 4564.2505, "encoder_q-layer.6": 5142.9917, "encoder_q-layer.7": 5683.0742, "encoder_q-layer.8": 6863.0039, "encoder_q-layer.9": 6405.1992, "epoch": 0.67, "inbatch_neg_score": 4.6328, "inbatch_pos_score": 5.2539, "learning_rate": 1.1973684210526315e-05, "loss": 3.3616, "norm_diff": 0.0712, "norm_loss": 0.0, "num_token_doc": 66.7657, "num_token_overlap": 11.6847, "num_token_query": 31.3199, "num_token_union": 65.1082, "num_word_context": 202.102, "num_word_doc": 49.8593, "num_word_query": 23.2819, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10390.677, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4764, "queue_k_norm": 2.5495, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3199, "sent_len_1": 66.7657, "sent_len_max_0": 127.9762, "sent_len_max_1": 205.8275, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154500 }, { "accuracy": 49.9512, "active_queue_size": 16384.0, "cl_loss": 3.3588, "doc_norm": 2.5514, "encoder_q-embeddings": 4926.5078, "encoder_q-layer.0": 3490.0129, "encoder_q-layer.1": 3671.4119, "encoder_q-layer.10": 6791.6343, "encoder_q-layer.11": 17193.0352, "encoder_q-layer.2": 4087.5759, "encoder_q-layer.3": 4263.1699, "encoder_q-layer.4": 4516.8159, "encoder_q-layer.5": 4598.2261, "encoder_q-layer.6": 5463.1528, "encoder_q-layer.7": 6188.4038, "encoder_q-layer.8": 7389.5181, "encoder_q-layer.9": 6570.5625, "epoch": 0.67, "inbatch_neg_score": 4.633, "inbatch_pos_score": 5.2266, "learning_rate": 1.1947368421052632e-05, "loss": 3.3588, "norm_diff": 0.0812, "norm_loss": 0.0, "num_token_doc": 66.7163, "num_token_overlap": 11.7026, "num_token_query": 31.474, "num_token_union": 65.1432, "num_word_context": 202.2684, "num_word_doc": 49.829, "num_word_query": 23.4017, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10852.9687, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4701, "queue_k_norm": 2.5497, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.474, "sent_len_1": 66.7163, "sent_len_max_0": 128.0, "sent_len_max_1": 207.97, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154600 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3773, "doc_norm": 2.5519, "encoder_q-embeddings": 5046.9653, "encoder_q-layer.0": 3526.2803, "encoder_q-layer.1": 3740.5461, "encoder_q-layer.10": 6984.9688, "encoder_q-layer.11": 17484.2539, "encoder_q-layer.2": 4189.6299, "encoder_q-layer.3": 4370.3462, "encoder_q-layer.4": 4760.0566, "encoder_q-layer.5": 4957.8564, "encoder_q-layer.6": 5652.375, "encoder_q-layer.7": 6053.1382, "encoder_q-layer.8": 7296.3462, "encoder_q-layer.9": 6798.437, "epoch": 0.67, "inbatch_neg_score": 4.6317, "inbatch_pos_score": 5.2305, "learning_rate": 1.1921052631578949e-05, "loss": 3.3773, "norm_diff": 0.0767, "norm_loss": 0.0, "num_token_doc": 66.7742, "num_token_overlap": 11.6749, "num_token_query": 31.3265, "num_token_union": 65.0364, "num_word_context": 202.1666, "num_word_doc": 49.8126, "num_word_query": 23.2755, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11123.3823, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4753, "queue_k_norm": 2.5495, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3265, "sent_len_1": 66.7742, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.8237, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154700 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3636, "doc_norm": 2.5513, "encoder_q-embeddings": 5238.1792, "encoder_q-layer.0": 3597.364, "encoder_q-layer.1": 3883.4385, "encoder_q-layer.10": 6983.0518, "encoder_q-layer.11": 17156.998, "encoder_q-layer.2": 4126.7681, "encoder_q-layer.3": 4266.4316, "encoder_q-layer.4": 4728.2939, "encoder_q-layer.5": 4792.4824, "encoder_q-layer.6": 5205.5596, "encoder_q-layer.7": 6052.9414, "encoder_q-layer.8": 7350.085, "encoder_q-layer.9": 6630.7524, "epoch": 0.67, "inbatch_neg_score": 4.6314, "inbatch_pos_score": 5.25, "learning_rate": 1.1894736842105264e-05, "loss": 3.3636, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.7161, "num_token_overlap": 11.6778, "num_token_query": 31.3886, "num_token_union": 65.0969, "num_word_context": 202.3157, "num_word_doc": 49.7683, "num_word_query": 23.3304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10838.1878, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4747, "queue_k_norm": 2.5513, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3886, "sent_len_1": 66.7161, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.135, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 154800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3573, "doc_norm": 2.5495, "encoder_q-embeddings": 5045.8306, "encoder_q-layer.0": 3546.6892, "encoder_q-layer.1": 3745.9741, "encoder_q-layer.10": 6401.5435, "encoder_q-layer.11": 16497.4746, "encoder_q-layer.2": 4269.3911, "encoder_q-layer.3": 4438.8359, "encoder_q-layer.4": 4775.4556, "encoder_q-layer.5": 5090.9751, "encoder_q-layer.6": 5196.165, "encoder_q-layer.7": 5805.7173, "encoder_q-layer.8": 6956.668, "encoder_q-layer.9": 6323.3916, "epoch": 0.67, "inbatch_neg_score": 4.6335, "inbatch_pos_score": 5.2578, "learning_rate": 1.186842105263158e-05, "loss": 3.3573, "norm_diff": 0.0759, "norm_loss": 0.0, "num_token_doc": 66.7153, "num_token_overlap": 11.7014, "num_token_query": 31.3931, "num_token_union": 65.0772, "num_word_context": 202.0042, "num_word_doc": 49.797, "num_word_query": 23.3256, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10499.7871, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6328, "query_norm": 2.4736, "queue_k_norm": 2.5505, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3931, "sent_len_1": 66.7153, "sent_len_max_0": 127.995, "sent_len_max_1": 207.6637, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 154900 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3648, "doc_norm": 2.5538, "encoder_q-embeddings": 5047.2642, "encoder_q-layer.0": 3583.9661, "encoder_q-layer.1": 4006.9131, "encoder_q-layer.10": 7393.3916, "encoder_q-layer.11": 17235.668, "encoder_q-layer.2": 4379.2026, "encoder_q-layer.3": 4687.4087, "encoder_q-layer.4": 4925.4771, "encoder_q-layer.5": 4994.9907, "encoder_q-layer.6": 5690.8281, "encoder_q-layer.7": 6081.4795, "encoder_q-layer.8": 7172.3105, "encoder_q-layer.9": 6493.3374, "epoch": 0.67, "inbatch_neg_score": 4.6365, "inbatch_pos_score": 5.2422, "learning_rate": 1.1842105263157895e-05, "loss": 3.3648, "norm_diff": 0.0836, "norm_loss": 0.0, "num_token_doc": 66.676, "num_token_overlap": 11.6312, "num_token_query": 31.2987, "num_token_union": 65.0418, "num_word_context": 202.2292, "num_word_doc": 49.7247, "num_word_query": 23.2444, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10836.9483, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6367, "query_norm": 2.4702, "queue_k_norm": 2.5522, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2987, "sent_len_1": 66.676, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.3225, "stdk": 0.0496, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 155000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3533, "doc_norm": 2.5513, "encoder_q-embeddings": 4645.6475, "encoder_q-layer.0": 3236.0208, "encoder_q-layer.1": 3455.7427, "encoder_q-layer.10": 6521.1147, "encoder_q-layer.11": 16348.0918, "encoder_q-layer.2": 3773.5229, "encoder_q-layer.3": 3868.5063, "encoder_q-layer.4": 4199.0146, "encoder_q-layer.5": 4523.2329, "encoder_q-layer.6": 5035.1021, "encoder_q-layer.7": 5738.2197, "encoder_q-layer.8": 6720.2427, "encoder_q-layer.9": 6168.7339, "epoch": 0.67, "inbatch_neg_score": 4.6432, "inbatch_pos_score": 5.2578, "learning_rate": 1.181578947368421e-05, "loss": 3.3533, "norm_diff": 0.0738, "norm_loss": 0.0, "num_token_doc": 66.7613, "num_token_overlap": 11.7146, "num_token_query": 31.4511, "num_token_union": 65.1194, "num_word_context": 202.2366, "num_word_doc": 49.8259, "num_word_query": 23.3691, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10058.0687, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6445, "query_norm": 2.4776, "queue_k_norm": 2.5506, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4511, "sent_len_1": 66.7613, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.6488, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 155100 }, { "accuracy": 52.9785, "active_queue_size": 16384.0, "cl_loss": 3.3601, "doc_norm": 2.5526, "encoder_q-embeddings": 5814.3037, "encoder_q-layer.0": 3930.4946, "encoder_q-layer.1": 4668.8784, "encoder_q-layer.10": 6741.8862, "encoder_q-layer.11": 16506.2324, "encoder_q-layer.2": 5857.3833, "encoder_q-layer.3": 5175.8062, "encoder_q-layer.4": 5369.8467, "encoder_q-layer.5": 5169.5234, "encoder_q-layer.6": 5337.7993, "encoder_q-layer.7": 6186.2471, "encoder_q-layer.8": 7426.0771, "encoder_q-layer.9": 6405.021, "epoch": 0.67, "inbatch_neg_score": 4.644, "inbatch_pos_score": 5.2734, "learning_rate": 1.1789473684210527e-05, "loss": 3.3601, "norm_diff": 0.0748, "norm_loss": 0.0, "num_token_doc": 66.8247, "num_token_overlap": 11.6727, "num_token_query": 31.3615, "num_token_union": 65.1476, "num_word_context": 202.4567, "num_word_doc": 49.8616, "num_word_query": 23.2934, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11076.2493, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6445, "query_norm": 2.4778, "queue_k_norm": 2.5518, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3615, "sent_len_1": 66.8247, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.5137, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155200 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.5514, "encoder_q-embeddings": 4659.4531, "encoder_q-layer.0": 3242.1797, "encoder_q-layer.1": 3544.8809, "encoder_q-layer.10": 6834.3257, "encoder_q-layer.11": 16309.0107, "encoder_q-layer.2": 4032.6687, "encoder_q-layer.3": 4054.6838, "encoder_q-layer.4": 4354.6523, "encoder_q-layer.5": 4404.1187, "encoder_q-layer.6": 5159.814, "encoder_q-layer.7": 5818.5151, "encoder_q-layer.8": 7106.9902, "encoder_q-layer.9": 6379.9219, "epoch": 0.67, "inbatch_neg_score": 4.6501, "inbatch_pos_score": 5.2656, "learning_rate": 1.1763157894736842e-05, "loss": 3.3508, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.7537, "num_token_overlap": 11.6769, "num_token_query": 31.3214, "num_token_union": 65.0898, "num_word_context": 202.3151, "num_word_doc": 49.8398, "num_word_query": 23.2639, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10254.6137, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6484, "query_norm": 2.4827, "queue_k_norm": 2.5523, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3214, "sent_len_1": 66.7537, "sent_len_max_0": 127.94, "sent_len_max_1": 209.1562, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155300 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3564, "doc_norm": 2.5526, "encoder_q-embeddings": 4822.2783, "encoder_q-layer.0": 3463.5034, "encoder_q-layer.1": 3751.8743, "encoder_q-layer.10": 6353.9434, "encoder_q-layer.11": 16070.5303, "encoder_q-layer.2": 4363.9541, "encoder_q-layer.3": 4535.1343, "encoder_q-layer.4": 4713.0137, "encoder_q-layer.5": 4998.5801, "encoder_q-layer.6": 5553.123, "encoder_q-layer.7": 5679.5513, "encoder_q-layer.8": 6509.833, "encoder_q-layer.9": 5986.3486, "epoch": 0.67, "inbatch_neg_score": 4.6572, "inbatch_pos_score": 5.2695, "learning_rate": 1.1736842105263158e-05, "loss": 3.3564, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.8539, "num_token_overlap": 11.7136, "num_token_query": 31.4605, "num_token_union": 65.1773, "num_word_context": 202.4068, "num_word_doc": 49.8842, "num_word_query": 23.3762, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10230.8901, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6562, "query_norm": 2.48, "queue_k_norm": 2.5544, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4605, "sent_len_1": 66.8539, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.6612, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 155400 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3643, "doc_norm": 2.555, "encoder_q-embeddings": 4798.9316, "encoder_q-layer.0": 3327.4846, "encoder_q-layer.1": 3560.8279, "encoder_q-layer.10": 7306.5347, "encoder_q-layer.11": 17049.4551, "encoder_q-layer.2": 3961.3889, "encoder_q-layer.3": 4182.4165, "encoder_q-layer.4": 4487.3853, "encoder_q-layer.5": 4467.8784, "encoder_q-layer.6": 5201.8022, "encoder_q-layer.7": 5706.4512, "encoder_q-layer.8": 6805.5469, "encoder_q-layer.9": 6590.2744, "epoch": 0.67, "inbatch_neg_score": 4.6589, "inbatch_pos_score": 5.2734, "learning_rate": 1.1710526315789475e-05, "loss": 3.3643, "norm_diff": 0.0694, "norm_loss": 0.0, "num_token_doc": 66.8055, "num_token_overlap": 11.6549, "num_token_query": 31.286, "num_token_union": 65.1119, "num_word_context": 202.1668, "num_word_doc": 49.8201, "num_word_query": 23.2235, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10554.1905, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6602, "query_norm": 2.4856, "queue_k_norm": 2.5538, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.286, "sent_len_1": 66.8055, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.4762, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155500 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3656, "doc_norm": 2.555, "encoder_q-embeddings": 5060.2271, "encoder_q-layer.0": 3493.6326, "encoder_q-layer.1": 3640.832, "encoder_q-layer.10": 7946.1963, "encoder_q-layer.11": 17748.6777, "encoder_q-layer.2": 4045.28, "encoder_q-layer.3": 4197.6665, "encoder_q-layer.4": 4477.6572, "encoder_q-layer.5": 4688.3433, "encoder_q-layer.6": 5095.3374, "encoder_q-layer.7": 5909.8691, "encoder_q-layer.8": 7451.8496, "encoder_q-layer.9": 6881.3911, "epoch": 0.68, "inbatch_neg_score": 4.6652, "inbatch_pos_score": 5.2695, "learning_rate": 1.168421052631579e-05, "loss": 3.3656, "norm_diff": 0.071, "norm_loss": 0.0, "num_token_doc": 66.7461, "num_token_overlap": 11.6777, "num_token_query": 31.4014, "num_token_union": 65.1249, "num_word_context": 202.3213, "num_word_doc": 49.7791, "num_word_query": 23.3166, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10948.3156, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6641, "query_norm": 2.484, "queue_k_norm": 2.5548, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4014, "sent_len_1": 66.7461, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.9075, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 155600 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3682, "doc_norm": 2.558, "encoder_q-embeddings": 4813.9331, "encoder_q-layer.0": 3316.5144, "encoder_q-layer.1": 3547.2883, "encoder_q-layer.10": 6662.1846, "encoder_q-layer.11": 16373.666, "encoder_q-layer.2": 3858.9604, "encoder_q-layer.3": 4049.4438, "encoder_q-layer.4": 4417.3071, "encoder_q-layer.5": 4564.7695, "encoder_q-layer.6": 5185.1714, "encoder_q-layer.7": 5804.917, "encoder_q-layer.8": 6717.3965, "encoder_q-layer.9": 6426.4199, "epoch": 0.68, "inbatch_neg_score": 4.6707, "inbatch_pos_score": 5.2891, "learning_rate": 1.1657894736842105e-05, "loss": 3.3682, "norm_diff": 0.0693, "norm_loss": 0.0, "num_token_doc": 66.8346, "num_token_overlap": 11.698, "num_token_query": 31.4205, "num_token_union": 65.176, "num_word_context": 202.2438, "num_word_doc": 49.8751, "num_word_query": 23.3395, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10386.2192, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6719, "query_norm": 2.4887, "queue_k_norm": 2.5547, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4205, "sent_len_1": 66.8346, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.605, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155700 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3671, "doc_norm": 2.5574, "encoder_q-embeddings": 5021.0542, "encoder_q-layer.0": 3561.95, "encoder_q-layer.1": 3630.8311, "encoder_q-layer.10": 6468.5596, "encoder_q-layer.11": 16634.8398, "encoder_q-layer.2": 3996.7683, "encoder_q-layer.3": 4105.8135, "encoder_q-layer.4": 4306.9712, "encoder_q-layer.5": 4625.4048, "encoder_q-layer.6": 5016.3794, "encoder_q-layer.7": 5450.6035, "encoder_q-layer.8": 7052.999, "encoder_q-layer.9": 6387.0293, "epoch": 0.68, "inbatch_neg_score": 4.6762, "inbatch_pos_score": 5.2852, "learning_rate": 1.1631578947368422e-05, "loss": 3.3671, "norm_diff": 0.0665, "norm_loss": 0.0, "num_token_doc": 66.9274, "num_token_overlap": 11.6981, "num_token_query": 31.3522, "num_token_union": 65.1827, "num_word_context": 202.2996, "num_word_doc": 49.9562, "num_word_query": 23.2794, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10447.7892, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6758, "query_norm": 2.4909, "queue_k_norm": 2.5552, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3522, "sent_len_1": 66.9274, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.3638, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155800 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3748, "doc_norm": 2.5558, "encoder_q-embeddings": 4539.8208, "encoder_q-layer.0": 3275.0845, "encoder_q-layer.1": 3506.5598, "encoder_q-layer.10": 6707.1821, "encoder_q-layer.11": 16464.7812, "encoder_q-layer.2": 3755.5305, "encoder_q-layer.3": 3928.8123, "encoder_q-layer.4": 4051.2612, "encoder_q-layer.5": 4404.4004, "encoder_q-layer.6": 5115.9468, "encoder_q-layer.7": 5990.8711, "encoder_q-layer.8": 6752.9141, "encoder_q-layer.9": 6269.1885, "epoch": 0.68, "inbatch_neg_score": 4.6803, "inbatch_pos_score": 5.3008, "learning_rate": 1.1605263157894737e-05, "loss": 3.3748, "norm_diff": 0.0675, "norm_loss": 0.0, "num_token_doc": 66.8025, "num_token_overlap": 11.6782, "num_token_query": 31.4442, "num_token_union": 65.1827, "num_word_context": 202.4455, "num_word_doc": 49.8349, "num_word_query": 23.3629, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10226.158, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6797, "query_norm": 2.4882, "queue_k_norm": 2.5555, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4442, "sent_len_1": 66.8025, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.715, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 155900 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3672, "doc_norm": 2.5558, "encoder_q-embeddings": 4609.7041, "encoder_q-layer.0": 3214.9009, "encoder_q-layer.1": 3367.064, "encoder_q-layer.10": 6952.7021, "encoder_q-layer.11": 16233.1084, "encoder_q-layer.2": 3805.9841, "encoder_q-layer.3": 3844.8545, "encoder_q-layer.4": 4118.5024, "encoder_q-layer.5": 4402.0942, "encoder_q-layer.6": 4920.8184, "encoder_q-layer.7": 5484.4072, "encoder_q-layer.8": 6654.4839, "encoder_q-layer.9": 6158.7158, "epoch": 0.68, "inbatch_neg_score": 4.6848, "inbatch_pos_score": 5.3047, "learning_rate": 1.1578947368421053e-05, "loss": 3.3672, "norm_diff": 0.0691, "norm_loss": 0.0, "num_token_doc": 66.7524, "num_token_overlap": 11.7041, "num_token_query": 31.5607, "num_token_union": 65.1979, "num_word_context": 202.2293, "num_word_doc": 49.8184, "num_word_query": 23.4546, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10112.1771, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6836, "query_norm": 2.4867, "queue_k_norm": 2.5558, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5607, "sent_len_1": 66.7524, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.95, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156000 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3716, "doc_norm": 2.555, "encoder_q-embeddings": 4972.5215, "encoder_q-layer.0": 3431.8701, "encoder_q-layer.1": 3626.3645, "encoder_q-layer.10": 7542.7759, "encoder_q-layer.11": 17146.8027, "encoder_q-layer.2": 4171.2466, "encoder_q-layer.3": 4435.7822, "encoder_q-layer.4": 4800.4551, "encoder_q-layer.5": 5025.7744, "encoder_q-layer.6": 5708.9072, "encoder_q-layer.7": 6416.8223, "encoder_q-layer.8": 7565.582, "encoder_q-layer.9": 6724.1074, "epoch": 0.68, "inbatch_neg_score": 4.6906, "inbatch_pos_score": 5.3086, "learning_rate": 1.1552631578947368e-05, "loss": 3.3716, "norm_diff": 0.0678, "norm_loss": 0.0, "num_token_doc": 66.713, "num_token_overlap": 11.6773, "num_token_query": 31.3885, "num_token_union": 65.0831, "num_word_context": 202.2886, "num_word_doc": 49.7871, "num_word_query": 23.3049, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10887.6794, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6914, "query_norm": 2.4872, "queue_k_norm": 2.5575, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3885, "sent_len_1": 66.713, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.2363, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156100 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.365, "doc_norm": 2.5592, "encoder_q-embeddings": 5611.3452, "encoder_q-layer.0": 3981.0181, "encoder_q-layer.1": 4269.6699, "encoder_q-layer.10": 6524.8218, "encoder_q-layer.11": 16603.0195, "encoder_q-layer.2": 4794.7598, "encoder_q-layer.3": 4920.9136, "encoder_q-layer.4": 5447.4219, "encoder_q-layer.5": 5770.3872, "encoder_q-layer.6": 6325.1641, "encoder_q-layer.7": 6418.9277, "encoder_q-layer.8": 7198.3818, "encoder_q-layer.9": 6331.3418, "epoch": 0.68, "inbatch_neg_score": 4.6964, "inbatch_pos_score": 5.3047, "learning_rate": 1.1526315789473685e-05, "loss": 3.365, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.7227, "num_token_overlap": 11.6888, "num_token_query": 31.3674, "num_token_union": 65.0525, "num_word_context": 202.288, "num_word_doc": 49.7916, "num_word_query": 23.2959, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11045.6476, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.6953, "query_norm": 2.4897, "queue_k_norm": 2.5578, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3674, "sent_len_1": 66.7227, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.0575, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156200 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3624, "doc_norm": 2.5577, "encoder_q-embeddings": 4962.9976, "encoder_q-layer.0": 3529.2341, "encoder_q-layer.1": 3806.6953, "encoder_q-layer.10": 6815.2393, "encoder_q-layer.11": 16740.8457, "encoder_q-layer.2": 4281.4551, "encoder_q-layer.3": 4489.4893, "encoder_q-layer.4": 4825.0464, "encoder_q-layer.5": 4919.3247, "encoder_q-layer.6": 5450.2441, "encoder_q-layer.7": 5905.377, "encoder_q-layer.8": 6899.4438, "encoder_q-layer.9": 6380.4834, "epoch": 0.68, "inbatch_neg_score": 4.7048, "inbatch_pos_score": 5.3164, "learning_rate": 1.1500000000000002e-05, "loss": 3.3624, "norm_diff": 0.0663, "norm_loss": 0.0, "num_token_doc": 66.7569, "num_token_overlap": 11.6798, "num_token_query": 31.348, "num_token_union": 65.09, "num_word_context": 202.5852, "num_word_doc": 49.8043, "num_word_query": 23.2788, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10758.6274, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.7031, "query_norm": 2.4915, "queue_k_norm": 2.5584, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.348, "sent_len_1": 66.7569, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.94, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156300 }, { "accuracy": 53.0762, "active_queue_size": 16384.0, "cl_loss": 3.3633, "doc_norm": 2.5585, "encoder_q-embeddings": 9437.2031, "encoder_q-layer.0": 6641.2788, "encoder_q-layer.1": 6887.8149, "encoder_q-layer.10": 12393.4834, "encoder_q-layer.11": 30790.8652, "encoder_q-layer.2": 7797.1426, "encoder_q-layer.3": 7977.3594, "encoder_q-layer.4": 8616.0723, "encoder_q-layer.5": 8849.7363, "encoder_q-layer.6": 9630.7656, "encoder_q-layer.7": 10986.5635, "encoder_q-layer.8": 12615.709, "encoder_q-layer.9": 11601.6104, "epoch": 0.68, "inbatch_neg_score": 4.7086, "inbatch_pos_score": 5.3281, "learning_rate": 1.1473684210526315e-05, "loss": 3.3633, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.7562, "num_token_overlap": 11.6814, "num_token_query": 31.3581, "num_token_union": 65.1044, "num_word_context": 202.4137, "num_word_doc": 49.8461, "num_word_query": 23.2853, "postclip_grad_norm": 1.0, "preclip_grad_norm": 19476.5671, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.707, "query_norm": 2.4904, "queue_k_norm": 2.5607, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3581, "sent_len_1": 66.7562, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.6962, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 156400 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3593, "doc_norm": 2.5601, "encoder_q-embeddings": 9203.7393, "encoder_q-layer.0": 6456.8882, "encoder_q-layer.1": 6818.6514, "encoder_q-layer.10": 13720.208, "encoder_q-layer.11": 33499.125, "encoder_q-layer.2": 7386.2988, "encoder_q-layer.3": 7558.8506, "encoder_q-layer.4": 8222.8633, "encoder_q-layer.5": 8772.5195, "encoder_q-layer.6": 10036.0088, "encoder_q-layer.7": 11265.4258, "encoder_q-layer.8": 14352.8125, "encoder_q-layer.9": 12952.6191, "epoch": 0.68, "inbatch_neg_score": 4.7102, "inbatch_pos_score": 5.3281, "learning_rate": 1.1447368421052632e-05, "loss": 3.3593, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.8333, "num_token_overlap": 11.664, "num_token_query": 31.3701, "num_token_union": 65.1592, "num_word_context": 202.2811, "num_word_doc": 49.8563, "num_word_query": 23.3072, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20726.7687, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7109, "query_norm": 2.4862, "queue_k_norm": 2.5606, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3701, "sent_len_1": 66.8333, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.5025, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156500 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.3736, "doc_norm": 2.5585, "encoder_q-embeddings": 9080.1436, "encoder_q-layer.0": 6660.1294, "encoder_q-layer.1": 6931.7065, "encoder_q-layer.10": 13342.6787, "encoder_q-layer.11": 32090.4043, "encoder_q-layer.2": 7696.7246, "encoder_q-layer.3": 7936.8364, "encoder_q-layer.4": 8142.3232, "encoder_q-layer.5": 8398.7607, "encoder_q-layer.6": 9489.7432, "encoder_q-layer.7": 10646.3164, "encoder_q-layer.8": 13656.3018, "encoder_q-layer.9": 12599.0947, "epoch": 0.68, "inbatch_neg_score": 4.7142, "inbatch_pos_score": 5.3047, "learning_rate": 1.1421052631578948e-05, "loss": 3.3736, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.7071, "num_token_overlap": 11.6179, "num_token_query": 31.2421, "num_token_union": 65.0094, "num_word_context": 202.2665, "num_word_doc": 49.7535, "num_word_query": 23.1912, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20008.7134, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7148, "query_norm": 2.4866, "queue_k_norm": 2.5618, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2421, "sent_len_1": 66.7071, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.985, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 156600 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.3677, "doc_norm": 2.5608, "encoder_q-embeddings": 9186.5645, "encoder_q-layer.0": 6333.9385, "encoder_q-layer.1": 6743.0991, "encoder_q-layer.10": 13264.7676, "encoder_q-layer.11": 33493.2344, "encoder_q-layer.2": 7451.8896, "encoder_q-layer.3": 7634.3843, "encoder_q-layer.4": 8206.9668, "encoder_q-layer.5": 8730.9541, "encoder_q-layer.6": 9697.7979, "encoder_q-layer.7": 11424.2051, "encoder_q-layer.8": 13633.6621, "encoder_q-layer.9": 13069.7529, "epoch": 0.68, "inbatch_neg_score": 4.715, "inbatch_pos_score": 5.332, "learning_rate": 1.1394736842105263e-05, "loss": 3.3677, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.6614, "num_token_overlap": 11.6714, "num_token_query": 31.4517, "num_token_union": 65.1356, "num_word_context": 202.0591, "num_word_doc": 49.7561, "num_word_query": 23.3419, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20402.9221, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7148, "query_norm": 2.491, "queue_k_norm": 2.5609, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4517, "sent_len_1": 66.6614, "sent_len_max_0": 127.9675, "sent_len_max_1": 206.185, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 156700 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.3746, "doc_norm": 2.5616, "encoder_q-embeddings": 10033.1504, "encoder_q-layer.0": 7055.0249, "encoder_q-layer.1": 7493.4126, "encoder_q-layer.10": 14280.6953, "encoder_q-layer.11": 33778.2969, "encoder_q-layer.2": 8568.3115, "encoder_q-layer.3": 9131.1279, "encoder_q-layer.4": 10055.5449, "encoder_q-layer.5": 10163.1553, "encoder_q-layer.6": 10955.5078, "encoder_q-layer.7": 11898.2314, "encoder_q-layer.8": 13649.2305, "encoder_q-layer.9": 12736.8281, "epoch": 0.68, "inbatch_neg_score": 4.7182, "inbatch_pos_score": 5.3203, "learning_rate": 1.136842105263158e-05, "loss": 3.3746, "norm_diff": 0.0707, "norm_loss": 0.0, "num_token_doc": 66.7703, "num_token_overlap": 11.6784, "num_token_query": 31.3624, "num_token_union": 65.1052, "num_word_context": 202.2551, "num_word_doc": 49.8564, "num_word_query": 23.3071, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21510.0217, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7188, "query_norm": 2.4909, "queue_k_norm": 2.5627, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3624, "sent_len_1": 66.7703, "sent_len_max_0": 127.98, "sent_len_max_1": 207.2188, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3404, "doc_norm": 2.5641, "encoder_q-embeddings": 10165.4238, "encoder_q-layer.0": 7040.4233, "encoder_q-layer.1": 7456.3926, "encoder_q-layer.10": 13856.0605, "encoder_q-layer.11": 34150.3984, "encoder_q-layer.2": 8500.3838, "encoder_q-layer.3": 8782.7197, "encoder_q-layer.4": 9398.8574, "encoder_q-layer.5": 9484.3857, "encoder_q-layer.6": 10608.7148, "encoder_q-layer.7": 13162.3877, "encoder_q-layer.8": 14402.0752, "encoder_q-layer.9": 13205.9385, "epoch": 0.68, "inbatch_neg_score": 4.7219, "inbatch_pos_score": 5.3281, "learning_rate": 1.1342105263157895e-05, "loss": 3.3404, "norm_diff": 0.0737, "norm_loss": 0.0, "num_token_doc": 66.8468, "num_token_overlap": 11.7326, "num_token_query": 31.4577, "num_token_union": 65.1693, "num_word_context": 202.1982, "num_word_doc": 49.8467, "num_word_query": 23.3702, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21573.7423, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7188, "query_norm": 2.4904, "queue_k_norm": 2.5637, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4577, "sent_len_1": 66.8468, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.4087, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 156900 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3569, "doc_norm": 2.5617, "encoder_q-embeddings": 9769.6133, "encoder_q-layer.0": 6679.5957, "encoder_q-layer.1": 7004.1401, "encoder_q-layer.10": 12840.501, "encoder_q-layer.11": 31739.6914, "encoder_q-layer.2": 7761.8877, "encoder_q-layer.3": 7883.7725, "encoder_q-layer.4": 8258.3193, "encoder_q-layer.5": 8445.4658, "encoder_q-layer.6": 9895.5576, "encoder_q-layer.7": 11465.416, "encoder_q-layer.8": 13344.1553, "encoder_q-layer.9": 12376.3848, "epoch": 0.68, "inbatch_neg_score": 4.7235, "inbatch_pos_score": 5.3359, "learning_rate": 1.1315789473684212e-05, "loss": 3.3569, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.749, "num_token_overlap": 11.6947, "num_token_query": 31.388, "num_token_union": 65.077, "num_word_context": 202.2111, "num_word_doc": 49.8145, "num_word_query": 23.3067, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20122.9389, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7227, "query_norm": 2.4911, "queue_k_norm": 2.5639, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.388, "sent_len_1": 66.749, "sent_len_max_0": 127.975, "sent_len_max_1": 208.9112, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157000 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3644, "doc_norm": 2.5645, "encoder_q-embeddings": 9287.8809, "encoder_q-layer.0": 6367.4204, "encoder_q-layer.1": 6936.5986, "encoder_q-layer.10": 13333.8213, "encoder_q-layer.11": 33183.8125, "encoder_q-layer.2": 7684.9429, "encoder_q-layer.3": 7861.9834, "encoder_q-layer.4": 8681.7803, "encoder_q-layer.5": 8876.3916, "encoder_q-layer.6": 9688.6016, "encoder_q-layer.7": 11128.0928, "encoder_q-layer.8": 13693.7559, "encoder_q-layer.9": 12786.043, "epoch": 0.68, "inbatch_neg_score": 4.7254, "inbatch_pos_score": 5.3477, "learning_rate": 1.1289473684210528e-05, "loss": 3.3644, "norm_diff": 0.0725, "norm_loss": 0.0, "num_token_doc": 66.7986, "num_token_overlap": 11.6803, "num_token_query": 31.3483, "num_token_union": 65.119, "num_word_context": 202.5117, "num_word_doc": 49.8342, "num_word_query": 23.2867, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20309.5161, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7266, "query_norm": 2.492, "queue_k_norm": 2.5647, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3483, "sent_len_1": 66.7986, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.0387, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157100 }, { "accuracy": 49.2188, "active_queue_size": 16384.0, "cl_loss": 3.353, "doc_norm": 2.5636, "encoder_q-embeddings": 9981.3506, "encoder_q-layer.0": 7116.21, "encoder_q-layer.1": 7570.7646, "encoder_q-layer.10": 13968.0898, "encoder_q-layer.11": 34150.2188, "encoder_q-layer.2": 8280.6123, "encoder_q-layer.3": 8585.2705, "encoder_q-layer.4": 9147.9678, "encoder_q-layer.5": 9593.9629, "encoder_q-layer.6": 10644.5684, "encoder_q-layer.7": 11153.4463, "encoder_q-layer.8": 13693.8682, "encoder_q-layer.9": 12718.3008, "epoch": 0.68, "inbatch_neg_score": 4.7294, "inbatch_pos_score": 5.3359, "learning_rate": 1.1263157894736842e-05, "loss": 3.353, "norm_diff": 0.0672, "norm_loss": 0.0, "num_token_doc": 66.8878, "num_token_overlap": 11.6889, "num_token_query": 31.327, "num_token_union": 65.1148, "num_word_context": 202.5402, "num_word_doc": 49.9033, "num_word_query": 23.2612, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21415.1052, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7305, "query_norm": 2.4964, "queue_k_norm": 2.5647, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.327, "sent_len_1": 66.8878, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.8462, "stdk": 0.0493, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157200 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3471, "doc_norm": 2.5657, "encoder_q-embeddings": 9442.3438, "encoder_q-layer.0": 6682.1729, "encoder_q-layer.1": 6842.2529, "encoder_q-layer.10": 13258.7383, "encoder_q-layer.11": 33295.3711, "encoder_q-layer.2": 7841.7705, "encoder_q-layer.3": 8023.0938, "encoder_q-layer.4": 8450.4844, "encoder_q-layer.5": 8830.9883, "encoder_q-layer.6": 9909.0635, "encoder_q-layer.7": 11151.3281, "encoder_q-layer.8": 13555.3584, "encoder_q-layer.9": 12597.0732, "epoch": 0.68, "inbatch_neg_score": 4.7347, "inbatch_pos_score": 5.3477, "learning_rate": 1.1236842105263158e-05, "loss": 3.3471, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.6699, "num_token_overlap": 11.7262, "num_token_query": 31.4058, "num_token_union": 65.0539, "num_word_context": 201.8689, "num_word_doc": 49.733, "num_word_query": 23.3207, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20802.4625, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7344, "query_norm": 2.4942, "queue_k_norm": 2.5662, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4058, "sent_len_1": 66.6699, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.4512, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157300 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3488, "doc_norm": 2.5675, "encoder_q-embeddings": 9740.8965, "encoder_q-layer.0": 6976.9854, "encoder_q-layer.1": 7510.9697, "encoder_q-layer.10": 13871.1641, "encoder_q-layer.11": 33186.3047, "encoder_q-layer.2": 8437.2988, "encoder_q-layer.3": 8484.3857, "encoder_q-layer.4": 9692.8691, "encoder_q-layer.5": 9736.0947, "encoder_q-layer.6": 10673.1367, "encoder_q-layer.7": 12638.1396, "encoder_q-layer.8": 14155.5068, "encoder_q-layer.9": 12668.9365, "epoch": 0.68, "inbatch_neg_score": 4.7339, "inbatch_pos_score": 5.3438, "learning_rate": 1.1210526315789475e-05, "loss": 3.3488, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.8726, "num_token_overlap": 11.711, "num_token_query": 31.3492, "num_token_union": 65.1227, "num_word_context": 202.3677, "num_word_doc": 49.8807, "num_word_query": 23.2762, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21346.9145, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7344, "query_norm": 2.492, "queue_k_norm": 2.5662, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3492, "sent_len_1": 66.8726, "sent_len_max_0": 127.995, "sent_len_max_1": 210.2188, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157400 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3565, "doc_norm": 2.5678, "encoder_q-embeddings": 9988.5107, "encoder_q-layer.0": 7028.4795, "encoder_q-layer.1": 7825.1597, "encoder_q-layer.10": 13995.1523, "encoder_q-layer.11": 34086.3359, "encoder_q-layer.2": 9045.0166, "encoder_q-layer.3": 9499.4033, "encoder_q-layer.4": 10511.876, "encoder_q-layer.5": 11074.6191, "encoder_q-layer.6": 12643.0508, "encoder_q-layer.7": 13783.7822, "encoder_q-layer.8": 14801.7686, "encoder_q-layer.9": 13358.6191, "epoch": 0.68, "inbatch_neg_score": 4.739, "inbatch_pos_score": 5.3555, "learning_rate": 1.118421052631579e-05, "loss": 3.3565, "norm_diff": 0.0696, "norm_loss": 0.0, "num_token_doc": 66.7593, "num_token_overlap": 11.6894, "num_token_query": 31.4524, "num_token_union": 65.1754, "num_word_context": 202.2271, "num_word_doc": 49.8069, "num_word_query": 23.3516, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21882.9374, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7383, "query_norm": 2.4982, "queue_k_norm": 2.5666, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4524, "sent_len_1": 66.7593, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.5825, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157500 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3516, "doc_norm": 2.5678, "encoder_q-embeddings": 10946.249, "encoder_q-layer.0": 7780.8145, "encoder_q-layer.1": 8764.1172, "encoder_q-layer.10": 12729.9102, "encoder_q-layer.11": 31034.6934, "encoder_q-layer.2": 9837.7227, "encoder_q-layer.3": 9314.5723, "encoder_q-layer.4": 9681.4375, "encoder_q-layer.5": 9720.8643, "encoder_q-layer.6": 10592.0645, "encoder_q-layer.7": 11129.0234, "encoder_q-layer.8": 13194.9053, "encoder_q-layer.9": 12602.543, "epoch": 0.68, "inbatch_neg_score": 4.7423, "inbatch_pos_score": 5.3633, "learning_rate": 1.1157894736842106e-05, "loss": 3.3516, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.66, "num_token_overlap": 11.7023, "num_token_query": 31.4921, "num_token_union": 65.1089, "num_word_context": 202.3898, "num_word_doc": 49.758, "num_word_query": 23.3924, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20509.0124, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7422, "query_norm": 2.4991, "queue_k_norm": 2.5686, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4921, "sent_len_1": 66.66, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.8025, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 157600 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3662, "doc_norm": 2.5713, "encoder_q-embeddings": 11810.5195, "encoder_q-layer.0": 8147.8823, "encoder_q-layer.1": 8902.6035, "encoder_q-layer.10": 14300.0957, "encoder_q-layer.11": 35684.4883, "encoder_q-layer.2": 10257.6338, "encoder_q-layer.3": 10088.6113, "encoder_q-layer.4": 10689.4932, "encoder_q-layer.5": 10971.5322, "encoder_q-layer.6": 11920.4209, "encoder_q-layer.7": 12692.0244, "encoder_q-layer.8": 14861.6396, "encoder_q-layer.9": 13485.291, "epoch": 0.68, "inbatch_neg_score": 4.7478, "inbatch_pos_score": 5.3516, "learning_rate": 1.1131578947368421e-05, "loss": 3.3662, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.7933, "num_token_overlap": 11.6608, "num_token_query": 31.2845, "num_token_union": 65.0945, "num_word_context": 202.2435, "num_word_doc": 49.806, "num_word_query": 23.2327, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22867.6518, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7461, "query_norm": 2.4961, "queue_k_norm": 2.5686, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2845, "sent_len_1": 66.7933, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.9675, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157700 }, { "accuracy": 53.7109, "active_queue_size": 16384.0, "cl_loss": 3.3458, "doc_norm": 2.5694, "encoder_q-embeddings": 9927.5879, "encoder_q-layer.0": 7020.3682, "encoder_q-layer.1": 7540.3711, "encoder_q-layer.10": 14286.1123, "encoder_q-layer.11": 34020.4922, "encoder_q-layer.2": 8251.5186, "encoder_q-layer.3": 8492.6133, "encoder_q-layer.4": 8941.5771, "encoder_q-layer.5": 8877.5381, "encoder_q-layer.6": 10007.8828, "encoder_q-layer.7": 11431.4482, "encoder_q-layer.8": 13539.6611, "encoder_q-layer.9": 13310.9307, "epoch": 0.68, "inbatch_neg_score": 4.7453, "inbatch_pos_score": 5.3711, "learning_rate": 1.1105263157894738e-05, "loss": 3.3458, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.8417, "num_token_overlap": 11.7012, "num_token_query": 31.3923, "num_token_union": 65.1356, "num_word_context": 202.2207, "num_word_doc": 49.8737, "num_word_query": 23.3003, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20827.0321, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7461, "query_norm": 2.4942, "queue_k_norm": 2.5692, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3923, "sent_len_1": 66.8417, "sent_len_max_0": 128.0, "sent_len_max_1": 207.9975, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157800 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.341, "doc_norm": 2.5697, "encoder_q-embeddings": 9657.3525, "encoder_q-layer.0": 7011.4019, "encoder_q-layer.1": 7420.7407, "encoder_q-layer.10": 14188.6953, "encoder_q-layer.11": 32077.8027, "encoder_q-layer.2": 8409.6113, "encoder_q-layer.3": 8432.4199, "encoder_q-layer.4": 9000.1143, "encoder_q-layer.5": 9506.9062, "encoder_q-layer.6": 10741.8018, "encoder_q-layer.7": 11803.6309, "encoder_q-layer.8": 13784.7578, "encoder_q-layer.9": 12495.3896, "epoch": 0.69, "inbatch_neg_score": 4.7483, "inbatch_pos_score": 5.3711, "learning_rate": 1.1078947368421053e-05, "loss": 3.341, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.8537, "num_token_overlap": 11.6884, "num_token_query": 31.4195, "num_token_union": 65.2089, "num_word_context": 202.3574, "num_word_doc": 49.8937, "num_word_query": 23.3451, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20701.935, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7461, "query_norm": 2.4979, "queue_k_norm": 2.57, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4195, "sent_len_1": 66.8537, "sent_len_max_0": 127.99, "sent_len_max_1": 208.3375, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 157900 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.3612, "doc_norm": 2.5694, "encoder_q-embeddings": 9711.3838, "encoder_q-layer.0": 6810.4868, "encoder_q-layer.1": 7060.7314, "encoder_q-layer.10": 13386.5957, "encoder_q-layer.11": 33170.6992, "encoder_q-layer.2": 7876.8789, "encoder_q-layer.3": 8319.4746, "encoder_q-layer.4": 8652.9062, "encoder_q-layer.5": 9027.2207, "encoder_q-layer.6": 9957.2783, "encoder_q-layer.7": 11152.4473, "encoder_q-layer.8": 13668.54, "encoder_q-layer.9": 12913.0273, "epoch": 0.69, "inbatch_neg_score": 4.7511, "inbatch_pos_score": 5.3633, "learning_rate": 1.1052631578947368e-05, "loss": 3.3612, "norm_diff": 0.0682, "norm_loss": 0.0, "num_token_doc": 66.6181, "num_token_overlap": 11.6213, "num_token_query": 31.238, "num_token_union": 64.9548, "num_word_context": 202.1263, "num_word_doc": 49.7367, "num_word_query": 23.1953, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20519.6611, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.75, "query_norm": 2.5012, "queue_k_norm": 2.5711, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.238, "sent_len_1": 66.6181, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.9238, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 158000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.352, "doc_norm": 2.5705, "encoder_q-embeddings": 9530.8955, "encoder_q-layer.0": 6863.6201, "encoder_q-layer.1": 7073.3032, "encoder_q-layer.10": 14871.5186, "encoder_q-layer.11": 34325.7539, "encoder_q-layer.2": 8066.7393, "encoder_q-layer.3": 8414.2686, "encoder_q-layer.4": 8981.3311, "encoder_q-layer.5": 9097.1973, "encoder_q-layer.6": 10382.7441, "encoder_q-layer.7": 11660.2363, "encoder_q-layer.8": 13735.2939, "encoder_q-layer.9": 12979.5996, "epoch": 0.69, "inbatch_neg_score": 4.7552, "inbatch_pos_score": 5.3633, "learning_rate": 1.1026315789473685e-05, "loss": 3.352, "norm_diff": 0.0712, "norm_loss": 0.0, "num_token_doc": 66.8202, "num_token_overlap": 11.7095, "num_token_query": 31.4263, "num_token_union": 65.1168, "num_word_context": 202.2482, "num_word_doc": 49.839, "num_word_query": 23.3495, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21029.7072, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7539, "query_norm": 2.4992, "queue_k_norm": 2.5699, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4263, "sent_len_1": 66.8202, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.615, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158100 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.358, "doc_norm": 2.5701, "encoder_q-embeddings": 9442.5293, "encoder_q-layer.0": 6698.4517, "encoder_q-layer.1": 7081.0034, "encoder_q-layer.10": 13906.1729, "encoder_q-layer.11": 33515.6172, "encoder_q-layer.2": 8045.248, "encoder_q-layer.3": 8427.1641, "encoder_q-layer.4": 9080.2227, "encoder_q-layer.5": 9910.6719, "encoder_q-layer.6": 10989.6494, "encoder_q-layer.7": 12231.7734, "encoder_q-layer.8": 14154.4414, "encoder_q-layer.9": 12716.6123, "epoch": 0.69, "inbatch_neg_score": 4.7563, "inbatch_pos_score": 5.3633, "learning_rate": 1.1000000000000001e-05, "loss": 3.358, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.6637, "num_token_overlap": 11.6438, "num_token_query": 31.3237, "num_token_union": 65.0696, "num_word_context": 202.2895, "num_word_doc": 49.7311, "num_word_query": 23.2575, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21281.6943, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7578, "query_norm": 2.4984, "queue_k_norm": 2.5717, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3237, "sent_len_1": 66.6637, "sent_len_max_0": 128.0, "sent_len_max_1": 207.2713, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158200 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.352, "doc_norm": 2.5725, "encoder_q-embeddings": 9292.0615, "encoder_q-layer.0": 6442.4209, "encoder_q-layer.1": 6954.4893, "encoder_q-layer.10": 14106.4238, "encoder_q-layer.11": 32804.6406, "encoder_q-layer.2": 7677.5176, "encoder_q-layer.3": 7697.9185, "encoder_q-layer.4": 8295.2832, "encoder_q-layer.5": 8412.7734, "encoder_q-layer.6": 9670.5889, "encoder_q-layer.7": 10840.5547, "encoder_q-layer.8": 12957.9512, "encoder_q-layer.9": 12635.5723, "epoch": 0.69, "inbatch_neg_score": 4.7575, "inbatch_pos_score": 5.3789, "learning_rate": 1.0973684210526316e-05, "loss": 3.352, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.8541, "num_token_overlap": 11.6932, "num_token_query": 31.3784, "num_token_union": 65.1694, "num_word_context": 202.2842, "num_word_doc": 49.8652, "num_word_query": 23.3071, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20390.5215, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7578, "query_norm": 2.5027, "queue_k_norm": 2.572, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3784, "sent_len_1": 66.8541, "sent_len_max_0": 128.0, "sent_len_max_1": 209.7262, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158300 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3478, "doc_norm": 2.5714, "encoder_q-embeddings": 19843.9102, "encoder_q-layer.0": 14086.1543, "encoder_q-layer.1": 15087.667, "encoder_q-layer.10": 26881.0078, "encoder_q-layer.11": 67869.7656, "encoder_q-layer.2": 17039.0605, "encoder_q-layer.3": 17210.541, "encoder_q-layer.4": 18824.2852, "encoder_q-layer.5": 18920.7266, "encoder_q-layer.6": 21601.6777, "encoder_q-layer.7": 23509.9043, "encoder_q-layer.8": 28087.1445, "encoder_q-layer.9": 25106.7012, "epoch": 0.69, "inbatch_neg_score": 4.7618, "inbatch_pos_score": 5.3711, "learning_rate": 1.0947368421052633e-05, "loss": 3.3478, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.8567, "num_token_overlap": 11.6963, "num_token_query": 31.392, "num_token_union": 65.1391, "num_word_context": 202.3744, "num_word_doc": 49.8795, "num_word_query": 23.3064, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42824.1642, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7617, "query_norm": 2.5016, "queue_k_norm": 2.5725, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.392, "sent_len_1": 66.8567, "sent_len_max_0": 127.9163, "sent_len_max_1": 209.3237, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158400 }, { "accuracy": 53.9551, "active_queue_size": 16384.0, "cl_loss": 3.3534, "doc_norm": 2.5745, "encoder_q-embeddings": 19144.0605, "encoder_q-layer.0": 13722.5312, "encoder_q-layer.1": 14623.9775, "encoder_q-layer.10": 25493.5254, "encoder_q-layer.11": 63106.9727, "encoder_q-layer.2": 16393.9766, "encoder_q-layer.3": 16865.1348, "encoder_q-layer.4": 18204.293, "encoder_q-layer.5": 18690.918, "encoder_q-layer.6": 20785.8125, "encoder_q-layer.7": 23954.5625, "encoder_q-layer.8": 27984.3652, "encoder_q-layer.9": 25537.9141, "epoch": 0.69, "inbatch_neg_score": 4.7616, "inbatch_pos_score": 5.3867, "learning_rate": 1.0921052631578948e-05, "loss": 3.3534, "norm_diff": 0.0691, "norm_loss": 0.0, "num_token_doc": 66.7839, "num_token_overlap": 11.6566, "num_token_query": 31.3753, "num_token_union": 65.1478, "num_word_context": 202.2462, "num_word_doc": 49.8356, "num_word_query": 23.3122, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40830.5117, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7617, "query_norm": 2.5054, "queue_k_norm": 2.5731, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3753, "sent_len_1": 66.7839, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.7025, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158500 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3659, "doc_norm": 2.5756, "encoder_q-embeddings": 19211.6172, "encoder_q-layer.0": 13687.6152, "encoder_q-layer.1": 14319.1953, "encoder_q-layer.10": 25920.2383, "encoder_q-layer.11": 64431.7188, "encoder_q-layer.2": 15895.6328, "encoder_q-layer.3": 16599.3965, "encoder_q-layer.4": 17378.8457, "encoder_q-layer.5": 17617.4609, "encoder_q-layer.6": 19700.2402, "encoder_q-layer.7": 22240.8359, "encoder_q-layer.8": 26518.8105, "encoder_q-layer.9": 24961.9922, "epoch": 0.69, "inbatch_neg_score": 4.7639, "inbatch_pos_score": 5.3828, "learning_rate": 1.0894736842105263e-05, "loss": 3.3659, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.7725, "num_token_overlap": 11.6483, "num_token_query": 31.2879, "num_token_union": 65.1039, "num_word_context": 202.2486, "num_word_doc": 49.8133, "num_word_query": 23.2238, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40336.8687, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7656, "query_norm": 2.5011, "queue_k_norm": 2.5739, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2879, "sent_len_1": 66.7725, "sent_len_max_0": 127.995, "sent_len_max_1": 207.265, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158600 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3581, "doc_norm": 2.5751, "encoder_q-embeddings": 19942.5078, "encoder_q-layer.0": 13417.8789, "encoder_q-layer.1": 14523.3516, "encoder_q-layer.10": 30025.9395, "encoder_q-layer.11": 67093.3516, "encoder_q-layer.2": 16318.957, "encoder_q-layer.3": 17390.5039, "encoder_q-layer.4": 18265.0078, "encoder_q-layer.5": 18712.9648, "encoder_q-layer.6": 20202.0684, "encoder_q-layer.7": 23854.8203, "encoder_q-layer.8": 27140.123, "encoder_q-layer.9": 25452.3711, "epoch": 0.69, "inbatch_neg_score": 4.7662, "inbatch_pos_score": 5.3711, "learning_rate": 1.086842105263158e-05, "loss": 3.3581, "norm_diff": 0.0702, "norm_loss": 0.0, "num_token_doc": 66.7748, "num_token_overlap": 11.6818, "num_token_query": 31.4188, "num_token_union": 65.1372, "num_word_context": 202.2982, "num_word_doc": 49.8653, "num_word_query": 23.3555, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42367.6869, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7656, "query_norm": 2.5049, "queue_k_norm": 2.5742, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4188, "sent_len_1": 66.7748, "sent_len_max_0": 127.9875, "sent_len_max_1": 205.2337, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158700 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3518, "doc_norm": 2.5741, "encoder_q-embeddings": 20591.791, "encoder_q-layer.0": 14712.8193, "encoder_q-layer.1": 15215.6035, "encoder_q-layer.10": 27011.1523, "encoder_q-layer.11": 65913.3125, "encoder_q-layer.2": 18190.4199, "encoder_q-layer.3": 18608.6309, "encoder_q-layer.4": 19814.1992, "encoder_q-layer.5": 19718.7988, "encoder_q-layer.6": 22231.4004, "encoder_q-layer.7": 24921.8594, "encoder_q-layer.8": 29136.1426, "encoder_q-layer.9": 25922.9805, "epoch": 0.69, "inbatch_neg_score": 4.7691, "inbatch_pos_score": 5.3906, "learning_rate": 1.0842105263157895e-05, "loss": 3.3518, "norm_diff": 0.0684, "norm_loss": 0.0, "num_token_doc": 66.8885, "num_token_overlap": 11.6916, "num_token_query": 31.379, "num_token_union": 65.2009, "num_word_context": 202.4489, "num_word_doc": 49.9303, "num_word_query": 23.3033, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42775.0265, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7695, "query_norm": 2.5057, "queue_k_norm": 2.5753, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.379, "sent_len_1": 66.8885, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.615, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158800 }, { "accuracy": 53.6133, "active_queue_size": 16384.0, "cl_loss": 3.3496, "doc_norm": 2.5761, "encoder_q-embeddings": 18356.0547, "encoder_q-layer.0": 12687.6543, "encoder_q-layer.1": 13721.835, "encoder_q-layer.10": 25888.5566, "encoder_q-layer.11": 64083.1406, "encoder_q-layer.2": 15371.1055, "encoder_q-layer.3": 15292.9209, "encoder_q-layer.4": 16384.0781, "encoder_q-layer.5": 16867.8945, "encoder_q-layer.6": 18292.8164, "encoder_q-layer.7": 21385.1055, "encoder_q-layer.8": 26106.7988, "encoder_q-layer.9": 24024.0859, "epoch": 0.69, "inbatch_neg_score": 4.773, "inbatch_pos_score": 5.4023, "learning_rate": 1.0815789473684211e-05, "loss": 3.3496, "norm_diff": 0.0699, "norm_loss": 0.0, "num_token_doc": 66.7782, "num_token_overlap": 11.6784, "num_token_query": 31.4399, "num_token_union": 65.1796, "num_word_context": 202.2082, "num_word_doc": 49.8474, "num_word_query": 23.3554, "postclip_grad_norm": 1.0, "preclip_grad_norm": 39730.5325, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7734, "query_norm": 2.5062, "queue_k_norm": 2.5757, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4399, "sent_len_1": 66.7782, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.7212, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 158900 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3514, "doc_norm": 2.5748, "encoder_q-embeddings": 20190.2969, "encoder_q-layer.0": 13177.4727, "encoder_q-layer.1": 13840.3115, "encoder_q-layer.10": 30976.123, "encoder_q-layer.11": 69128.8984, "encoder_q-layer.2": 15610.2939, "encoder_q-layer.3": 15912.1621, "encoder_q-layer.4": 17128.4941, "encoder_q-layer.5": 17296.1992, "encoder_q-layer.6": 20166.3184, "encoder_q-layer.7": 23385.7891, "encoder_q-layer.8": 28969.5312, "encoder_q-layer.9": 27341.9941, "epoch": 0.69, "inbatch_neg_score": 4.7772, "inbatch_pos_score": 5.3906, "learning_rate": 1.0789473684210526e-05, "loss": 3.3514, "norm_diff": 0.0636, "norm_loss": 0.0, "num_token_doc": 66.7596, "num_token_overlap": 11.6843, "num_token_query": 31.3831, "num_token_union": 65.1184, "num_word_context": 202.1986, "num_word_doc": 49.8258, "num_word_query": 23.3114, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42217.7809, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7773, "query_norm": 2.5113, "queue_k_norm": 2.5762, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3831, "sent_len_1": 66.7596, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.465, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 159000 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3526, "doc_norm": 2.5767, "encoder_q-embeddings": 19087.9023, "encoder_q-layer.0": 13276.0947, "encoder_q-layer.1": 14198.0625, "encoder_q-layer.10": 27072.9863, "encoder_q-layer.11": 66485.6094, "encoder_q-layer.2": 15914.7031, "encoder_q-layer.3": 16320.0137, "encoder_q-layer.4": 17487.0859, "encoder_q-layer.5": 18049.9512, "encoder_q-layer.6": 20198.3301, "encoder_q-layer.7": 23994.1172, "encoder_q-layer.8": 28974.0762, "encoder_q-layer.9": 26071.1172, "epoch": 0.69, "inbatch_neg_score": 4.7861, "inbatch_pos_score": 5.3906, "learning_rate": 1.0763157894736843e-05, "loss": 3.3526, "norm_diff": 0.0658, "norm_loss": 0.0, "num_token_doc": 66.8665, "num_token_overlap": 11.6958, "num_token_query": 31.3577, "num_token_union": 65.1668, "num_word_context": 202.5195, "num_word_doc": 49.9037, "num_word_query": 23.2964, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41742.7686, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7852, "query_norm": 2.5109, "queue_k_norm": 2.5768, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3577, "sent_len_1": 66.8665, "sent_len_max_0": 127.9488, "sent_len_max_1": 207.3413, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 159100 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3469, "doc_norm": 2.5775, "encoder_q-embeddings": 19418.0195, "encoder_q-layer.0": 13473.9443, "encoder_q-layer.1": 14535.082, "encoder_q-layer.10": 26347.4551, "encoder_q-layer.11": 63580.1875, "encoder_q-layer.2": 16518.9375, "encoder_q-layer.3": 17171.9297, "encoder_q-layer.4": 18132.5977, "encoder_q-layer.5": 18686.1797, "encoder_q-layer.6": 20643.5957, "encoder_q-layer.7": 22895.2578, "encoder_q-layer.8": 27147.2793, "encoder_q-layer.9": 24958.8516, "epoch": 0.69, "inbatch_neg_score": 4.7895, "inbatch_pos_score": 5.4102, "learning_rate": 1.073684210526316e-05, "loss": 3.3469, "norm_diff": 0.0681, "norm_loss": 0.0, "num_token_doc": 66.8116, "num_token_overlap": 11.7002, "num_token_query": 31.4209, "num_token_union": 65.1503, "num_word_context": 202.2261, "num_word_doc": 49.8546, "num_word_query": 23.3642, "postclip_grad_norm": 1.0, "preclip_grad_norm": 40838.0277, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.7891, "query_norm": 2.5095, "queue_k_norm": 2.5784, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4209, "sent_len_1": 66.8116, "sent_len_max_0": 127.9638, "sent_len_max_1": 207.455, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 159200 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3778, "doc_norm": 2.5765, "encoder_q-embeddings": 9178.9531, "encoder_q-layer.0": 6515.249, "encoder_q-layer.1": 6950.3682, "encoder_q-layer.10": 13861.7061, "encoder_q-layer.11": 33318.9336, "encoder_q-layer.2": 7921.5527, "encoder_q-layer.3": 8128.0571, "encoder_q-layer.4": 8662.041, "encoder_q-layer.5": 9091.9209, "encoder_q-layer.6": 10538.1201, "encoder_q-layer.7": 11670.0957, "encoder_q-layer.8": 13892.7666, "encoder_q-layer.9": 13127.2197, "epoch": 0.69, "inbatch_neg_score": 4.7975, "inbatch_pos_score": 5.4102, "learning_rate": 1.0710526315789473e-05, "loss": 3.3778, "norm_diff": 0.0689, "norm_loss": 0.0, "num_token_doc": 66.7194, "num_token_overlap": 11.635, "num_token_query": 31.2467, "num_token_union": 65.0418, "num_word_context": 202.2799, "num_word_doc": 49.8207, "num_word_query": 23.1927, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20953.3889, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.7969, "query_norm": 2.5075, "queue_k_norm": 2.578, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2467, "sent_len_1": 66.7194, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.315, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 159300 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3629, "doc_norm": 2.5779, "encoder_q-embeddings": 5693.8872, "encoder_q-layer.0": 4169.0244, "encoder_q-layer.1": 4641.374, "encoder_q-layer.10": 6440.2979, "encoder_q-layer.11": 16645.2949, "encoder_q-layer.2": 5525.2549, "encoder_q-layer.3": 5423.7739, "encoder_q-layer.4": 5510.7266, "encoder_q-layer.5": 5508.5215, "encoder_q-layer.6": 5829.2212, "encoder_q-layer.7": 6032.1445, "encoder_q-layer.8": 6565.9194, "encoder_q-layer.9": 6103.7769, "epoch": 0.69, "inbatch_neg_score": 4.8039, "inbatch_pos_score": 5.4023, "learning_rate": 1.068421052631579e-05, "loss": 3.3629, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.6343, "num_token_overlap": 11.6995, "num_token_query": 31.3424, "num_token_union": 65.0021, "num_word_context": 202.2225, "num_word_doc": 49.7184, "num_word_query": 23.2654, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10980.5538, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8047, "query_norm": 2.5087, "queue_k_norm": 2.5802, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3424, "sent_len_1": 66.6343, "sent_len_max_0": 127.9525, "sent_len_max_1": 208.1163, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 159400 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3588, "doc_norm": 2.5809, "encoder_q-embeddings": 4742.0557, "encoder_q-layer.0": 3390.7593, "encoder_q-layer.1": 3546.7205, "encoder_q-layer.10": 6804.6968, "encoder_q-layer.11": 16284.6211, "encoder_q-layer.2": 3968.7251, "encoder_q-layer.3": 3994.7024, "encoder_q-layer.4": 4313.6064, "encoder_q-layer.5": 4484.1851, "encoder_q-layer.6": 4977.3604, "encoder_q-layer.7": 5554.6943, "encoder_q-layer.8": 6370.2275, "encoder_q-layer.9": 6344.1812, "epoch": 0.69, "inbatch_neg_score": 4.8103, "inbatch_pos_score": 5.4297, "learning_rate": 1.0657894736842106e-05, "loss": 3.3588, "norm_diff": 0.0645, "norm_loss": 0.0, "num_token_doc": 66.7862, "num_token_overlap": 11.6846, "num_token_query": 31.4004, "num_token_union": 65.1474, "num_word_context": 202.241, "num_word_doc": 49.8362, "num_word_query": 23.3177, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10256.8751, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8125, "query_norm": 2.5164, "queue_k_norm": 2.579, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4004, "sent_len_1": 66.7862, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.2225, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 159500 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3512, "doc_norm": 2.5811, "encoder_q-embeddings": 4849.7446, "encoder_q-layer.0": 3394.7222, "encoder_q-layer.1": 3658.7402, "encoder_q-layer.10": 7060.0801, "encoder_q-layer.11": 17223.2754, "encoder_q-layer.2": 4196.5547, "encoder_q-layer.3": 4221.8545, "encoder_q-layer.4": 4611.4321, "encoder_q-layer.5": 4868.4956, "encoder_q-layer.6": 5436.7261, "encoder_q-layer.7": 6239.3477, "encoder_q-layer.8": 7467.0117, "encoder_q-layer.9": 6545.5791, "epoch": 0.69, "inbatch_neg_score": 4.8172, "inbatch_pos_score": 5.4258, "learning_rate": 1.0631578947368421e-05, "loss": 3.3512, "norm_diff": 0.0647, "norm_loss": 0.0, "num_token_doc": 66.7287, "num_token_overlap": 11.6778, "num_token_query": 31.3403, "num_token_union": 65.0734, "num_word_context": 202.1905, "num_word_doc": 49.791, "num_word_query": 23.2724, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11073.5215, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8164, "query_norm": 2.5164, "queue_k_norm": 2.5818, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3403, "sent_len_1": 66.7287, "sent_len_max_0": 127.985, "sent_len_max_1": 207.5275, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 159600 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3282, "doc_norm": 2.58, "encoder_q-embeddings": 4937.7993, "encoder_q-layer.0": 3442.0593, "encoder_q-layer.1": 3743.0901, "encoder_q-layer.10": 6202.2896, "encoder_q-layer.11": 16062.2588, "encoder_q-layer.2": 4329.6797, "encoder_q-layer.3": 4341.9146, "encoder_q-layer.4": 4593.1255, "encoder_q-layer.5": 4854.1836, "encoder_q-layer.6": 5225.8901, "encoder_q-layer.7": 5896.5078, "encoder_q-layer.8": 6804.0947, "encoder_q-layer.9": 6255.1699, "epoch": 0.69, "inbatch_neg_score": 4.8213, "inbatch_pos_score": 5.4375, "learning_rate": 1.0605263157894738e-05, "loss": 3.3282, "norm_diff": 0.0626, "norm_loss": 0.0, "num_token_doc": 66.7842, "num_token_overlap": 11.7235, "num_token_query": 31.4264, "num_token_union": 65.1207, "num_word_context": 201.9075, "num_word_doc": 49.7934, "num_word_query": 23.3462, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10284.4155, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8203, "query_norm": 2.5175, "queue_k_norm": 2.5809, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4264, "sent_len_1": 66.7842, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.6587, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 159700 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.356, "doc_norm": 2.5844, "encoder_q-embeddings": 4791.2773, "encoder_q-layer.0": 3302.478, "encoder_q-layer.1": 3469.522, "encoder_q-layer.10": 7977.4399, "encoder_q-layer.11": 17368.3242, "encoder_q-layer.2": 3901.0083, "encoder_q-layer.3": 4131.6606, "encoder_q-layer.4": 4497.8472, "encoder_q-layer.5": 4622.0234, "encoder_q-layer.6": 5241.9165, "encoder_q-layer.7": 5989.6309, "encoder_q-layer.8": 7143.5737, "encoder_q-layer.9": 6710.9126, "epoch": 0.69, "inbatch_neg_score": 4.8283, "inbatch_pos_score": 5.4492, "learning_rate": 1.0578947368421053e-05, "loss": 3.356, "norm_diff": 0.0622, "norm_loss": 0.0, "num_token_doc": 66.8454, "num_token_overlap": 11.6855, "num_token_query": 31.3479, "num_token_union": 65.1542, "num_word_context": 202.3598, "num_word_doc": 49.8666, "num_word_query": 23.2797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10843.399, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8281, "query_norm": 2.5223, "queue_k_norm": 2.5837, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3479, "sent_len_1": 66.8454, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.5175, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 159800 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.356, "doc_norm": 2.5834, "encoder_q-embeddings": 4813.5381, "encoder_q-layer.0": 3322.7844, "encoder_q-layer.1": 3536.5256, "encoder_q-layer.10": 6649.7773, "encoder_q-layer.11": 17256.2988, "encoder_q-layer.2": 3991.8818, "encoder_q-layer.3": 4114.7178, "encoder_q-layer.4": 4520.729, "encoder_q-layer.5": 4596.957, "encoder_q-layer.6": 5054.8936, "encoder_q-layer.7": 5641.9722, "encoder_q-layer.8": 7081.6284, "encoder_q-layer.9": 6462.8096, "epoch": 0.69, "inbatch_neg_score": 4.8331, "inbatch_pos_score": 5.4414, "learning_rate": 1.055263157894737e-05, "loss": 3.356, "norm_diff": 0.0641, "norm_loss": 0.0, "num_token_doc": 66.7925, "num_token_overlap": 11.698, "num_token_query": 31.4149, "num_token_union": 65.1344, "num_word_context": 202.3472, "num_word_doc": 49.8374, "num_word_query": 23.3413, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10661.4451, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8359, "query_norm": 2.5192, "queue_k_norm": 2.5848, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4149, "sent_len_1": 66.7925, "sent_len_max_0": 127.995, "sent_len_max_1": 208.295, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 159900 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3456, "doc_norm": 2.5856, "encoder_q-embeddings": 4665.2275, "encoder_q-layer.0": 3261.6602, "encoder_q-layer.1": 3431.8198, "encoder_q-layer.10": 7644.0503, "encoder_q-layer.11": 17024.7793, "encoder_q-layer.2": 3835.0513, "encoder_q-layer.3": 4108.27, "encoder_q-layer.4": 4442.6172, "encoder_q-layer.5": 4401.3975, "encoder_q-layer.6": 4901.7324, "encoder_q-layer.7": 5856.0615, "encoder_q-layer.8": 7060.6504, "encoder_q-layer.9": 6544.8984, "epoch": 0.69, "inbatch_neg_score": 4.8396, "inbatch_pos_score": 5.4609, "learning_rate": 1.0526315789473684e-05, "loss": 3.3456, "norm_diff": 0.0702, "norm_loss": 0.0, "num_token_doc": 66.8537, "num_token_overlap": 11.6933, "num_token_query": 31.368, "num_token_union": 65.1281, "num_word_context": 201.932, "num_word_doc": 49.8742, "num_word_query": 23.2992, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10646.5838, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8398, "query_norm": 2.5154, "queue_k_norm": 2.5837, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.368, "sent_len_1": 66.8537, "sent_len_max_0": 127.9638, "sent_len_max_1": 210.725, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160000 }, { "dev_runtime": 29.3181, "dev_samples_per_second": 1.091, "dev_steps_per_second": 0.034, "epoch": 0.69, "step": 160000, "test_accuracy": 94.189453125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3224518299102783, "test_doc_norm": 2.570589542388916, "test_inbatch_neg_score": 5.137938499450684, "test_inbatch_pos_score": 6.1638922691345215, "test_loss": 0.3224518299102783, "test_loss_align": 1.0055606365203857, "test_loss_unif": -41.18000793457031, "test_loss_unif_q@queue": -41.18000793457031, "test_norm_diff": 0.020814530551433563, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.839037895202637, "test_query_norm": 2.5914039611816406, "test_queue_k_norm": 2.5835928916931152, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044252026826143265, "test_stdq": 0.04398474842309952, "test_stdqueue_k": 0.049492381513118744, "test_stdqueue_q": 0.0 }, { "dev_runtime": 29.3181, "dev_samples_per_second": 1.091, "dev_steps_per_second": 0.034, "epoch": 0.69, "eval_beir-arguana_ndcg@10": 0.41198, "eval_beir-arguana_recall@10": 0.69346, "eval_beir-arguana_recall@100": 0.95021, "eval_beir-arguana_recall@20": 0.82432, "eval_beir-avg_ndcg@10": 0.40711733333333344, "eval_beir-avg_recall@10": 0.47762674999999993, "eval_beir-avg_recall@100": 0.6560624166666666, "eval_beir-avg_recall@20": 0.539704, "eval_beir-cqadupstack_ndcg@10": 0.2955333333333333, "eval_beir-cqadupstack_recall@10": 0.3947274999999999, "eval_beir-cqadupstack_recall@100": 0.6315241666666666, "eval_beir-cqadupstack_recall@20": 0.4654000000000001, "eval_beir-fiqa_ndcg@10": 0.27305, "eval_beir-fiqa_recall@10": 0.33971, "eval_beir-fiqa_recall@100": 0.61858, "eval_beir-fiqa_recall@20": 0.4194, "eval_beir-nfcorpus_ndcg@10": 0.32773, "eval_beir-nfcorpus_recall@10": 0.16206, "eval_beir-nfcorpus_recall@100": 0.3166, "eval_beir-nfcorpus_recall@20": 0.20502, "eval_beir-nq_ndcg@10": 0.3306, "eval_beir-nq_recall@10": 0.52197, "eval_beir-nq_recall@100": 0.84004, "eval_beir-nq_recall@20": 0.63386, "eval_beir-quora_ndcg@10": 0.79245, "eval_beir-quora_recall@10": 0.89779, "eval_beir-quora_recall@100": 0.98074, "eval_beir-quora_recall@20": 0.9372, "eval_beir-scidocs_ndcg@10": 0.1669, "eval_beir-scidocs_recall@10": 0.17652, "eval_beir-scidocs_recall@100": 0.39182, "eval_beir-scidocs_recall@20": 0.23485, "eval_beir-scifact_ndcg@10": 0.68079, "eval_beir-scifact_recall@10": 0.81233, "eval_beir-scifact_recall@100": 0.92656, "eval_beir-scifact_recall@20": 0.85078, "eval_beir-trec-covid_ndcg@10": 0.61168, "eval_beir-trec-covid_recall@10": 0.648, "eval_beir-trec-covid_recall@100": 0.4694, "eval_beir-trec-covid_recall@20": 0.615, "eval_beir-webis-touche2020_ndcg@10": 0.18046, "eval_beir-webis-touche2020_recall@10": 0.1297, "eval_beir-webis-touche2020_recall@100": 0.43515, "eval_beir-webis-touche2020_recall@20": 0.21121, "eval_senteval-avg_sts": 0.7335002503746486, "eval_senteval-sickr_spearman": 0.727487035544493, "eval_senteval-stsb_spearman": 0.7395134652048041, "step": 160000, "test_accuracy": 94.189453125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3224518299102783, "test_doc_norm": 2.570589542388916, "test_inbatch_neg_score": 5.137938499450684, "test_inbatch_pos_score": 6.1638922691345215, "test_loss": 0.3224518299102783, "test_loss_align": 1.0055606365203857, "test_loss_unif": -41.18000793457031, "test_loss_unif_q@queue": -41.18000793457031, "test_norm_diff": 0.020814530551433563, "test_norm_loss": 0.0, "test_q@queue_neg_score": 4.839037895202637, "test_query_norm": 2.5914039611816406, "test_queue_k_norm": 2.5835928916931152, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044252026826143265, "test_stdq": 0.04398474842309952, "test_stdqueue_k": 0.049492381513118744, "test_stdqueue_q": 0.0 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3582, "doc_norm": 2.5846, "encoder_q-embeddings": 4944.1455, "encoder_q-layer.0": 3451.821, "encoder_q-layer.1": 3497.0227, "encoder_q-layer.10": 6678.418, "encoder_q-layer.11": 16317.9082, "encoder_q-layer.2": 3906.6785, "encoder_q-layer.3": 3925.1475, "encoder_q-layer.4": 4183.4272, "encoder_q-layer.5": 4235.0293, "encoder_q-layer.6": 4755.2949, "encoder_q-layer.7": 5430.6733, "encoder_q-layer.8": 6737.1558, "encoder_q-layer.9": 6202.2681, "epoch": 0.69, "inbatch_neg_score": 4.8414, "inbatch_pos_score": 5.4492, "learning_rate": 1.05e-05, "loss": 3.3582, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.8341, "num_token_overlap": 11.6683, "num_token_query": 31.1665, "num_token_union": 65.0138, "num_word_context": 202.1954, "num_word_doc": 49.8432, "num_word_query": 23.1464, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10114.4391, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8398, "query_norm": 2.513, "queue_k_norm": 2.5856, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.1665, "sent_len_1": 66.8341, "sent_len_max_0": 127.9562, "sent_len_max_1": 208.7512, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 160100 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3561, "doc_norm": 2.5881, "encoder_q-embeddings": 4517.3525, "encoder_q-layer.0": 3227.9802, "encoder_q-layer.1": 3460.7402, "encoder_q-layer.10": 6637.6611, "encoder_q-layer.11": 15985.752, "encoder_q-layer.2": 3967.6096, "encoder_q-layer.3": 3937.5603, "encoder_q-layer.4": 4095.4861, "encoder_q-layer.5": 4146.6821, "encoder_q-layer.6": 4614.5591, "encoder_q-layer.7": 5318.7881, "encoder_q-layer.8": 6531.0732, "encoder_q-layer.9": 6263.9316, "epoch": 0.7, "inbatch_neg_score": 4.8446, "inbatch_pos_score": 5.4648, "learning_rate": 1.0473684210526316e-05, "loss": 3.3561, "norm_diff": 0.0632, "norm_loss": 0.0, "num_token_doc": 66.8645, "num_token_overlap": 11.6805, "num_token_query": 31.4475, "num_token_union": 65.2195, "num_word_context": 202.6382, "num_word_doc": 49.8804, "num_word_query": 23.3589, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9946.4808, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8438, "query_norm": 2.5249, "queue_k_norm": 2.5856, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4475, "sent_len_1": 66.8645, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.9425, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160200 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3499, "doc_norm": 2.5843, "encoder_q-embeddings": 4665.9434, "encoder_q-layer.0": 3344.4861, "encoder_q-layer.1": 3535.6597, "encoder_q-layer.10": 7626.8691, "encoder_q-layer.11": 17492.0195, "encoder_q-layer.2": 4065.3374, "encoder_q-layer.3": 4154.4551, "encoder_q-layer.4": 4438.1304, "encoder_q-layer.5": 4528.3813, "encoder_q-layer.6": 5171.2593, "encoder_q-layer.7": 5839.5659, "encoder_q-layer.8": 7435.6387, "encoder_q-layer.9": 7066.1187, "epoch": 0.7, "inbatch_neg_score": 4.846, "inbatch_pos_score": 5.457, "learning_rate": 1.0447368421052633e-05, "loss": 3.3499, "norm_diff": 0.0643, "norm_loss": 0.0, "num_token_doc": 66.7975, "num_token_overlap": 11.7232, "num_token_query": 31.514, "num_token_union": 65.1787, "num_word_context": 202.5386, "num_word_doc": 49.8293, "num_word_query": 23.4101, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10745.3747, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8477, "query_norm": 2.52, "queue_k_norm": 2.586, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.514, "sent_len_1": 66.7975, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.535, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160300 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3531, "doc_norm": 2.5873, "encoder_q-embeddings": 4694.873, "encoder_q-layer.0": 3306.3867, "encoder_q-layer.1": 3507.2021, "encoder_q-layer.10": 6411.3652, "encoder_q-layer.11": 16434.2793, "encoder_q-layer.2": 3962.4746, "encoder_q-layer.3": 3992.0024, "encoder_q-layer.4": 4222.749, "encoder_q-layer.5": 4255.9897, "encoder_q-layer.6": 4769.1348, "encoder_q-layer.7": 5579.2158, "encoder_q-layer.8": 6701.127, "encoder_q-layer.9": 6261.1499, "epoch": 0.7, "inbatch_neg_score": 4.8499, "inbatch_pos_score": 5.457, "learning_rate": 1.0421052631578948e-05, "loss": 3.3531, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.6729, "num_token_overlap": 11.6961, "num_token_query": 31.4496, "num_token_union": 65.0922, "num_word_context": 202.1332, "num_word_doc": 49.7544, "num_word_query": 23.3681, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10362.4925, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8516, "query_norm": 2.5133, "queue_k_norm": 2.5868, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4496, "sent_len_1": 66.6729, "sent_len_max_0": 127.9475, "sent_len_max_1": 207.8537, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160400 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3496, "doc_norm": 2.5915, "encoder_q-embeddings": 5136.2007, "encoder_q-layer.0": 3606.5251, "encoder_q-layer.1": 3715.7495, "encoder_q-layer.10": 6870.9541, "encoder_q-layer.11": 17032.6523, "encoder_q-layer.2": 4207.062, "encoder_q-layer.3": 4400.6626, "encoder_q-layer.4": 4755.7832, "encoder_q-layer.5": 4913.0708, "encoder_q-layer.6": 5370.3423, "encoder_q-layer.7": 6191.1904, "encoder_q-layer.8": 7129.0688, "encoder_q-layer.9": 6529.3418, "epoch": 0.7, "inbatch_neg_score": 4.85, "inbatch_pos_score": 5.4492, "learning_rate": 1.0394736842105264e-05, "loss": 3.3496, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.8073, "num_token_overlap": 11.6468, "num_token_query": 31.3844, "num_token_union": 65.1787, "num_word_context": 202.4413, "num_word_doc": 49.8454, "num_word_query": 23.3051, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10834.5571, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8516, "query_norm": 2.5092, "queue_k_norm": 2.5889, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3844, "sent_len_1": 66.8073, "sent_len_max_0": 127.9725, "sent_len_max_1": 210.0725, "stdk": 0.0497, "stdq": 0.0426, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 160500 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3529, "doc_norm": 2.5891, "encoder_q-embeddings": 5326.0464, "encoder_q-layer.0": 3568.687, "encoder_q-layer.1": 3903.1089, "encoder_q-layer.10": 6775.0601, "encoder_q-layer.11": 16252.3711, "encoder_q-layer.2": 4476.3843, "encoder_q-layer.3": 4330.9434, "encoder_q-layer.4": 4523.3301, "encoder_q-layer.5": 4627.8271, "encoder_q-layer.6": 5172.1128, "encoder_q-layer.7": 5843.7188, "encoder_q-layer.8": 6859.6211, "encoder_q-layer.9": 6331.6045, "epoch": 0.7, "inbatch_neg_score": 4.8518, "inbatch_pos_score": 5.4648, "learning_rate": 1.036842105263158e-05, "loss": 3.3529, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.7652, "num_token_overlap": 11.6746, "num_token_query": 31.3163, "num_token_union": 65.1138, "num_word_context": 202.1889, "num_word_doc": 49.8228, "num_word_query": 23.2428, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10524.3734, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8516, "query_norm": 2.5176, "queue_k_norm": 2.5892, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3163, "sent_len_1": 66.7652, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.6125, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160600 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3459, "doc_norm": 2.5919, "encoder_q-embeddings": 5179.3843, "encoder_q-layer.0": 3681.502, "encoder_q-layer.1": 4143.7319, "encoder_q-layer.10": 7005.4976, "encoder_q-layer.11": 16546.3008, "encoder_q-layer.2": 4645.2539, "encoder_q-layer.3": 4880.8218, "encoder_q-layer.4": 5445.2832, "encoder_q-layer.5": 5891.2573, "encoder_q-layer.6": 6328.6143, "encoder_q-layer.7": 6912.2461, "encoder_q-layer.8": 7868.2383, "encoder_q-layer.9": 6490.9609, "epoch": 0.7, "inbatch_neg_score": 4.8518, "inbatch_pos_score": 5.4766, "learning_rate": 1.0342105263157894e-05, "loss": 3.3459, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.8421, "num_token_overlap": 11.7068, "num_token_query": 31.4824, "num_token_union": 65.2152, "num_word_context": 202.4169, "num_word_doc": 49.8591, "num_word_query": 23.3776, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11077.3068, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8516, "query_norm": 2.5168, "queue_k_norm": 2.59, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4824, "sent_len_1": 66.8421, "sent_len_max_0": 127.985, "sent_len_max_1": 208.44, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 160700 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3574, "doc_norm": 2.5899, "encoder_q-embeddings": 4850.4209, "encoder_q-layer.0": 3321.1475, "encoder_q-layer.1": 3537.2026, "encoder_q-layer.10": 6836.6675, "encoder_q-layer.11": 16682.4434, "encoder_q-layer.2": 3942.2664, "encoder_q-layer.3": 4053.8933, "encoder_q-layer.4": 4375.7773, "encoder_q-layer.5": 4503.8218, "encoder_q-layer.6": 4947.8433, "encoder_q-layer.7": 5979.585, "encoder_q-layer.8": 7211.3291, "encoder_q-layer.9": 6652.2729, "epoch": 0.7, "inbatch_neg_score": 4.8536, "inbatch_pos_score": 5.4688, "learning_rate": 1.0315789473684211e-05, "loss": 3.3574, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.8843, "num_token_overlap": 11.6949, "num_token_query": 31.3395, "num_token_union": 65.1506, "num_word_context": 202.3726, "num_word_doc": 49.9129, "num_word_query": 23.2707, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10647.4411, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8516, "query_norm": 2.5196, "queue_k_norm": 2.5918, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3395, "sent_len_1": 66.8843, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.645, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 160800 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3454, "doc_norm": 2.5892, "encoder_q-embeddings": 4690.9443, "encoder_q-layer.0": 3271.7166, "encoder_q-layer.1": 3484.0632, "encoder_q-layer.10": 6747.5527, "encoder_q-layer.11": 16577.4531, "encoder_q-layer.2": 3915.9211, "encoder_q-layer.3": 4076.3635, "encoder_q-layer.4": 4270.4487, "encoder_q-layer.5": 4355.9097, "encoder_q-layer.6": 4756.4121, "encoder_q-layer.7": 5522.3452, "encoder_q-layer.8": 6850.5107, "encoder_q-layer.9": 6364.3398, "epoch": 0.7, "inbatch_neg_score": 4.8552, "inbatch_pos_score": 5.457, "learning_rate": 1.0289473684210526e-05, "loss": 3.3454, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.8698, "num_token_overlap": 11.7118, "num_token_query": 31.4376, "num_token_union": 65.1985, "num_word_context": 202.6014, "num_word_doc": 49.9029, "num_word_query": 23.3449, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10361.4009, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8555, "query_norm": 2.515, "queue_k_norm": 2.5908, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4376, "sent_len_1": 66.8698, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.8862, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 160900 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3498, "doc_norm": 2.5896, "encoder_q-embeddings": 4688.1367, "encoder_q-layer.0": 3206.0251, "encoder_q-layer.1": 3474.7336, "encoder_q-layer.10": 7103.5293, "encoder_q-layer.11": 16657.6934, "encoder_q-layer.2": 3836.4158, "encoder_q-layer.3": 3842.3323, "encoder_q-layer.4": 4222.6631, "encoder_q-layer.5": 4369.2783, "encoder_q-layer.6": 5018.166, "encoder_q-layer.7": 5795.8628, "encoder_q-layer.8": 6653.0103, "encoder_q-layer.9": 6307.481, "epoch": 0.7, "inbatch_neg_score": 4.8567, "inbatch_pos_score": 5.4844, "learning_rate": 1.0263157894736843e-05, "loss": 3.3498, "norm_diff": 0.0697, "norm_loss": 0.0, "num_token_doc": 66.8651, "num_token_overlap": 11.6761, "num_token_query": 31.3814, "num_token_union": 65.1717, "num_word_context": 202.4398, "num_word_doc": 49.88, "num_word_query": 23.2965, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10291.4379, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8555, "query_norm": 2.52, "queue_k_norm": 2.5915, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3814, "sent_len_1": 66.8651, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.6687, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 161000 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3594, "doc_norm": 2.5937, "encoder_q-embeddings": 4858.3135, "encoder_q-layer.0": 3453.1216, "encoder_q-layer.1": 3686.9866, "encoder_q-layer.10": 6807.3188, "encoder_q-layer.11": 16523.4102, "encoder_q-layer.2": 4039.2998, "encoder_q-layer.3": 4152.27, "encoder_q-layer.4": 4523.1958, "encoder_q-layer.5": 4628.7842, "encoder_q-layer.6": 5030.1924, "encoder_q-layer.7": 5613.8657, "encoder_q-layer.8": 6900.8862, "encoder_q-layer.9": 6658.23, "epoch": 0.7, "inbatch_neg_score": 4.8551, "inbatch_pos_score": 5.4727, "learning_rate": 1.023684210526316e-05, "loss": 3.3594, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.748, "num_token_overlap": 11.6215, "num_token_query": 31.2497, "num_token_union": 65.0936, "num_word_context": 202.3099, "num_word_doc": 49.8107, "num_word_query": 23.1996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10329.1137, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8555, "query_norm": 2.5183, "queue_k_norm": 2.593, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2497, "sent_len_1": 66.748, "sent_len_max_0": 127.99, "sent_len_max_1": 209.2212, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161100 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3635, "doc_norm": 2.5941, "encoder_q-embeddings": 4721.3867, "encoder_q-layer.0": 3383.6169, "encoder_q-layer.1": 3548.6431, "encoder_q-layer.10": 6751.9932, "encoder_q-layer.11": 16882.0605, "encoder_q-layer.2": 3960.0806, "encoder_q-layer.3": 4008.0674, "encoder_q-layer.4": 4464.8862, "encoder_q-layer.5": 4392.2275, "encoder_q-layer.6": 5133.813, "encoder_q-layer.7": 6099.7837, "encoder_q-layer.8": 7016.7739, "encoder_q-layer.9": 6327.665, "epoch": 0.7, "inbatch_neg_score": 4.8579, "inbatch_pos_score": 5.4805, "learning_rate": 1.0210526315789474e-05, "loss": 3.3635, "norm_diff": 0.08, "norm_loss": 0.0, "num_token_doc": 66.6615, "num_token_overlap": 11.697, "num_token_query": 31.386, "num_token_union": 65.0341, "num_word_context": 202.0367, "num_word_doc": 49.754, "num_word_query": 23.3165, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10613.9368, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8594, "query_norm": 2.5141, "queue_k_norm": 2.5937, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.386, "sent_len_1": 66.6615, "sent_len_max_0": 127.945, "sent_len_max_1": 206.7625, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161200 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3656, "doc_norm": 2.5945, "encoder_q-embeddings": 4724.2075, "encoder_q-layer.0": 3442.6267, "encoder_q-layer.1": 3598.1938, "encoder_q-layer.10": 6928.9419, "encoder_q-layer.11": 17294.1504, "encoder_q-layer.2": 4028.3357, "encoder_q-layer.3": 4139.7378, "encoder_q-layer.4": 4395.0581, "encoder_q-layer.5": 4418.8892, "encoder_q-layer.6": 4857.7261, "encoder_q-layer.7": 5824.6636, "encoder_q-layer.8": 6828.5605, "encoder_q-layer.9": 6403.2192, "epoch": 0.7, "inbatch_neg_score": 4.8601, "inbatch_pos_score": 5.4688, "learning_rate": 1.0184210526315791e-05, "loss": 3.3656, "norm_diff": 0.0777, "norm_loss": 0.0, "num_token_doc": 66.8794, "num_token_overlap": 11.6859, "num_token_query": 31.4365, "num_token_union": 65.1823, "num_word_context": 202.6086, "num_word_doc": 49.9145, "num_word_query": 23.3485, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10491.7051, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.8594, "query_norm": 2.5167, "queue_k_norm": 2.5942, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4365, "sent_len_1": 66.8794, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.6175, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161300 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.344, "doc_norm": 2.5933, "encoder_q-embeddings": 9402.7783, "encoder_q-layer.0": 6695.0581, "encoder_q-layer.1": 7089.1035, "encoder_q-layer.10": 15572.627, "encoder_q-layer.11": 34697.7188, "encoder_q-layer.2": 8244.4932, "encoder_q-layer.3": 8247.4316, "encoder_q-layer.4": 8754.9004, "encoder_q-layer.5": 8932.1982, "encoder_q-layer.6": 10157.3848, "encoder_q-layer.7": 12014.2305, "encoder_q-layer.8": 14773.373, "encoder_q-layer.9": 13693.3066, "epoch": 0.7, "inbatch_neg_score": 4.8608, "inbatch_pos_score": 5.4766, "learning_rate": 1.0157894736842106e-05, "loss": 3.344, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.7465, "num_token_overlap": 11.6906, "num_token_query": 31.4288, "num_token_union": 65.1579, "num_word_context": 202.2055, "num_word_doc": 49.8176, "num_word_query": 23.3457, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21848.0172, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8594, "query_norm": 2.5187, "queue_k_norm": 2.5946, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4288, "sent_len_1": 66.7465, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.2738, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161400 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3652, "doc_norm": 2.594, "encoder_q-embeddings": 9704.96, "encoder_q-layer.0": 6891.3135, "encoder_q-layer.1": 7592.1299, "encoder_q-layer.10": 12682.7344, "encoder_q-layer.11": 32664.1328, "encoder_q-layer.2": 8684.9082, "encoder_q-layer.3": 8967.2129, "encoder_q-layer.4": 9475.0029, "encoder_q-layer.5": 9263.0986, "encoder_q-layer.6": 10135.8438, "encoder_q-layer.7": 11621.6875, "encoder_q-layer.8": 12813.1748, "encoder_q-layer.9": 11934.3828, "epoch": 0.7, "inbatch_neg_score": 4.8668, "inbatch_pos_score": 5.4727, "learning_rate": 1.0131578947368421e-05, "loss": 3.3652, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.8196, "num_token_overlap": 11.6788, "num_token_query": 31.3361, "num_token_union": 65.1529, "num_word_context": 202.4156, "num_word_doc": 49.8827, "num_word_query": 23.2723, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20828.5894, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8672, "query_norm": 2.5147, "queue_k_norm": 2.595, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3361, "sent_len_1": 66.8196, "sent_len_max_0": 127.985, "sent_len_max_1": 206.7438, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161500 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3464, "doc_norm": 2.5963, "encoder_q-embeddings": 9854.9766, "encoder_q-layer.0": 6894.1274, "encoder_q-layer.1": 7307.9507, "encoder_q-layer.10": 13354.5078, "encoder_q-layer.11": 33223.3984, "encoder_q-layer.2": 8019.4829, "encoder_q-layer.3": 8070.522, "encoder_q-layer.4": 8795.1094, "encoder_q-layer.5": 8904.0488, "encoder_q-layer.6": 9966.9775, "encoder_q-layer.7": 11689.2236, "encoder_q-layer.8": 13651.6318, "encoder_q-layer.9": 13007.667, "epoch": 0.7, "inbatch_neg_score": 4.8711, "inbatch_pos_score": 5.4883, "learning_rate": 1.0105263157894738e-05, "loss": 3.3464, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.7576, "num_token_overlap": 11.719, "num_token_query": 31.4836, "num_token_union": 65.15, "num_word_context": 202.3353, "num_word_doc": 49.8444, "num_word_query": 23.3812, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20958.1541, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8711, "query_norm": 2.5208, "queue_k_norm": 2.5948, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4836, "sent_len_1": 66.7576, "sent_len_max_0": 127.97, "sent_len_max_1": 208.6788, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161600 }, { "accuracy": 49.5605, "active_queue_size": 16384.0, "cl_loss": 3.3566, "doc_norm": 2.5955, "encoder_q-embeddings": 9964.1387, "encoder_q-layer.0": 7067.8027, "encoder_q-layer.1": 7643.4058, "encoder_q-layer.10": 13763.0244, "encoder_q-layer.11": 34248.0391, "encoder_q-layer.2": 8654.374, "encoder_q-layer.3": 9093.999, "encoder_q-layer.4": 9585.376, "encoder_q-layer.5": 10079.2988, "encoder_q-layer.6": 11747.623, "encoder_q-layer.7": 12151.2832, "encoder_q-layer.8": 14756.7686, "encoder_q-layer.9": 13214.6816, "epoch": 0.7, "inbatch_neg_score": 4.872, "inbatch_pos_score": 5.4688, "learning_rate": 1.0078947368421053e-05, "loss": 3.3566, "norm_diff": 0.0836, "norm_loss": 0.0, "num_token_doc": 66.6517, "num_token_overlap": 11.6642, "num_token_query": 31.4286, "num_token_union": 65.1065, "num_word_context": 202.4127, "num_word_doc": 49.729, "num_word_query": 23.3532, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21792.7348, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8711, "query_norm": 2.5119, "queue_k_norm": 2.595, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4286, "sent_len_1": 66.6517, "sent_len_max_0": 127.9737, "sent_len_max_1": 205.1275, "stdk": 0.0495, "stdq": 0.0425, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 161700 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3493, "doc_norm": 2.5946, "encoder_q-embeddings": 11274.9785, "encoder_q-layer.0": 8096.7363, "encoder_q-layer.1": 8485.21, "encoder_q-layer.10": 13938.0615, "encoder_q-layer.11": 34992.4727, "encoder_q-layer.2": 10133.8633, "encoder_q-layer.3": 10149.5322, "encoder_q-layer.4": 11311.5459, "encoder_q-layer.5": 12350.415, "encoder_q-layer.6": 12682.7617, "encoder_q-layer.7": 13449.0156, "encoder_q-layer.8": 15874.5273, "encoder_q-layer.9": 13698.7734, "epoch": 0.7, "inbatch_neg_score": 4.8757, "inbatch_pos_score": 5.4922, "learning_rate": 1.005263157894737e-05, "loss": 3.3493, "norm_diff": 0.0746, "norm_loss": 0.0, "num_token_doc": 66.7414, "num_token_overlap": 11.6834, "num_token_query": 31.3191, "num_token_union": 65.0404, "num_word_context": 202.2134, "num_word_doc": 49.799, "num_word_query": 23.2445, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23616.9486, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.875, "query_norm": 2.5201, "queue_k_norm": 2.5956, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3191, "sent_len_1": 66.7414, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.9087, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 161800 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3559, "doc_norm": 2.5974, "encoder_q-embeddings": 9406.6445, "encoder_q-layer.0": 6796.3354, "encoder_q-layer.1": 7317.4033, "encoder_q-layer.10": 13002.3477, "encoder_q-layer.11": 31974.2188, "encoder_q-layer.2": 8178.4453, "encoder_q-layer.3": 8581.5703, "encoder_q-layer.4": 9271.2363, "encoder_q-layer.5": 9495.4521, "encoder_q-layer.6": 10708.9023, "encoder_q-layer.7": 12028.2598, "encoder_q-layer.8": 13673.5215, "encoder_q-layer.9": 12485.4297, "epoch": 0.7, "inbatch_neg_score": 4.8781, "inbatch_pos_score": 5.4766, "learning_rate": 1.0026315789473686e-05, "loss": 3.3559, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.713, "num_token_overlap": 11.6522, "num_token_query": 31.3428, "num_token_union": 65.0838, "num_word_context": 202.2714, "num_word_doc": 49.7778, "num_word_query": 23.2843, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20522.471, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8789, "query_norm": 2.5205, "queue_k_norm": 2.5967, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3428, "sent_len_1": 66.713, "sent_len_max_0": 127.98, "sent_len_max_1": 207.9888, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 161900 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3581, "doc_norm": 2.5975, "encoder_q-embeddings": 9180.2109, "encoder_q-layer.0": 6610.7163, "encoder_q-layer.1": 7063.7412, "encoder_q-layer.10": 13532.5225, "encoder_q-layer.11": 33941.6953, "encoder_q-layer.2": 7963.8442, "encoder_q-layer.3": 8003.9565, "encoder_q-layer.4": 8367.0312, "encoder_q-layer.5": 8601.9561, "encoder_q-layer.6": 10021.0566, "encoder_q-layer.7": 11116.9805, "encoder_q-layer.8": 13854.5889, "encoder_q-layer.9": 13073.6494, "epoch": 0.7, "inbatch_neg_score": 4.8818, "inbatch_pos_score": 5.4883, "learning_rate": 1e-05, "loss": 3.3581, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.7198, "num_token_overlap": 11.6501, "num_token_query": 31.303, "num_token_union": 65.0674, "num_word_context": 202.0853, "num_word_doc": 49.7698, "num_word_query": 23.2458, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20845.9166, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8828, "query_norm": 2.5201, "queue_k_norm": 2.5967, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.303, "sent_len_1": 66.7198, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.6525, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162000 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3464, "doc_norm": 2.5948, "encoder_q-embeddings": 9278.9043, "encoder_q-layer.0": 6658.7476, "encoder_q-layer.1": 7130.6357, "encoder_q-layer.10": 14833.3965, "encoder_q-layer.11": 35293.2656, "encoder_q-layer.2": 8098.6685, "encoder_q-layer.3": 8392.3506, "encoder_q-layer.4": 8762.3223, "encoder_q-layer.5": 8843.3525, "encoder_q-layer.6": 9990.7783, "encoder_q-layer.7": 11390.6318, "encoder_q-layer.8": 14044.9785, "encoder_q-layer.9": 13020.0244, "epoch": 0.7, "inbatch_neg_score": 4.8853, "inbatch_pos_score": 5.4844, "learning_rate": 9.973684210526316e-06, "loss": 3.3464, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.8126, "num_token_overlap": 11.6716, "num_token_query": 31.4298, "num_token_union": 65.2033, "num_word_context": 202.3826, "num_word_doc": 49.846, "num_word_query": 23.3424, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21596.0152, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8828, "query_norm": 2.52, "queue_k_norm": 2.597, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4298, "sent_len_1": 66.8126, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.7312, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162100 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3609, "doc_norm": 2.5988, "encoder_q-embeddings": 11687.2861, "encoder_q-layer.0": 8670.2188, "encoder_q-layer.1": 9414.2598, "encoder_q-layer.10": 14621.9189, "encoder_q-layer.11": 34197.4297, "encoder_q-layer.2": 10921.9746, "encoder_q-layer.3": 11186.2451, "encoder_q-layer.4": 12810.4238, "encoder_q-layer.5": 14390.1689, "encoder_q-layer.6": 14499.7705, "encoder_q-layer.7": 14375.7207, "encoder_q-layer.8": 16336.1553, "encoder_q-layer.9": 12959.4863, "epoch": 0.7, "inbatch_neg_score": 4.8882, "inbatch_pos_score": 5.5078, "learning_rate": 9.947368421052632e-06, "loss": 3.3609, "norm_diff": 0.0772, "norm_loss": 0.0, "num_token_doc": 66.6851, "num_token_overlap": 11.6907, "num_token_query": 31.3702, "num_token_union": 65.044, "num_word_context": 201.9998, "num_word_doc": 49.776, "num_word_query": 23.303, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23776.3015, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8867, "query_norm": 2.5216, "queue_k_norm": 2.5979, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3702, "sent_len_1": 66.6851, "sent_len_max_0": 127.9463, "sent_len_max_1": 209.6475, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162200 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3425, "doc_norm": 2.6011, "encoder_q-embeddings": 10384.1621, "encoder_q-layer.0": 7049.7861, "encoder_q-layer.1": 7698.1499, "encoder_q-layer.10": 12681.9287, "encoder_q-layer.11": 32552.6836, "encoder_q-layer.2": 8474.3428, "encoder_q-layer.3": 9088.207, "encoder_q-layer.4": 10162.5322, "encoder_q-layer.5": 10016.7031, "encoder_q-layer.6": 11282.2539, "encoder_q-layer.7": 11964.6787, "encoder_q-layer.8": 13572.7451, "encoder_q-layer.9": 12375.4561, "epoch": 0.7, "inbatch_neg_score": 4.8901, "inbatch_pos_score": 5.5, "learning_rate": 9.921052631578947e-06, "loss": 3.3425, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.9448, "num_token_overlap": 11.6902, "num_token_query": 31.3093, "num_token_union": 65.1454, "num_word_context": 202.5156, "num_word_doc": 49.9268, "num_word_query": 23.2363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21299.7941, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8906, "query_norm": 2.5226, "queue_k_norm": 2.5989, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3093, "sent_len_1": 66.9448, "sent_len_max_0": 127.965, "sent_len_max_1": 208.9112, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 162300 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.354, "doc_norm": 2.5984, "encoder_q-embeddings": 9175.5479, "encoder_q-layer.0": 6641.3604, "encoder_q-layer.1": 7016.749, "encoder_q-layer.10": 14929.1562, "encoder_q-layer.11": 33441.6641, "encoder_q-layer.2": 8109.9229, "encoder_q-layer.3": 8184.2842, "encoder_q-layer.4": 8794.084, "encoder_q-layer.5": 9001.7988, "encoder_q-layer.6": 10463.2236, "encoder_q-layer.7": 11312.0703, "encoder_q-layer.8": 13715.8691, "encoder_q-layer.9": 12800.5137, "epoch": 0.7, "inbatch_neg_score": 4.8926, "inbatch_pos_score": 5.5156, "learning_rate": 9.894736842105264e-06, "loss": 3.354, "norm_diff": 0.0698, "norm_loss": 0.0, "num_token_doc": 66.7961, "num_token_overlap": 11.7219, "num_token_query": 31.4672, "num_token_union": 65.1706, "num_word_context": 202.581, "num_word_doc": 49.8287, "num_word_query": 23.3891, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21149.4109, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8906, "query_norm": 2.5286, "queue_k_norm": 2.5997, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4672, "sent_len_1": 66.7961, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.9412, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 162400 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3491, "doc_norm": 2.5988, "encoder_q-embeddings": 9855.9014, "encoder_q-layer.0": 7069.2075, "encoder_q-layer.1": 7607.9824, "encoder_q-layer.10": 12127.1377, "encoder_q-layer.11": 32375.0801, "encoder_q-layer.2": 8645.3799, "encoder_q-layer.3": 8709.5176, "encoder_q-layer.4": 9416.3662, "encoder_q-layer.5": 9629.3545, "encoder_q-layer.6": 10433.3047, "encoder_q-layer.7": 11392.5576, "encoder_q-layer.8": 12807.624, "encoder_q-layer.9": 12170.3701, "epoch": 0.71, "inbatch_neg_score": 4.8959, "inbatch_pos_score": 5.5117, "learning_rate": 9.868421052631579e-06, "loss": 3.3491, "norm_diff": 0.0721, "norm_loss": 0.0, "num_token_doc": 66.742, "num_token_overlap": 11.7166, "num_token_query": 31.3428, "num_token_union": 65.0802, "num_word_context": 202.1469, "num_word_doc": 49.8155, "num_word_query": 23.2712, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20639.712, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8945, "query_norm": 2.5267, "queue_k_norm": 2.5992, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3428, "sent_len_1": 66.742, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.2262, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162500 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3573, "doc_norm": 2.5984, "encoder_q-embeddings": 9883.8701, "encoder_q-layer.0": 6818.2441, "encoder_q-layer.1": 7066.3384, "encoder_q-layer.10": 15339.1211, "encoder_q-layer.11": 33865.6328, "encoder_q-layer.2": 7984.6377, "encoder_q-layer.3": 8264.2764, "encoder_q-layer.4": 8829.0615, "encoder_q-layer.5": 9122.6709, "encoder_q-layer.6": 10361.1299, "encoder_q-layer.7": 11933.5996, "encoder_q-layer.8": 13600.1836, "encoder_q-layer.9": 12796.9922, "epoch": 0.71, "inbatch_neg_score": 4.9004, "inbatch_pos_score": 5.5078, "learning_rate": 9.842105263157896e-06, "loss": 3.3573, "norm_diff": 0.0708, "norm_loss": 0.0, "num_token_doc": 66.6613, "num_token_overlap": 11.6779, "num_token_query": 31.3362, "num_token_union": 64.9992, "num_word_context": 202.2, "num_word_doc": 49.7408, "num_word_query": 23.2714, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21495.0044, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.8984, "query_norm": 2.5276, "queue_k_norm": 2.5994, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3362, "sent_len_1": 66.6613, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.71, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162600 }, { "accuracy": 53.3203, "active_queue_size": 16384.0, "cl_loss": 3.3502, "doc_norm": 2.6034, "encoder_q-embeddings": 9365.6953, "encoder_q-layer.0": 6653.772, "encoder_q-layer.1": 7287.103, "encoder_q-layer.10": 14108.085, "encoder_q-layer.11": 32148.8477, "encoder_q-layer.2": 7912.6279, "encoder_q-layer.3": 8061.082, "encoder_q-layer.4": 8638.6865, "encoder_q-layer.5": 9006.0781, "encoder_q-layer.6": 9555.9941, "encoder_q-layer.7": 11635.9297, "encoder_q-layer.8": 14597.4824, "encoder_q-layer.9": 12863.832, "epoch": 0.71, "inbatch_neg_score": 4.9001, "inbatch_pos_score": 5.5312, "learning_rate": 9.81578947368421e-06, "loss": 3.3502, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.6887, "num_token_overlap": 11.6596, "num_token_query": 31.3017, "num_token_union": 65.0107, "num_word_context": 202.1739, "num_word_doc": 49.792, "num_word_query": 23.2601, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20607.9293, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9023, "query_norm": 2.5302, "queue_k_norm": 2.601, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3017, "sent_len_1": 66.6887, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.6712, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 162700 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3556, "doc_norm": 2.6012, "encoder_q-embeddings": 9278.0342, "encoder_q-layer.0": 6606.2231, "encoder_q-layer.1": 7070.5137, "encoder_q-layer.10": 13577.8672, "encoder_q-layer.11": 33734.6641, "encoder_q-layer.2": 8044.5571, "encoder_q-layer.3": 8096.7793, "encoder_q-layer.4": 8618.9512, "encoder_q-layer.5": 9061.7002, "encoder_q-layer.6": 10237.9922, "encoder_q-layer.7": 11320.5098, "encoder_q-layer.8": 13718.7793, "encoder_q-layer.9": 12881.4619, "epoch": 0.71, "inbatch_neg_score": 4.9037, "inbatch_pos_score": 5.5195, "learning_rate": 9.789473684210526e-06, "loss": 3.3556, "norm_diff": 0.0701, "norm_loss": 0.0, "num_token_doc": 66.5904, "num_token_overlap": 11.6541, "num_token_query": 31.4142, "num_token_union": 65.0547, "num_word_context": 202.17, "num_word_doc": 49.7047, "num_word_query": 23.3345, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20912.8923, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9023, "query_norm": 2.5312, "queue_k_norm": 2.601, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4142, "sent_len_1": 66.5904, "sent_len_max_0": 127.975, "sent_len_max_1": 207.4387, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 162800 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3628, "doc_norm": 2.6023, "encoder_q-embeddings": 9466.5127, "encoder_q-layer.0": 6662.2363, "encoder_q-layer.1": 6989.9082, "encoder_q-layer.10": 12936.7148, "encoder_q-layer.11": 33450.8359, "encoder_q-layer.2": 7848.1836, "encoder_q-layer.3": 8297.7178, "encoder_q-layer.4": 8888.1865, "encoder_q-layer.5": 8922.5664, "encoder_q-layer.6": 10323.6309, "encoder_q-layer.7": 11255.5, "encoder_q-layer.8": 13602.1816, "encoder_q-layer.9": 12303.4893, "epoch": 0.71, "inbatch_neg_score": 4.9052, "inbatch_pos_score": 5.5195, "learning_rate": 9.763157894736842e-06, "loss": 3.3628, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.7757, "num_token_overlap": 11.6939, "num_token_query": 31.4321, "num_token_union": 65.1421, "num_word_context": 202.3124, "num_word_doc": 49.806, "num_word_query": 23.3442, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21031.1, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9062, "query_norm": 2.5266, "queue_k_norm": 2.6018, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4321, "sent_len_1": 66.7757, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.055, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 162900 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3454, "doc_norm": 2.602, "encoder_q-embeddings": 10006.1826, "encoder_q-layer.0": 7048.1724, "encoder_q-layer.1": 7689.021, "encoder_q-layer.10": 13241.501, "encoder_q-layer.11": 32212.5254, "encoder_q-layer.2": 8858.2852, "encoder_q-layer.3": 9097.1885, "encoder_q-layer.4": 9798.6973, "encoder_q-layer.5": 9894.5449, "encoder_q-layer.6": 10917.8828, "encoder_q-layer.7": 12289.249, "encoder_q-layer.8": 14271.4961, "encoder_q-layer.9": 12777.3975, "epoch": 0.71, "inbatch_neg_score": 4.9108, "inbatch_pos_score": 5.5195, "learning_rate": 9.736842105263157e-06, "loss": 3.3454, "norm_diff": 0.0702, "norm_loss": 0.0, "num_token_doc": 66.726, "num_token_overlap": 11.6645, "num_token_query": 31.337, "num_token_union": 65.0849, "num_word_context": 202.3755, "num_word_doc": 49.7417, "num_word_query": 23.2679, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20869.1543, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9102, "query_norm": 2.5317, "queue_k_norm": 2.6014, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.337, "sent_len_1": 66.726, "sent_len_max_0": 127.9975, "sent_len_max_1": 211.18, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163000 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.3499, "doc_norm": 2.6026, "encoder_q-embeddings": 10230.1426, "encoder_q-layer.0": 7160.0469, "encoder_q-layer.1": 7475.8042, "encoder_q-layer.10": 15033.1025, "encoder_q-layer.11": 34949.4102, "encoder_q-layer.2": 8792.4365, "encoder_q-layer.3": 9222.3359, "encoder_q-layer.4": 9878.7568, "encoder_q-layer.5": 10413.9609, "encoder_q-layer.6": 11254.6016, "encoder_q-layer.7": 12678.2441, "encoder_q-layer.8": 14644.2803, "encoder_q-layer.9": 13776.1543, "epoch": 0.71, "inbatch_neg_score": 4.9126, "inbatch_pos_score": 5.5078, "learning_rate": 9.710526315789474e-06, "loss": 3.3499, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.8222, "num_token_overlap": 11.6752, "num_token_query": 31.3742, "num_token_union": 65.1686, "num_word_context": 202.5784, "num_word_doc": 49.8558, "num_word_query": 23.2971, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22020.5554, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9141, "query_norm": 2.5258, "queue_k_norm": 2.6024, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3742, "sent_len_1": 66.8222, "sent_len_max_0": 127.99, "sent_len_max_1": 209.2388, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163100 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3539, "doc_norm": 2.6044, "encoder_q-embeddings": 9677.4834, "encoder_q-layer.0": 6916.3691, "encoder_q-layer.1": 7264.3652, "encoder_q-layer.10": 14471.0996, "encoder_q-layer.11": 35501.8711, "encoder_q-layer.2": 8325.3271, "encoder_q-layer.3": 8661.8008, "encoder_q-layer.4": 9223.3281, "encoder_q-layer.5": 9581.1367, "encoder_q-layer.6": 10676.7051, "encoder_q-layer.7": 11880.1787, "encoder_q-layer.8": 14433.6055, "encoder_q-layer.9": 13162.1113, "epoch": 0.71, "inbatch_neg_score": 4.9125, "inbatch_pos_score": 5.5234, "learning_rate": 9.68421052631579e-06, "loss": 3.3539, "norm_diff": 0.0763, "norm_loss": 0.0, "num_token_doc": 66.9792, "num_token_overlap": 11.6421, "num_token_query": 31.2668, "num_token_union": 65.185, "num_word_context": 202.4821, "num_word_doc": 49.9888, "num_word_query": 23.2202, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21941.2898, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9141, "query_norm": 2.5281, "queue_k_norm": 2.603, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2668, "sent_len_1": 66.9792, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.9525, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163200 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3561, "doc_norm": 2.6039, "encoder_q-embeddings": 10037.7266, "encoder_q-layer.0": 6849.4937, "encoder_q-layer.1": 7201.3267, "encoder_q-layer.10": 13259.502, "encoder_q-layer.11": 32314.5156, "encoder_q-layer.2": 8072.665, "encoder_q-layer.3": 8238.6699, "encoder_q-layer.4": 8933.8545, "encoder_q-layer.5": 9061.7607, "encoder_q-layer.6": 9779.4727, "encoder_q-layer.7": 11131.5752, "encoder_q-layer.8": 13481.6631, "encoder_q-layer.9": 12736.2441, "epoch": 0.71, "inbatch_neg_score": 4.9169, "inbatch_pos_score": 5.5273, "learning_rate": 9.657894736842106e-06, "loss": 3.3561, "norm_diff": 0.0782, "norm_loss": 0.0, "num_token_doc": 66.6566, "num_token_overlap": 11.6865, "num_token_query": 31.3652, "num_token_union": 65.0471, "num_word_context": 202.3175, "num_word_doc": 49.7469, "num_word_query": 23.2908, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20699.1293, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.918, "query_norm": 2.5257, "queue_k_norm": 2.6036, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3652, "sent_len_1": 66.6566, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.8325, "stdk": 0.0495, "stdq": 0.0427, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163300 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.3665, "doc_norm": 2.6027, "encoder_q-embeddings": 19218.7793, "encoder_q-layer.0": 13540.5713, "encoder_q-layer.1": 14485.1465, "encoder_q-layer.10": 26150.998, "encoder_q-layer.11": 66401.375, "encoder_q-layer.2": 15874.5088, "encoder_q-layer.3": 16603.0977, "encoder_q-layer.4": 18129.8926, "encoder_q-layer.5": 19286.9785, "encoder_q-layer.6": 21473.1602, "encoder_q-layer.7": 24525.6016, "encoder_q-layer.8": 28763.1348, "encoder_q-layer.9": 25675.832, "epoch": 0.71, "inbatch_neg_score": 4.919, "inbatch_pos_score": 5.5273, "learning_rate": 9.631578947368422e-06, "loss": 3.3665, "norm_diff": 0.0702, "norm_loss": 0.0, "num_token_doc": 66.6209, "num_token_overlap": 11.652, "num_token_query": 31.3031, "num_token_union": 64.9717, "num_word_context": 202.1975, "num_word_doc": 49.6977, "num_word_query": 23.2448, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42122.7372, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 4.9219, "query_norm": 2.5325, "queue_k_norm": 2.6041, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3031, "sent_len_1": 66.6209, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.8925, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163400 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3399, "doc_norm": 2.6073, "encoder_q-embeddings": 19591.4258, "encoder_q-layer.0": 15594.9658, "encoder_q-layer.1": 17653.3359, "encoder_q-layer.10": 12804.3262, "encoder_q-layer.11": 32369.334, "encoder_q-layer.2": 20377.5977, "encoder_q-layer.3": 20541.2598, "encoder_q-layer.4": 22556.5938, "encoder_q-layer.5": 19112.6387, "encoder_q-layer.6": 18201.1602, "encoder_q-layer.7": 19079.6641, "encoder_q-layer.8": 17320.4062, "encoder_q-layer.9": 13285.3711, "epoch": 0.71, "inbatch_neg_score": 4.9236, "inbatch_pos_score": 5.5312, "learning_rate": 9.605263157894737e-06, "loss": 3.3399, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.7503, "num_token_overlap": 11.6722, "num_token_query": 31.3199, "num_token_union": 65.0235, "num_word_context": 202.0103, "num_word_doc": 49.7791, "num_word_query": 23.2619, "postclip_grad_norm": 1.0, "preclip_grad_norm": 30582.0867, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 4.9219, "query_norm": 2.5321, "queue_k_norm": 2.6049, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3199, "sent_len_1": 66.7503, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.6362, "stdk": 0.0497, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163500 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3449, "doc_norm": 2.6058, "encoder_q-embeddings": 12123.8643, "encoder_q-layer.0": 8884.8721, "encoder_q-layer.1": 10241.8301, "encoder_q-layer.10": 13182.6094, "encoder_q-layer.11": 33070.8867, "encoder_q-layer.2": 12696.0352, "encoder_q-layer.3": 13838.9414, "encoder_q-layer.4": 14767.5732, "encoder_q-layer.5": 16652.3418, "encoder_q-layer.6": 18608.293, "encoder_q-layer.7": 19839.7539, "encoder_q-layer.8": 17439.1191, "encoder_q-layer.9": 13591.7021, "epoch": 0.71, "inbatch_neg_score": 4.9291, "inbatch_pos_score": 5.5273, "learning_rate": 9.578947368421052e-06, "loss": 3.3449, "norm_diff": 0.0733, "norm_loss": 0.0, "num_token_doc": 66.7581, "num_token_overlap": 11.6516, "num_token_query": 31.312, "num_token_union": 65.0831, "num_word_context": 202.0927, "num_word_doc": 49.8215, "num_word_query": 23.257, "postclip_grad_norm": 1.0, "preclip_grad_norm": 25622.5346, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9297, "query_norm": 2.5325, "queue_k_norm": 2.6051, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.312, "sent_len_1": 66.7581, "sent_len_max_0": 127.9938, "sent_len_max_1": 206.6025, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 163600 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3318, "doc_norm": 2.6097, "encoder_q-embeddings": 9308.5391, "encoder_q-layer.0": 6517.7637, "encoder_q-layer.1": 6975.7114, "encoder_q-layer.10": 14815.7129, "encoder_q-layer.11": 33751.5898, "encoder_q-layer.2": 8014.4092, "encoder_q-layer.3": 8064.3926, "encoder_q-layer.4": 8562.8711, "encoder_q-layer.5": 8876.8154, "encoder_q-layer.6": 10896.3408, "encoder_q-layer.7": 12619.6074, "encoder_q-layer.8": 14907.4199, "encoder_q-layer.9": 13481.8057, "epoch": 0.71, "inbatch_neg_score": 4.934, "inbatch_pos_score": 5.5664, "learning_rate": 9.552631578947369e-06, "loss": 3.3318, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.8062, "num_token_overlap": 11.6523, "num_token_query": 31.3167, "num_token_union": 65.1084, "num_word_context": 202.1569, "num_word_doc": 49.8392, "num_word_query": 23.2597, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21315.9291, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9336, "query_norm": 2.5378, "queue_k_norm": 2.6073, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3167, "sent_len_1": 66.8062, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.0387, "stdk": 0.0498, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 163700 }, { "accuracy": 53.5156, "active_queue_size": 16384.0, "cl_loss": 3.3564, "doc_norm": 2.6068, "encoder_q-embeddings": 9097.1426, "encoder_q-layer.0": 6375.0908, "encoder_q-layer.1": 6903.6304, "encoder_q-layer.10": 15487.085, "encoder_q-layer.11": 32102.0254, "encoder_q-layer.2": 7839.3267, "encoder_q-layer.3": 8131.5542, "encoder_q-layer.4": 8859.0781, "encoder_q-layer.5": 8889.1758, "encoder_q-layer.6": 9913.1758, "encoder_q-layer.7": 11023.083, "encoder_q-layer.8": 14073.332, "encoder_q-layer.9": 12555.1006, "epoch": 0.71, "inbatch_neg_score": 4.9394, "inbatch_pos_score": 5.5547, "learning_rate": 9.526315789473684e-06, "loss": 3.3564, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.887, "num_token_overlap": 11.7189, "num_token_query": 31.4843, "num_token_union": 65.2213, "num_word_context": 202.3759, "num_word_doc": 49.892, "num_word_query": 23.3956, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20116.3838, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9375, "query_norm": 2.5355, "queue_k_norm": 2.6072, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4843, "sent_len_1": 66.887, "sent_len_max_0": 127.9638, "sent_len_max_1": 210.2038, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 163800 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3683, "doc_norm": 2.6097, "encoder_q-embeddings": 9459.4385, "encoder_q-layer.0": 6704.2202, "encoder_q-layer.1": 7192.0186, "encoder_q-layer.10": 14335.415, "encoder_q-layer.11": 35720.3203, "encoder_q-layer.2": 7903.625, "encoder_q-layer.3": 8352.7432, "encoder_q-layer.4": 9327.9209, "encoder_q-layer.5": 9284.7803, "encoder_q-layer.6": 10433.3008, "encoder_q-layer.7": 12449.2217, "encoder_q-layer.8": 14759.9062, "encoder_q-layer.9": 14217.0771, "epoch": 0.71, "inbatch_neg_score": 4.9433, "inbatch_pos_score": 5.5625, "learning_rate": 9.5e-06, "loss": 3.3683, "norm_diff": 0.0657, "norm_loss": 0.0, "num_token_doc": 66.783, "num_token_overlap": 11.6592, "num_token_query": 31.3135, "num_token_union": 65.0932, "num_word_context": 202.1748, "num_word_doc": 49.824, "num_word_query": 23.2602, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21780.5691, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 4.9453, "query_norm": 2.544, "queue_k_norm": 2.6078, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3135, "sent_len_1": 66.783, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.8413, "stdk": 0.0497, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 163900 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3461, "doc_norm": 2.6097, "encoder_q-embeddings": 4847.2031, "encoder_q-layer.0": 3284.6311, "encoder_q-layer.1": 3521.7029, "encoder_q-layer.10": 6700.7715, "encoder_q-layer.11": 16494.1621, "encoder_q-layer.2": 3896.6663, "encoder_q-layer.3": 4052.2578, "encoder_q-layer.4": 4479.9707, "encoder_q-layer.5": 4529.3174, "encoder_q-layer.6": 5100.2466, "encoder_q-layer.7": 6097.1235, "encoder_q-layer.8": 7051.1621, "encoder_q-layer.9": 6321.7822, "epoch": 0.71, "inbatch_neg_score": 4.9526, "inbatch_pos_score": 5.5664, "learning_rate": 9.473684210526317e-06, "loss": 3.3461, "norm_diff": 0.0656, "norm_loss": 0.0, "num_token_doc": 66.7552, "num_token_overlap": 11.7064, "num_token_query": 31.4914, "num_token_union": 65.1567, "num_word_context": 202.2721, "num_word_doc": 49.8372, "num_word_query": 23.396, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10549.4038, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9531, "query_norm": 2.5441, "queue_k_norm": 2.6085, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4914, "sent_len_1": 66.7552, "sent_len_max_0": 127.995, "sent_len_max_1": 211.2275, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164000 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3556, "doc_norm": 2.6088, "encoder_q-embeddings": 4710.2959, "encoder_q-layer.0": 3413.9958, "encoder_q-layer.1": 3600.468, "encoder_q-layer.10": 8502.0137, "encoder_q-layer.11": 18403.627, "encoder_q-layer.2": 4030.8171, "encoder_q-layer.3": 4134.9668, "encoder_q-layer.4": 4434.6045, "encoder_q-layer.5": 4402.585, "encoder_q-layer.6": 4932.3301, "encoder_q-layer.7": 6001.8433, "encoder_q-layer.8": 7100.4502, "encoder_q-layer.9": 6675.168, "epoch": 0.71, "inbatch_neg_score": 4.9535, "inbatch_pos_score": 5.5664, "learning_rate": 9.447368421052632e-06, "loss": 3.3556, "norm_diff": 0.0691, "norm_loss": 0.0, "num_token_doc": 66.6689, "num_token_overlap": 11.6754, "num_token_query": 31.4235, "num_token_union": 65.0874, "num_word_context": 202.1565, "num_word_doc": 49.7318, "num_word_query": 23.333, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11365.3226, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9531, "query_norm": 2.5398, "queue_k_norm": 2.6081, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4235, "sent_len_1": 66.6689, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.1375, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 164100 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3554, "doc_norm": 2.6138, "encoder_q-embeddings": 4877.7173, "encoder_q-layer.0": 3334.1484, "encoder_q-layer.1": 3605.6128, "encoder_q-layer.10": 7195.6836, "encoder_q-layer.11": 16708.832, "encoder_q-layer.2": 4025.1328, "encoder_q-layer.3": 4076.8196, "encoder_q-layer.4": 4275.4277, "encoder_q-layer.5": 4391.5498, "encoder_q-layer.6": 4958.0103, "encoder_q-layer.7": 5780.0005, "encoder_q-layer.8": 7087.6021, "encoder_q-layer.9": 6312.502, "epoch": 0.71, "inbatch_neg_score": 4.9572, "inbatch_pos_score": 5.582, "learning_rate": 9.421052631578947e-06, "loss": 3.3554, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.7573, "num_token_overlap": 11.6246, "num_token_query": 31.2237, "num_token_union": 65.0362, "num_word_context": 202.237, "num_word_doc": 49.8056, "num_word_query": 23.1691, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10438.8108, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.957, "query_norm": 2.5451, "queue_k_norm": 2.61, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2237, "sent_len_1": 66.7573, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.5613, "stdk": 0.0498, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164200 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3444, "doc_norm": 2.6105, "encoder_q-embeddings": 4813.7178, "encoder_q-layer.0": 3447.4592, "encoder_q-layer.1": 3570.7866, "encoder_q-layer.10": 6808.9585, "encoder_q-layer.11": 17570.2715, "encoder_q-layer.2": 4068.6736, "encoder_q-layer.3": 4226.3125, "encoder_q-layer.4": 4652.7095, "encoder_q-layer.5": 4804.9253, "encoder_q-layer.6": 5240.9102, "encoder_q-layer.7": 5809.2461, "encoder_q-layer.8": 7133.9312, "encoder_q-layer.9": 6546.0449, "epoch": 0.71, "inbatch_neg_score": 4.9627, "inbatch_pos_score": 5.5703, "learning_rate": 9.394736842105264e-06, "loss": 3.3444, "norm_diff": 0.0686, "norm_loss": 0.0, "num_token_doc": 66.8429, "num_token_overlap": 11.6807, "num_token_query": 31.3293, "num_token_union": 65.1271, "num_word_context": 202.3838, "num_word_doc": 49.8943, "num_word_query": 23.2577, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10994.5794, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9609, "query_norm": 2.5419, "queue_k_norm": 2.6101, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3293, "sent_len_1": 66.8429, "sent_len_max_0": 127.995, "sent_len_max_1": 208.7025, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164300 }, { "accuracy": 53.8086, "active_queue_size": 16384.0, "cl_loss": 3.353, "doc_norm": 2.612, "encoder_q-embeddings": 4852.9736, "encoder_q-layer.0": 3451.6338, "encoder_q-layer.1": 3632.1118, "encoder_q-layer.10": 7958.4053, "encoder_q-layer.11": 18462.3125, "encoder_q-layer.2": 4125.8623, "encoder_q-layer.3": 4199.1797, "encoder_q-layer.4": 4554.0791, "encoder_q-layer.5": 4749.0962, "encoder_q-layer.6": 5460.1147, "encoder_q-layer.7": 6176.4482, "encoder_q-layer.8": 7434.4619, "encoder_q-layer.9": 7369.749, "epoch": 0.71, "inbatch_neg_score": 4.9654, "inbatch_pos_score": 5.6094, "learning_rate": 9.368421052631579e-06, "loss": 3.353, "norm_diff": 0.0668, "norm_loss": 0.0, "num_token_doc": 66.7409, "num_token_overlap": 11.6674, "num_token_query": 31.3931, "num_token_union": 65.1147, "num_word_context": 202.1696, "num_word_doc": 49.784, "num_word_query": 23.3272, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11175.28, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9648, "query_norm": 2.5453, "queue_k_norm": 2.6107, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3931, "sent_len_1": 66.7409, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.42, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 164400 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.357, "doc_norm": 2.6095, "encoder_q-embeddings": 4718.5352, "encoder_q-layer.0": 3292.4141, "encoder_q-layer.1": 3446.0889, "encoder_q-layer.10": 6518.522, "encoder_q-layer.11": 16670.2168, "encoder_q-layer.2": 3818.6243, "encoder_q-layer.3": 3920.9666, "encoder_q-layer.4": 4163.7021, "encoder_q-layer.5": 4175.5645, "encoder_q-layer.6": 5010.9766, "encoder_q-layer.7": 5616.854, "encoder_q-layer.8": 6866.8633, "encoder_q-layer.9": 6274.1636, "epoch": 0.71, "inbatch_neg_score": 4.97, "inbatch_pos_score": 5.5859, "learning_rate": 9.342105263157895e-06, "loss": 3.357, "norm_diff": 0.0663, "norm_loss": 0.0, "num_token_doc": 66.7095, "num_token_overlap": 11.6332, "num_token_query": 31.2678, "num_token_union": 65.0111, "num_word_context": 202.2377, "num_word_doc": 49.7742, "num_word_query": 23.2212, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10308.8791, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9688, "query_norm": 2.5433, "queue_k_norm": 2.6115, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2678, "sent_len_1": 66.7095, "sent_len_max_0": 127.96, "sent_len_max_1": 208.3125, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 164500 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3536, "doc_norm": 2.6105, "encoder_q-embeddings": 4384.0088, "encoder_q-layer.0": 3137.8477, "encoder_q-layer.1": 3362.3926, "encoder_q-layer.10": 7432.5132, "encoder_q-layer.11": 16617.3125, "encoder_q-layer.2": 3671.3171, "encoder_q-layer.3": 3938.0586, "encoder_q-layer.4": 4206.2588, "encoder_q-layer.5": 4108.8745, "encoder_q-layer.6": 4787.4551, "encoder_q-layer.7": 5596.7666, "encoder_q-layer.8": 6848.7402, "encoder_q-layer.9": 6427.5024, "epoch": 0.71, "inbatch_neg_score": 4.9708, "inbatch_pos_score": 5.5898, "learning_rate": 9.31578947368421e-06, "loss": 3.3536, "norm_diff": 0.0674, "norm_loss": 0.0, "num_token_doc": 66.7214, "num_token_overlap": 11.6579, "num_token_query": 31.2708, "num_token_union": 65.0242, "num_word_context": 202.0973, "num_word_doc": 49.782, "num_word_query": 23.2153, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10262.621, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9727, "query_norm": 2.5431, "queue_k_norm": 2.6125, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2708, "sent_len_1": 66.7214, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.5938, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164600 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.364, "doc_norm": 2.6111, "encoder_q-embeddings": 4886.7046, "encoder_q-layer.0": 3429.7651, "encoder_q-layer.1": 3728.626, "encoder_q-layer.10": 7121.2314, "encoder_q-layer.11": 17355.75, "encoder_q-layer.2": 4115.8936, "encoder_q-layer.3": 4265.5679, "encoder_q-layer.4": 4513.5869, "encoder_q-layer.5": 4762.8857, "encoder_q-layer.6": 5314.6646, "encoder_q-layer.7": 5871.6494, "encoder_q-layer.8": 6823.1831, "encoder_q-layer.9": 6393.9707, "epoch": 0.71, "inbatch_neg_score": 4.9745, "inbatch_pos_score": 5.5898, "learning_rate": 9.289473684210527e-06, "loss": 3.364, "norm_diff": 0.0717, "norm_loss": 0.0, "num_token_doc": 66.7767, "num_token_overlap": 11.6546, "num_token_query": 31.3327, "num_token_union": 65.1338, "num_word_context": 202.4595, "num_word_doc": 49.8176, "num_word_query": 23.2572, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10837.834, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9727, "query_norm": 2.5393, "queue_k_norm": 2.6138, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3327, "sent_len_1": 66.7767, "sent_len_max_0": 127.995, "sent_len_max_1": 210.7138, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164700 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.3551, "doc_norm": 2.6142, "encoder_q-embeddings": 4745.3662, "encoder_q-layer.0": 3417.2129, "encoder_q-layer.1": 3758.9111, "encoder_q-layer.10": 6854.2549, "encoder_q-layer.11": 17059.2012, "encoder_q-layer.2": 4188.0762, "encoder_q-layer.3": 4241.356, "encoder_q-layer.4": 4504.0078, "encoder_q-layer.5": 4661.9385, "encoder_q-layer.6": 5191.7246, "encoder_q-layer.7": 5630.7065, "encoder_q-layer.8": 6592.6421, "encoder_q-layer.9": 6343.5361, "epoch": 0.72, "inbatch_neg_score": 4.9781, "inbatch_pos_score": 5.5781, "learning_rate": 9.263157894736844e-06, "loss": 3.3551, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.8627, "num_token_overlap": 11.6949, "num_token_query": 31.4764, "num_token_union": 65.2453, "num_word_context": 202.6141, "num_word_doc": 49.9012, "num_word_query": 23.3831, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10511.1031, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9766, "query_norm": 2.5385, "queue_k_norm": 2.6148, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4764, "sent_len_1": 66.8627, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.8587, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164800 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3425, "doc_norm": 2.6174, "encoder_q-embeddings": 4493.0396, "encoder_q-layer.0": 3302.0427, "encoder_q-layer.1": 3516.7422, "encoder_q-layer.10": 7074.6226, "encoder_q-layer.11": 16754.373, "encoder_q-layer.2": 3805.3118, "encoder_q-layer.3": 3909.4966, "encoder_q-layer.4": 4205.2544, "encoder_q-layer.5": 4347.9556, "encoder_q-layer.6": 4926.3247, "encoder_q-layer.7": 5587.6353, "encoder_q-layer.8": 6914.021, "encoder_q-layer.9": 6265.4458, "epoch": 0.72, "inbatch_neg_score": 4.9791, "inbatch_pos_score": 5.5977, "learning_rate": 9.236842105263159e-06, "loss": 3.3425, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.8427, "num_token_overlap": 11.7062, "num_token_query": 31.4589, "num_token_union": 65.1712, "num_word_context": 202.3874, "num_word_doc": 49.8892, "num_word_query": 23.3678, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10432.1467, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9805, "query_norm": 2.5424, "queue_k_norm": 2.6145, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4589, "sent_len_1": 66.8427, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.6238, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 164900 }, { "accuracy": 54.248, "active_queue_size": 16384.0, "cl_loss": 3.3541, "doc_norm": 2.6153, "encoder_q-embeddings": 4498.3257, "encoder_q-layer.0": 3264.3923, "encoder_q-layer.1": 3486.9839, "encoder_q-layer.10": 6195.6611, "encoder_q-layer.11": 15639.6465, "encoder_q-layer.2": 3765.822, "encoder_q-layer.3": 3849.5854, "encoder_q-layer.4": 4114.4272, "encoder_q-layer.5": 4384.4448, "encoder_q-layer.6": 4633.4692, "encoder_q-layer.7": 5076.5488, "encoder_q-layer.8": 6091.7939, "encoder_q-layer.9": 5832.6152, "epoch": 0.72, "inbatch_neg_score": 4.9786, "inbatch_pos_score": 5.6133, "learning_rate": 9.210526315789474e-06, "loss": 3.3541, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.7207, "num_token_overlap": 11.6771, "num_token_query": 31.3805, "num_token_union": 65.0565, "num_word_context": 202.2362, "num_word_doc": 49.7558, "num_word_query": 23.3109, "postclip_grad_norm": 1.0, "preclip_grad_norm": 9691.0334, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9805, "query_norm": 2.5417, "queue_k_norm": 2.6165, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3805, "sent_len_1": 66.7207, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.37, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 165000 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.342, "doc_norm": 2.6138, "encoder_q-embeddings": 5009.4399, "encoder_q-layer.0": 3452.6353, "encoder_q-layer.1": 3742.7939, "encoder_q-layer.10": 7220.1489, "encoder_q-layer.11": 17257.1504, "encoder_q-layer.2": 4129.4746, "encoder_q-layer.3": 4160.6567, "encoder_q-layer.4": 4580.5312, "encoder_q-layer.5": 4660.9897, "encoder_q-layer.6": 5149.6382, "encoder_q-layer.7": 5843.8638, "encoder_q-layer.8": 6971.7041, "encoder_q-layer.9": 6507.8276, "epoch": 0.72, "inbatch_neg_score": 4.9826, "inbatch_pos_score": 5.6016, "learning_rate": 9.18421052631579e-06, "loss": 3.342, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.765, "num_token_overlap": 11.7192, "num_token_query": 31.4364, "num_token_union": 65.1149, "num_word_context": 202.2011, "num_word_doc": 49.8168, "num_word_query": 23.3293, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10918.9288, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9844, "query_norm": 2.5426, "queue_k_norm": 2.6153, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4364, "sent_len_1": 66.765, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.9875, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165100 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3458, "doc_norm": 2.6198, "encoder_q-embeddings": 4773.4756, "encoder_q-layer.0": 3342.4377, "encoder_q-layer.1": 3504.553, "encoder_q-layer.10": 6630.1938, "encoder_q-layer.11": 17321.7051, "encoder_q-layer.2": 3947.5649, "encoder_q-layer.3": 4020.9561, "encoder_q-layer.4": 4330.1372, "encoder_q-layer.5": 4456.4424, "encoder_q-layer.6": 5140.3403, "encoder_q-layer.7": 5660.1621, "encoder_q-layer.8": 6821.124, "encoder_q-layer.9": 6355.7949, "epoch": 0.72, "inbatch_neg_score": 4.9839, "inbatch_pos_score": 5.5977, "learning_rate": 9.157894736842105e-06, "loss": 3.3458, "norm_diff": 0.0791, "norm_loss": 0.0, "num_token_doc": 66.894, "num_token_overlap": 11.6723, "num_token_query": 31.3252, "num_token_union": 65.1539, "num_word_context": 202.2651, "num_word_doc": 49.8773, "num_word_query": 23.2688, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10818.6539, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9844, "query_norm": 2.5407, "queue_k_norm": 2.6147, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3252, "sent_len_1": 66.894, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.745, "stdk": 0.0498, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165200 }, { "accuracy": 53.3203, "active_queue_size": 16384.0, "cl_loss": 3.3476, "doc_norm": 2.6187, "encoder_q-embeddings": 5398.2969, "encoder_q-layer.0": 3933.8608, "encoder_q-layer.1": 4239.1992, "encoder_q-layer.10": 7403.7378, "encoder_q-layer.11": 16606.2852, "encoder_q-layer.2": 4817.9863, "encoder_q-layer.3": 4916.7202, "encoder_q-layer.4": 5148.8081, "encoder_q-layer.5": 5181.9395, "encoder_q-layer.6": 5710.1553, "encoder_q-layer.7": 6333.7441, "encoder_q-layer.8": 7037.0771, "encoder_q-layer.9": 6765.2998, "epoch": 0.72, "inbatch_neg_score": 4.9852, "inbatch_pos_score": 5.6094, "learning_rate": 9.131578947368422e-06, "loss": 3.3476, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.8378, "num_token_overlap": 11.6952, "num_token_query": 31.4264, "num_token_union": 65.1593, "num_word_context": 202.416, "num_word_doc": 49.8606, "num_word_query": 23.348, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10871.7783, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9844, "query_norm": 2.5408, "queue_k_norm": 2.616, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4264, "sent_len_1": 66.8378, "sent_len_max_0": 127.99, "sent_len_max_1": 210.1275, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165300 }, { "accuracy": 52.832, "active_queue_size": 16384.0, "cl_loss": 3.335, "doc_norm": 2.6171, "encoder_q-embeddings": 4539.4658, "encoder_q-layer.0": 3213.5212, "encoder_q-layer.1": 3441.4661, "encoder_q-layer.10": 6363.9829, "encoder_q-layer.11": 17114.2656, "encoder_q-layer.2": 3836.4187, "encoder_q-layer.3": 4052.2981, "encoder_q-layer.4": 4210.8423, "encoder_q-layer.5": 4319.1309, "encoder_q-layer.6": 4769.7485, "encoder_q-layer.7": 5623.3198, "encoder_q-layer.8": 6684.6597, "encoder_q-layer.9": 6172.0078, "epoch": 0.72, "inbatch_neg_score": 4.9873, "inbatch_pos_score": 5.6133, "learning_rate": 9.105263157894737e-06, "loss": 3.335, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.9371, "num_token_overlap": 11.7131, "num_token_query": 31.4705, "num_token_union": 65.2484, "num_word_context": 202.546, "num_word_doc": 49.9459, "num_word_query": 23.364, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10397.272, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9883, "query_norm": 2.5402, "queue_k_norm": 2.6168, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4705, "sent_len_1": 66.9371, "sent_len_max_0": 127.985, "sent_len_max_1": 209.6163, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165400 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3641, "doc_norm": 2.6184, "encoder_q-embeddings": 4849.9458, "encoder_q-layer.0": 3438.6116, "encoder_q-layer.1": 3655.8777, "encoder_q-layer.10": 6943.395, "encoder_q-layer.11": 16429.5527, "encoder_q-layer.2": 4080.6672, "encoder_q-layer.3": 4086.5923, "encoder_q-layer.4": 4554.4731, "encoder_q-layer.5": 4761.6646, "encoder_q-layer.6": 5063.0322, "encoder_q-layer.7": 5770.9487, "encoder_q-layer.8": 7016.9238, "encoder_q-layer.9": 6288.793, "epoch": 0.72, "inbatch_neg_score": 4.9855, "inbatch_pos_score": 5.6055, "learning_rate": 9.078947368421054e-06, "loss": 3.3641, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.754, "num_token_overlap": 11.6594, "num_token_query": 31.2395, "num_token_union": 65.0142, "num_word_context": 202.1576, "num_word_doc": 49.7968, "num_word_query": 23.1868, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10324.11, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9844, "query_norm": 2.5412, "queue_k_norm": 2.6171, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2395, "sent_len_1": 66.754, "sent_len_max_0": 127.9737, "sent_len_max_1": 211.5025, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165500 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3418, "doc_norm": 2.6173, "encoder_q-embeddings": 5230.6733, "encoder_q-layer.0": 3666.251, "encoder_q-layer.1": 3792.4463, "encoder_q-layer.10": 6713.7207, "encoder_q-layer.11": 16715.3691, "encoder_q-layer.2": 4222.2969, "encoder_q-layer.3": 4479.8013, "encoder_q-layer.4": 4558.9697, "encoder_q-layer.5": 4469.8618, "encoder_q-layer.6": 4987.2021, "encoder_q-layer.7": 5884.457, "encoder_q-layer.8": 7056.7793, "encoder_q-layer.9": 6589.5903, "epoch": 0.72, "inbatch_neg_score": 4.9882, "inbatch_pos_score": 5.6055, "learning_rate": 9.05263157894737e-06, "loss": 3.3418, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.7077, "num_token_overlap": 11.6858, "num_token_query": 31.385, "num_token_union": 65.0745, "num_word_context": 202.3299, "num_word_doc": 49.7672, "num_word_query": 23.3229, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10656.676, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9883, "query_norm": 2.545, "queue_k_norm": 2.6184, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.385, "sent_len_1": 66.7077, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.0475, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 165600 }, { "accuracy": 53.6621, "active_queue_size": 16384.0, "cl_loss": 3.3291, "doc_norm": 2.6197, "encoder_q-embeddings": 4990.04, "encoder_q-layer.0": 3586.8257, "encoder_q-layer.1": 3826.7227, "encoder_q-layer.10": 6680.8608, "encoder_q-layer.11": 16933.5879, "encoder_q-layer.2": 4332.6855, "encoder_q-layer.3": 4459.0581, "encoder_q-layer.4": 5056.3921, "encoder_q-layer.5": 5200.2036, "encoder_q-layer.6": 5738.0718, "encoder_q-layer.7": 6091.853, "encoder_q-layer.8": 7083.1748, "encoder_q-layer.9": 6438.7627, "epoch": 0.72, "inbatch_neg_score": 4.9916, "inbatch_pos_score": 5.6211, "learning_rate": 9.026315789473684e-06, "loss": 3.3291, "norm_diff": 0.0695, "norm_loss": 0.0, "num_token_doc": 66.8199, "num_token_overlap": 11.7408, "num_token_query": 31.4988, "num_token_union": 65.1698, "num_word_context": 202.3657, "num_word_doc": 49.8722, "num_word_query": 23.3941, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10764.5663, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9922, "query_norm": 2.5503, "queue_k_norm": 2.6189, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4988, "sent_len_1": 66.8199, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.6575, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 165700 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3525, "doc_norm": 2.6222, "encoder_q-embeddings": 4588.7222, "encoder_q-layer.0": 3278.7947, "encoder_q-layer.1": 3481.4519, "encoder_q-layer.10": 6908.5483, "encoder_q-layer.11": 17229.6816, "encoder_q-layer.2": 3898.1316, "encoder_q-layer.3": 4059.1069, "encoder_q-layer.4": 4514.6841, "encoder_q-layer.5": 4454.1899, "encoder_q-layer.6": 5032.6807, "encoder_q-layer.7": 5698.1528, "encoder_q-layer.8": 6805.1797, "encoder_q-layer.9": 6534.6133, "epoch": 0.72, "inbatch_neg_score": 4.9947, "inbatch_pos_score": 5.6133, "learning_rate": 9e-06, "loss": 3.3525, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.7089, "num_token_overlap": 11.6795, "num_token_query": 31.3806, "num_token_union": 65.0791, "num_word_context": 202.1887, "num_word_doc": 49.7234, "num_word_query": 23.2913, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10521.2053, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9961, "query_norm": 2.5398, "queue_k_norm": 2.6198, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3806, "sent_len_1": 66.7089, "sent_len_max_0": 128.0, "sent_len_max_1": 210.3487, "stdk": 0.0497, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 165800 }, { "accuracy": 49.7559, "active_queue_size": 16384.0, "cl_loss": 3.3338, "doc_norm": 2.6191, "encoder_q-embeddings": 4791.7676, "encoder_q-layer.0": 3420.0681, "encoder_q-layer.1": 3691.6047, "encoder_q-layer.10": 7369.7236, "encoder_q-layer.11": 17509.9043, "encoder_q-layer.2": 4135.3823, "encoder_q-layer.3": 4289.7432, "encoder_q-layer.4": 4566.3345, "encoder_q-layer.5": 4722.8789, "encoder_q-layer.6": 5384.9688, "encoder_q-layer.7": 6157.7876, "encoder_q-layer.8": 7511.9316, "encoder_q-layer.9": 6954.1582, "epoch": 0.72, "inbatch_neg_score": 4.9976, "inbatch_pos_score": 5.6133, "learning_rate": 8.973684210526317e-06, "loss": 3.3338, "norm_diff": 0.0738, "norm_loss": 0.0, "num_token_doc": 66.8439, "num_token_overlap": 11.7163, "num_token_query": 31.4234, "num_token_union": 65.1912, "num_word_context": 202.5419, "num_word_doc": 49.9242, "num_word_query": 23.3456, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10889.8312, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 4.9961, "query_norm": 2.5454, "queue_k_norm": 2.6193, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4234, "sent_len_1": 66.8439, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.2325, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 165900 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.347, "doc_norm": 2.6199, "encoder_q-embeddings": 9399.5977, "encoder_q-layer.0": 6580.9292, "encoder_q-layer.1": 7078.1299, "encoder_q-layer.10": 13299.3086, "encoder_q-layer.11": 33202.0469, "encoder_q-layer.2": 7848.3755, "encoder_q-layer.3": 8021.0439, "encoder_q-layer.4": 8530.7324, "encoder_q-layer.5": 9077.6826, "encoder_q-layer.6": 9937.4043, "encoder_q-layer.7": 10911.9785, "encoder_q-layer.8": 13099.2344, "encoder_q-layer.9": 12464.2275, "epoch": 0.72, "inbatch_neg_score": 4.9987, "inbatch_pos_score": 5.625, "learning_rate": 8.947368421052632e-06, "loss": 3.347, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.7749, "num_token_overlap": 11.6801, "num_token_query": 31.3619, "num_token_union": 65.098, "num_word_context": 202.2337, "num_word_doc": 49.8171, "num_word_query": 23.3014, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20748.9698, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0, "query_norm": 2.5476, "queue_k_norm": 2.6199, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3619, "sent_len_1": 66.7749, "sent_len_max_0": 127.9813, "sent_len_max_1": 210.0025, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 166000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3623, "doc_norm": 2.6207, "encoder_q-embeddings": 10173.9375, "encoder_q-layer.0": 7113.4883, "encoder_q-layer.1": 7558.3008, "encoder_q-layer.10": 13839.3789, "encoder_q-layer.11": 34304.1523, "encoder_q-layer.2": 8364.1816, "encoder_q-layer.3": 8715.8076, "encoder_q-layer.4": 9059.7578, "encoder_q-layer.5": 9653.3936, "encoder_q-layer.6": 10362.041, "encoder_q-layer.7": 12749.7588, "encoder_q-layer.8": 14360.4629, "encoder_q-layer.9": 12955.3086, "epoch": 0.72, "inbatch_neg_score": 5.0013, "inbatch_pos_score": 5.6094, "learning_rate": 8.921052631578949e-06, "loss": 3.3623, "norm_diff": 0.0731, "norm_loss": 0.0, "num_token_doc": 66.6844, "num_token_overlap": 11.6163, "num_token_query": 31.3229, "num_token_union": 65.0811, "num_word_context": 202.0636, "num_word_doc": 49.7644, "num_word_query": 23.2464, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21667.6832, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0, "query_norm": 2.5476, "queue_k_norm": 2.6198, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3229, "sent_len_1": 66.6844, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.4863, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166100 }, { "accuracy": 52.3926, "active_queue_size": 16384.0, "cl_loss": 3.3462, "doc_norm": 2.6211, "encoder_q-embeddings": 10477.0645, "encoder_q-layer.0": 7457.7598, "encoder_q-layer.1": 8496.6846, "encoder_q-layer.10": 13620.7432, "encoder_q-layer.11": 33721.3984, "encoder_q-layer.2": 9660.9756, "encoder_q-layer.3": 10097.5752, "encoder_q-layer.4": 10947.1387, "encoder_q-layer.5": 11197.0068, "encoder_q-layer.6": 12501.0459, "encoder_q-layer.7": 13876.3242, "encoder_q-layer.8": 15306.5869, "encoder_q-layer.9": 12557.5273, "epoch": 0.72, "inbatch_neg_score": 5.0035, "inbatch_pos_score": 5.6172, "learning_rate": 8.894736842105264e-06, "loss": 3.3462, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.829, "num_token_overlap": 11.7024, "num_token_query": 31.4213, "num_token_union": 65.1291, "num_word_context": 202.5282, "num_word_doc": 49.8902, "num_word_query": 23.3353, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22555.2328, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0039, "query_norm": 2.5476, "queue_k_norm": 2.6205, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4213, "sent_len_1": 66.829, "sent_len_max_0": 128.0, "sent_len_max_1": 209.8325, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166200 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3407, "doc_norm": 2.6206, "encoder_q-embeddings": 9703.3057, "encoder_q-layer.0": 6643.2236, "encoder_q-layer.1": 7111.7329, "encoder_q-layer.10": 13169.7441, "encoder_q-layer.11": 33563.1641, "encoder_q-layer.2": 8344.4521, "encoder_q-layer.3": 8112.2407, "encoder_q-layer.4": 8800.1299, "encoder_q-layer.5": 8717.1299, "encoder_q-layer.6": 9674.6748, "encoder_q-layer.7": 11111.9141, "encoder_q-layer.8": 13009.8779, "encoder_q-layer.9": 11946.6709, "epoch": 0.72, "inbatch_neg_score": 5.0047, "inbatch_pos_score": 5.6289, "learning_rate": 8.86842105263158e-06, "loss": 3.3407, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.7066, "num_token_overlap": 11.6939, "num_token_query": 31.3614, "num_token_union": 65.0566, "num_word_context": 202.0768, "num_word_doc": 49.7659, "num_word_query": 23.2876, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20808.5452, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0039, "query_norm": 2.5488, "queue_k_norm": 2.6207, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3614, "sent_len_1": 66.7066, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.455, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166300 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3406, "doc_norm": 2.6225, "encoder_q-embeddings": 9811.5205, "encoder_q-layer.0": 7120.4507, "encoder_q-layer.1": 7504.373, "encoder_q-layer.10": 14190.0986, "encoder_q-layer.11": 35366.6523, "encoder_q-layer.2": 8350.5117, "encoder_q-layer.3": 8792.0918, "encoder_q-layer.4": 9377.2412, "encoder_q-layer.5": 9541.9639, "encoder_q-layer.6": 10813.5146, "encoder_q-layer.7": 12025.6709, "encoder_q-layer.8": 14437.8428, "encoder_q-layer.9": 12953.4463, "epoch": 0.72, "inbatch_neg_score": 5.0109, "inbatch_pos_score": 5.6289, "learning_rate": 8.842105263157895e-06, "loss": 3.3406, "norm_diff": 0.0714, "norm_loss": 0.0, "num_token_doc": 66.6914, "num_token_overlap": 11.7033, "num_token_query": 31.4044, "num_token_union": 65.0589, "num_word_context": 202.0952, "num_word_doc": 49.8179, "num_word_query": 23.3253, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22294.136, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0078, "query_norm": 2.5511, "queue_k_norm": 2.6216, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4044, "sent_len_1": 66.6914, "sent_len_max_0": 127.9613, "sent_len_max_1": 204.8038, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166400 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3558, "doc_norm": 2.623, "encoder_q-embeddings": 9549.0068, "encoder_q-layer.0": 6692.4141, "encoder_q-layer.1": 7227.0112, "encoder_q-layer.10": 13830.0684, "encoder_q-layer.11": 33807.6328, "encoder_q-layer.2": 7862.9434, "encoder_q-layer.3": 8389.3047, "encoder_q-layer.4": 8801.2109, "encoder_q-layer.5": 9037.7197, "encoder_q-layer.6": 10152.168, "encoder_q-layer.7": 11296.4023, "encoder_q-layer.8": 13910.1045, "encoder_q-layer.9": 12766.9727, "epoch": 0.72, "inbatch_neg_score": 5.0143, "inbatch_pos_score": 5.6172, "learning_rate": 8.81578947368421e-06, "loss": 3.3558, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.7291, "num_token_overlap": 11.6816, "num_token_query": 31.3796, "num_token_union": 65.0689, "num_word_context": 202.2601, "num_word_doc": 49.7905, "num_word_query": 23.3039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21294.4472, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0156, "query_norm": 2.5462, "queue_k_norm": 2.6217, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3796, "sent_len_1": 66.7291, "sent_len_max_0": 127.995, "sent_len_max_1": 211.3288, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166500 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3409, "doc_norm": 2.6228, "encoder_q-embeddings": 9405.5469, "encoder_q-layer.0": 6666.96, "encoder_q-layer.1": 6907.8018, "encoder_q-layer.10": 12786.165, "encoder_q-layer.11": 32444.4766, "encoder_q-layer.2": 7800.8003, "encoder_q-layer.3": 7931.7236, "encoder_q-layer.4": 8594.626, "encoder_q-layer.5": 8743.5312, "encoder_q-layer.6": 9978.3086, "encoder_q-layer.7": 11114.8711, "encoder_q-layer.8": 13515.0752, "encoder_q-layer.9": 12169.9199, "epoch": 0.72, "inbatch_neg_score": 5.0171, "inbatch_pos_score": 5.6367, "learning_rate": 8.789473684210527e-06, "loss": 3.3409, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.764, "num_token_overlap": 11.6931, "num_token_query": 31.4661, "num_token_union": 65.1648, "num_word_context": 202.4532, "num_word_doc": 49.8256, "num_word_query": 23.3845, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20474.9012, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0156, "query_norm": 2.5484, "queue_k_norm": 2.6218, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4661, "sent_len_1": 66.764, "sent_len_max_0": 127.985, "sent_len_max_1": 208.3675, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166600 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.3396, "doc_norm": 2.6219, "encoder_q-embeddings": 9275.0244, "encoder_q-layer.0": 6551.769, "encoder_q-layer.1": 7007.3999, "encoder_q-layer.10": 14828.4746, "encoder_q-layer.11": 35938.5977, "encoder_q-layer.2": 8161.8896, "encoder_q-layer.3": 8357.4141, "encoder_q-layer.4": 8784.583, "encoder_q-layer.5": 9406.2363, "encoder_q-layer.6": 10373.2344, "encoder_q-layer.7": 12051.834, "encoder_q-layer.8": 15214.543, "encoder_q-layer.9": 14471.9268, "epoch": 0.72, "inbatch_neg_score": 5.0198, "inbatch_pos_score": 5.6289, "learning_rate": 8.763157894736842e-06, "loss": 3.3396, "norm_diff": 0.0688, "norm_loss": 0.0, "num_token_doc": 66.6944, "num_token_overlap": 11.6565, "num_token_query": 31.3229, "num_token_union": 65.0795, "num_word_context": 202.3698, "num_word_doc": 49.7586, "num_word_query": 23.2485, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21833.115, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0195, "query_norm": 2.5531, "queue_k_norm": 2.6223, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3229, "sent_len_1": 66.6944, "sent_len_max_0": 127.9488, "sent_len_max_1": 208.4613, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 166700 }, { "accuracy": 52.9297, "active_queue_size": 16384.0, "cl_loss": 3.3424, "doc_norm": 2.6209, "encoder_q-embeddings": 4629.0093, "encoder_q-layer.0": 3262.5532, "encoder_q-layer.1": 3499.7295, "encoder_q-layer.10": 6957.002, "encoder_q-layer.11": 16243.0703, "encoder_q-layer.2": 3976.4995, "encoder_q-layer.3": 4086.0051, "encoder_q-layer.4": 4239.6323, "encoder_q-layer.5": 4445.7734, "encoder_q-layer.6": 4942.2666, "encoder_q-layer.7": 5427.437, "encoder_q-layer.8": 6761.8477, "encoder_q-layer.9": 6197.5547, "epoch": 0.72, "inbatch_neg_score": 5.0238, "inbatch_pos_score": 5.6367, "learning_rate": 8.736842105263158e-06, "loss": 3.3424, "norm_diff": 0.068, "norm_loss": 0.0, "num_token_doc": 66.7569, "num_token_overlap": 11.6669, "num_token_query": 31.3975, "num_token_union": 65.1278, "num_word_context": 202.4508, "num_word_doc": 49.7953, "num_word_query": 23.3295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10117.1424, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0234, "query_norm": 2.5529, "queue_k_norm": 2.6242, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3975, "sent_len_1": 66.7569, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.8675, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 166800 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3396, "doc_norm": 2.623, "encoder_q-embeddings": 4790.7231, "encoder_q-layer.0": 3527.3291, "encoder_q-layer.1": 3715.1074, "encoder_q-layer.10": 6596.0356, "encoder_q-layer.11": 16424.1855, "encoder_q-layer.2": 4113.4326, "encoder_q-layer.3": 4183.7095, "encoder_q-layer.4": 4478.2681, "encoder_q-layer.5": 4559.4985, "encoder_q-layer.6": 4897.8242, "encoder_q-layer.7": 5503.9917, "encoder_q-layer.8": 6562.4102, "encoder_q-layer.9": 6336.9102, "epoch": 0.72, "inbatch_neg_score": 5.0277, "inbatch_pos_score": 5.6406, "learning_rate": 8.710526315789475e-06, "loss": 3.3396, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 67.002, "num_token_overlap": 11.7205, "num_token_query": 31.3955, "num_token_union": 65.2399, "num_word_context": 202.3032, "num_word_doc": 50.0214, "num_word_query": 23.3166, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10420.1414, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0273, "query_norm": 2.5504, "queue_k_norm": 2.6246, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3955, "sent_len_1": 67.002, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.3113, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 166900 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.357, "doc_norm": 2.6301, "encoder_q-embeddings": 4765.9731, "encoder_q-layer.0": 3265.1343, "encoder_q-layer.1": 3502.8164, "encoder_q-layer.10": 7171.1216, "encoder_q-layer.11": 16666.0488, "encoder_q-layer.2": 3907.3489, "encoder_q-layer.3": 3917.2961, "encoder_q-layer.4": 4252.2622, "encoder_q-layer.5": 4495.3608, "encoder_q-layer.6": 5178.1919, "encoder_q-layer.7": 5696.7573, "encoder_q-layer.8": 7202.7715, "encoder_q-layer.9": 6684.833, "epoch": 0.72, "inbatch_neg_score": 5.0288, "inbatch_pos_score": 5.6484, "learning_rate": 8.68421052631579e-06, "loss": 3.357, "norm_diff": 0.0803, "norm_loss": 0.0, "num_token_doc": 66.7564, "num_token_overlap": 11.6588, "num_token_query": 31.3491, "num_token_union": 65.0964, "num_word_context": 202.1484, "num_word_doc": 49.805, "num_word_query": 23.28, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10480.4869, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0312, "query_norm": 2.5498, "queue_k_norm": 2.6243, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3491, "sent_len_1": 66.7564, "sent_len_max_0": 127.97, "sent_len_max_1": 208.0163, "stdk": 0.0499, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 167000 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3495, "doc_norm": 2.6248, "encoder_q-embeddings": 4652.0225, "encoder_q-layer.0": 3251.9624, "encoder_q-layer.1": 3488.853, "encoder_q-layer.10": 7033.5542, "encoder_q-layer.11": 17208.6504, "encoder_q-layer.2": 3788.8281, "encoder_q-layer.3": 3949.0476, "encoder_q-layer.4": 4170.1021, "encoder_q-layer.5": 4250.2754, "encoder_q-layer.6": 4990.2466, "encoder_q-layer.7": 5783.8311, "encoder_q-layer.8": 7491.2334, "encoder_q-layer.9": 6749.0283, "epoch": 0.73, "inbatch_neg_score": 5.034, "inbatch_pos_score": 5.6484, "learning_rate": 8.657894736842105e-06, "loss": 3.3495, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.7887, "num_token_overlap": 11.6419, "num_token_query": 31.2595, "num_token_union": 65.0559, "num_word_context": 202.094, "num_word_doc": 49.8355, "num_word_query": 23.2134, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10660.8442, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0352, "query_norm": 2.5545, "queue_k_norm": 2.6249, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2595, "sent_len_1": 66.7887, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.47, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 167100 }, { "accuracy": 53.2227, "active_queue_size": 16384.0, "cl_loss": 3.3368, "doc_norm": 2.6267, "encoder_q-embeddings": 4885.3398, "encoder_q-layer.0": 3445.3291, "encoder_q-layer.1": 3674.1636, "encoder_q-layer.10": 7023.6279, "encoder_q-layer.11": 17214.9336, "encoder_q-layer.2": 4059.6006, "encoder_q-layer.3": 4176.5166, "encoder_q-layer.4": 4479.7744, "encoder_q-layer.5": 4679.6826, "encoder_q-layer.6": 5316.4951, "encoder_q-layer.7": 5842.0767, "encoder_q-layer.8": 6889.8369, "encoder_q-layer.9": 6502.2856, "epoch": 0.73, "inbatch_neg_score": 5.0369, "inbatch_pos_score": 5.6562, "learning_rate": 8.631578947368422e-06, "loss": 3.3368, "norm_diff": 0.071, "norm_loss": 0.0, "num_token_doc": 66.7212, "num_token_overlap": 11.7174, "num_token_query": 31.5344, "num_token_union": 65.1379, "num_word_context": 202.0999, "num_word_doc": 49.7885, "num_word_query": 23.4176, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10777.1692, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0352, "query_norm": 2.5557, "queue_k_norm": 2.6248, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5344, "sent_len_1": 66.7212, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.6075, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 167200 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3578, "doc_norm": 2.6274, "encoder_q-embeddings": 4785.6494, "encoder_q-layer.0": 3322.8406, "encoder_q-layer.1": 3547.0972, "encoder_q-layer.10": 6498.3516, "encoder_q-layer.11": 16305.0322, "encoder_q-layer.2": 4089.1213, "encoder_q-layer.3": 4365.9263, "encoder_q-layer.4": 4615.9243, "encoder_q-layer.5": 4823.5405, "encoder_q-layer.6": 5496.7075, "encoder_q-layer.7": 6048.5049, "encoder_q-layer.8": 7071.9146, "encoder_q-layer.9": 6446.6689, "epoch": 0.73, "inbatch_neg_score": 5.0381, "inbatch_pos_score": 5.6641, "learning_rate": 8.605263157894737e-06, "loss": 3.3578, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.706, "num_token_overlap": 11.6479, "num_token_query": 31.2697, "num_token_union": 65.0411, "num_word_context": 202.0514, "num_word_doc": 49.7752, "num_word_query": 23.2299, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10530.4365, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0391, "query_norm": 2.556, "queue_k_norm": 2.6267, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2697, "sent_len_1": 66.706, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.36, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167300 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3492, "doc_norm": 2.6281, "encoder_q-embeddings": 6288.7441, "encoder_q-layer.0": 4474.3638, "encoder_q-layer.1": 5073.2817, "encoder_q-layer.10": 6558.6348, "encoder_q-layer.11": 16758.2852, "encoder_q-layer.2": 5538.7583, "encoder_q-layer.3": 5366.1074, "encoder_q-layer.4": 5387.2705, "encoder_q-layer.5": 5176.4712, "encoder_q-layer.6": 5483.4595, "encoder_q-layer.7": 5927.8408, "encoder_q-layer.8": 6868.8462, "encoder_q-layer.9": 6026.479, "epoch": 0.73, "inbatch_neg_score": 5.0429, "inbatch_pos_score": 5.6523, "learning_rate": 8.578947368421053e-06, "loss": 3.3492, "norm_diff": 0.0696, "norm_loss": 0.0, "num_token_doc": 66.7599, "num_token_overlap": 11.6721, "num_token_query": 31.3216, "num_token_union": 65.0998, "num_word_context": 202.5005, "num_word_doc": 49.8267, "num_word_query": 23.258, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11328.1512, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.043, "query_norm": 2.5585, "queue_k_norm": 2.6276, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3216, "sent_len_1": 66.7599, "sent_len_max_0": 127.975, "sent_len_max_1": 208.425, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167400 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3474, "doc_norm": 2.627, "encoder_q-embeddings": 4688.4512, "encoder_q-layer.0": 3282.1262, "encoder_q-layer.1": 3413.2791, "encoder_q-layer.10": 6572.5791, "encoder_q-layer.11": 16641.0273, "encoder_q-layer.2": 3811.499, "encoder_q-layer.3": 3916.0618, "encoder_q-layer.4": 4265.5156, "encoder_q-layer.5": 4509.4888, "encoder_q-layer.6": 4909.0898, "encoder_q-layer.7": 5709.709, "encoder_q-layer.8": 6893.7144, "encoder_q-layer.9": 6535.5347, "epoch": 0.73, "inbatch_neg_score": 5.0465, "inbatch_pos_score": 5.6562, "learning_rate": 8.552631578947368e-06, "loss": 3.3474, "norm_diff": 0.071, "norm_loss": 0.0, "num_token_doc": 66.8401, "num_token_overlap": 11.6996, "num_token_query": 31.4152, "num_token_union": 65.1404, "num_word_context": 202.4036, "num_word_doc": 49.8465, "num_word_query": 23.3469, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10460.8001, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0469, "query_norm": 2.5561, "queue_k_norm": 2.6279, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4152, "sent_len_1": 66.8401, "sent_len_max_0": 127.98, "sent_len_max_1": 210.4412, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167500 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3544, "doc_norm": 2.6278, "encoder_q-embeddings": 4682.1572, "encoder_q-layer.0": 3288.5925, "encoder_q-layer.1": 3569.2009, "encoder_q-layer.10": 7680.9453, "encoder_q-layer.11": 17665.4512, "encoder_q-layer.2": 4018.6179, "encoder_q-layer.3": 4136.0684, "encoder_q-layer.4": 4518.6743, "encoder_q-layer.5": 4632.2646, "encoder_q-layer.6": 5222.1328, "encoder_q-layer.7": 6000.9321, "encoder_q-layer.8": 7448.6465, "encoder_q-layer.9": 6997.5874, "epoch": 0.73, "inbatch_neg_score": 5.0482, "inbatch_pos_score": 5.6641, "learning_rate": 8.526315789473685e-06, "loss": 3.3544, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.9062, "num_token_overlap": 11.7073, "num_token_query": 31.4408, "num_token_union": 65.2568, "num_word_context": 202.6137, "num_word_doc": 49.9489, "num_word_query": 23.3534, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10932.5741, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0469, "query_norm": 2.5523, "queue_k_norm": 2.6287, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4408, "sent_len_1": 66.9062, "sent_len_max_0": 128.0, "sent_len_max_1": 207.4638, "stdk": 0.0495, "stdq": 0.0427, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167600 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3443, "doc_norm": 2.6278, "encoder_q-embeddings": 4814.666, "encoder_q-layer.0": 3387.9739, "encoder_q-layer.1": 3664.9597, "encoder_q-layer.10": 7112.4146, "encoder_q-layer.11": 17774.4473, "encoder_q-layer.2": 4164.1274, "encoder_q-layer.3": 4372.147, "encoder_q-layer.4": 4660.1123, "encoder_q-layer.5": 4758.0957, "encoder_q-layer.6": 5524.9873, "encoder_q-layer.7": 6085.4707, "encoder_q-layer.8": 7774.3213, "encoder_q-layer.9": 7051.002, "epoch": 0.73, "inbatch_neg_score": 5.0538, "inbatch_pos_score": 5.6602, "learning_rate": 8.500000000000002e-06, "loss": 3.3443, "norm_diff": 0.0696, "norm_loss": 0.0, "num_token_doc": 66.8083, "num_token_overlap": 11.6818, "num_token_query": 31.3447, "num_token_union": 65.0961, "num_word_context": 202.3961, "num_word_doc": 49.8562, "num_word_query": 23.269, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11266.1264, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0547, "query_norm": 2.5582, "queue_k_norm": 2.6291, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3447, "sent_len_1": 66.8083, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.07, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167700 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3431, "doc_norm": 2.63, "encoder_q-embeddings": 4793.7378, "encoder_q-layer.0": 3404.1826, "encoder_q-layer.1": 3553.8247, "encoder_q-layer.10": 6677.2412, "encoder_q-layer.11": 16653.7422, "encoder_q-layer.2": 3967.6799, "encoder_q-layer.3": 4034.8274, "encoder_q-layer.4": 4317.6021, "encoder_q-layer.5": 4463.3799, "encoder_q-layer.6": 5079.4824, "encoder_q-layer.7": 5920.9663, "encoder_q-layer.8": 6914.9634, "encoder_q-layer.9": 6589.0889, "epoch": 0.73, "inbatch_neg_score": 5.0586, "inbatch_pos_score": 5.6719, "learning_rate": 8.473684210526315e-06, "loss": 3.3431, "norm_diff": 0.0711, "norm_loss": 0.0, "num_token_doc": 66.8197, "num_token_overlap": 11.6779, "num_token_query": 31.3631, "num_token_union": 65.1228, "num_word_context": 202.5495, "num_word_doc": 49.8552, "num_word_query": 23.2874, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10511.0101, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0586, "query_norm": 2.5589, "queue_k_norm": 2.6299, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3631, "sent_len_1": 66.8197, "sent_len_max_0": 127.99, "sent_len_max_1": 209.595, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 167800 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3449, "doc_norm": 2.6335, "encoder_q-embeddings": 4793.2461, "encoder_q-layer.0": 3290.7007, "encoder_q-layer.1": 3485.6509, "encoder_q-layer.10": 7301.9604, "encoder_q-layer.11": 17213.7207, "encoder_q-layer.2": 3845.8621, "encoder_q-layer.3": 3998.5505, "encoder_q-layer.4": 4201.0571, "encoder_q-layer.5": 4355.1812, "encoder_q-layer.6": 5093.125, "encoder_q-layer.7": 5757.9517, "encoder_q-layer.8": 6990.7563, "encoder_q-layer.9": 6518.9072, "epoch": 0.73, "inbatch_neg_score": 5.0629, "inbatch_pos_score": 5.6875, "learning_rate": 8.447368421052632e-06, "loss": 3.3449, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.6753, "num_token_overlap": 11.6627, "num_token_query": 31.3371, "num_token_union": 65.0547, "num_word_context": 202.0861, "num_word_doc": 49.7596, "num_word_query": 23.2792, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10830.327, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0625, "query_norm": 2.5631, "queue_k_norm": 2.6292, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3371, "sent_len_1": 66.6753, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.135, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 167900 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3424, "doc_norm": 2.63, "encoder_q-embeddings": 4505.9175, "encoder_q-layer.0": 3265.7317, "encoder_q-layer.1": 3389.2473, "encoder_q-layer.10": 6583.6367, "encoder_q-layer.11": 16330.3457, "encoder_q-layer.2": 3781.8098, "encoder_q-layer.3": 3884.9575, "encoder_q-layer.4": 4133.4854, "encoder_q-layer.5": 4132.7202, "encoder_q-layer.6": 4682.4355, "encoder_q-layer.7": 5436.8506, "encoder_q-layer.8": 6499.689, "encoder_q-layer.9": 6105.1523, "epoch": 0.73, "inbatch_neg_score": 5.0662, "inbatch_pos_score": 5.6836, "learning_rate": 8.421052631578948e-06, "loss": 3.3424, "norm_diff": 0.0716, "norm_loss": 0.0, "num_token_doc": 66.7388, "num_token_overlap": 11.6706, "num_token_query": 31.4086, "num_token_union": 65.142, "num_word_context": 202.2511, "num_word_doc": 49.8172, "num_word_query": 23.3366, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10174.7457, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0664, "query_norm": 2.5584, "queue_k_norm": 2.6299, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4086, "sent_len_1": 66.7388, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.685, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 168000 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.6318, "encoder_q-embeddings": 4662.2183, "encoder_q-layer.0": 3347.8003, "encoder_q-layer.1": 3613.7175, "encoder_q-layer.10": 7615.1787, "encoder_q-layer.11": 17753.875, "encoder_q-layer.2": 3896.9492, "encoder_q-layer.3": 4016.7588, "encoder_q-layer.4": 4288.4336, "encoder_q-layer.5": 4476.7051, "encoder_q-layer.6": 4955.7529, "encoder_q-layer.7": 5827.0776, "encoder_q-layer.8": 7079.1841, "encoder_q-layer.9": 6677.8701, "epoch": 0.73, "inbatch_neg_score": 5.0706, "inbatch_pos_score": 5.6953, "learning_rate": 8.394736842105263e-06, "loss": 3.3508, "norm_diff": 0.0693, "norm_loss": 0.0, "num_token_doc": 66.7512, "num_token_overlap": 11.6308, "num_token_query": 31.2482, "num_token_union": 65.0452, "num_word_context": 202.1384, "num_word_doc": 49.8014, "num_word_query": 23.2049, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10871.3948, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0703, "query_norm": 2.5625, "queue_k_norm": 2.6311, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2482, "sent_len_1": 66.7512, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.705, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 168100 }, { "accuracy": 50.2441, "active_queue_size": 16384.0, "cl_loss": 3.347, "doc_norm": 2.6317, "encoder_q-embeddings": 5484.1758, "encoder_q-layer.0": 3905.6047, "encoder_q-layer.1": 4093.5503, "encoder_q-layer.10": 7621.9434, "encoder_q-layer.11": 17098.6367, "encoder_q-layer.2": 4693.4155, "encoder_q-layer.3": 4791.6162, "encoder_q-layer.4": 5000.3267, "encoder_q-layer.5": 5161.6431, "encoder_q-layer.6": 5572.6841, "encoder_q-layer.7": 6304.2427, "encoder_q-layer.8": 6842.1924, "encoder_q-layer.9": 6153.0464, "epoch": 0.73, "inbatch_neg_score": 5.0734, "inbatch_pos_score": 5.668, "learning_rate": 8.36842105263158e-06, "loss": 3.347, "norm_diff": 0.0746, "norm_loss": 0.0, "num_token_doc": 66.9279, "num_token_overlap": 11.6606, "num_token_query": 31.4203, "num_token_union": 65.2454, "num_word_context": 202.4571, "num_word_doc": 49.9418, "num_word_query": 23.3377, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11176.5366, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0742, "query_norm": 2.5571, "queue_k_norm": 2.6327, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4203, "sent_len_1": 66.9279, "sent_len_max_0": 127.9663, "sent_len_max_1": 210.1062, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168200 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3362, "doc_norm": 2.635, "encoder_q-embeddings": 5829.6714, "encoder_q-layer.0": 4072.3577, "encoder_q-layer.1": 4396.6372, "encoder_q-layer.10": 6683.2231, "encoder_q-layer.11": 16888.9062, "encoder_q-layer.2": 5051.5898, "encoder_q-layer.3": 5332.0303, "encoder_q-layer.4": 5485.001, "encoder_q-layer.5": 5689.0234, "encoder_q-layer.6": 6442.6807, "encoder_q-layer.7": 6712.4214, "encoder_q-layer.8": 7562.4873, "encoder_q-layer.9": 6720.9624, "epoch": 0.73, "inbatch_neg_score": 5.0769, "inbatch_pos_score": 5.6914, "learning_rate": 8.342105263157895e-06, "loss": 3.3362, "norm_diff": 0.0694, "norm_loss": 0.0, "num_token_doc": 66.6838, "num_token_overlap": 11.6987, "num_token_query": 31.482, "num_token_union": 65.1203, "num_word_context": 202.2227, "num_word_doc": 49.7643, "num_word_query": 23.3841, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11476.9956, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0781, "query_norm": 2.5656, "queue_k_norm": 2.632, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.482, "sent_len_1": 66.6838, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.7862, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 168300 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3401, "doc_norm": 2.6319, "encoder_q-embeddings": 5164.2305, "encoder_q-layer.0": 3641.4028, "encoder_q-layer.1": 3997.6226, "encoder_q-layer.10": 7302.6401, "encoder_q-layer.11": 16892.4512, "encoder_q-layer.2": 4554.1255, "encoder_q-layer.3": 4638.9106, "encoder_q-layer.4": 5184.8916, "encoder_q-layer.5": 5235.9019, "encoder_q-layer.6": 5873.6152, "encoder_q-layer.7": 6501.3613, "encoder_q-layer.8": 7640.5889, "encoder_q-layer.9": 6887.2524, "epoch": 0.73, "inbatch_neg_score": 5.0811, "inbatch_pos_score": 5.707, "learning_rate": 8.315789473684212e-06, "loss": 3.3401, "norm_diff": 0.0641, "norm_loss": 0.0, "num_token_doc": 66.8231, "num_token_overlap": 11.6862, "num_token_query": 31.3779, "num_token_union": 65.1226, "num_word_context": 202.2779, "num_word_doc": 49.8663, "num_word_query": 23.2953, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10911.2451, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.082, "query_norm": 2.5679, "queue_k_norm": 2.6327, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3779, "sent_len_1": 66.8231, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.4675, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 168400 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3463, "doc_norm": 2.6334, "encoder_q-embeddings": 4655.2246, "encoder_q-layer.0": 3353.4778, "encoder_q-layer.1": 3558.6221, "encoder_q-layer.10": 6564.2046, "encoder_q-layer.11": 16341.2139, "encoder_q-layer.2": 3848.9817, "encoder_q-layer.3": 3906.2698, "encoder_q-layer.4": 4313.5171, "encoder_q-layer.5": 4310.771, "encoder_q-layer.6": 4904.52, "encoder_q-layer.7": 5584.9717, "encoder_q-layer.8": 6725.2529, "encoder_q-layer.9": 6084.9043, "epoch": 0.73, "inbatch_neg_score": 5.084, "inbatch_pos_score": 5.6953, "learning_rate": 8.289473684210526e-06, "loss": 3.3463, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.8428, "num_token_overlap": 11.7063, "num_token_query": 31.4583, "num_token_union": 65.1821, "num_word_context": 202.4531, "num_word_doc": 49.8657, "num_word_query": 23.3756, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10325.3327, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0859, "query_norm": 2.5599, "queue_k_norm": 2.6343, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4583, "sent_len_1": 66.8428, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.685, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168500 }, { "accuracy": 49.0723, "active_queue_size": 16384.0, "cl_loss": 3.3456, "doc_norm": 2.6345, "encoder_q-embeddings": 4939.334, "encoder_q-layer.0": 3527.4856, "encoder_q-layer.1": 3771.2073, "encoder_q-layer.10": 6876.6099, "encoder_q-layer.11": 16869.1113, "encoder_q-layer.2": 4173.6226, "encoder_q-layer.3": 4060.5503, "encoder_q-layer.4": 4177.8979, "encoder_q-layer.5": 4315.3096, "encoder_q-layer.6": 4797.5938, "encoder_q-layer.7": 5504.2524, "encoder_q-layer.8": 6985.4038, "encoder_q-layer.9": 6515.9004, "epoch": 0.73, "inbatch_neg_score": 5.09, "inbatch_pos_score": 5.6953, "learning_rate": 8.263157894736841e-06, "loss": 3.3456, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.666, "num_token_overlap": 11.6549, "num_token_query": 31.3508, "num_token_union": 65.0304, "num_word_context": 201.9942, "num_word_doc": 49.7376, "num_word_query": 23.2872, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10576.204, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0898, "query_norm": 2.5659, "queue_k_norm": 2.6356, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3508, "sent_len_1": 66.666, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.4038, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168600 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.352, "doc_norm": 2.6353, "encoder_q-embeddings": 4712.4541, "encoder_q-layer.0": 3404.415, "encoder_q-layer.1": 3540.2493, "encoder_q-layer.10": 7165.8804, "encoder_q-layer.11": 16823.4668, "encoder_q-layer.2": 3957.7732, "encoder_q-layer.3": 4273.3379, "encoder_q-layer.4": 4516.0645, "encoder_q-layer.5": 4794.5366, "encoder_q-layer.6": 5288.7241, "encoder_q-layer.7": 5783.647, "encoder_q-layer.8": 7073.2617, "encoder_q-layer.9": 6367.2188, "epoch": 0.73, "inbatch_neg_score": 5.0902, "inbatch_pos_score": 5.7109, "learning_rate": 8.236842105263158e-06, "loss": 3.352, "norm_diff": 0.0662, "norm_loss": 0.0, "num_token_doc": 66.761, "num_token_overlap": 11.6844, "num_token_query": 31.3906, "num_token_union": 65.1272, "num_word_context": 202.3955, "num_word_doc": 49.8549, "num_word_query": 23.3169, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10591.0442, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.0898, "query_norm": 2.5691, "queue_k_norm": 2.6355, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3906, "sent_len_1": 66.761, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.755, "stdk": 0.0495, "stdq": 0.0436, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168700 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3631, "doc_norm": 2.6349, "encoder_q-embeddings": 9732.9902, "encoder_q-layer.0": 7066.6377, "encoder_q-layer.1": 7320.8794, "encoder_q-layer.10": 13809.3154, "encoder_q-layer.11": 34375.2891, "encoder_q-layer.2": 8315.8594, "encoder_q-layer.3": 8626.9043, "encoder_q-layer.4": 9477.959, "encoder_q-layer.5": 9838.0625, "encoder_q-layer.6": 10800.3408, "encoder_q-layer.7": 12033.7275, "encoder_q-layer.8": 15004.792, "encoder_q-layer.9": 13252.4268, "epoch": 0.73, "inbatch_neg_score": 5.0919, "inbatch_pos_score": 5.7031, "learning_rate": 8.210526315789475e-06, "loss": 3.3631, "norm_diff": 0.0659, "norm_loss": 0.0, "num_token_doc": 66.8481, "num_token_overlap": 11.6509, "num_token_query": 31.3726, "num_token_union": 65.1968, "num_word_context": 202.6543, "num_word_doc": 49.8809, "num_word_query": 23.3135, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21723.0678, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0898, "query_norm": 2.569, "queue_k_norm": 2.6364, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3726, "sent_len_1": 66.8481, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.155, "stdk": 0.0494, "stdq": 0.0436, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168800 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3326, "doc_norm": 2.6368, "encoder_q-embeddings": 9993.5176, "encoder_q-layer.0": 6892.4131, "encoder_q-layer.1": 7343.3496, "encoder_q-layer.10": 14667.6826, "encoder_q-layer.11": 34934.3789, "encoder_q-layer.2": 8314.4121, "encoder_q-layer.3": 8498.3896, "encoder_q-layer.4": 9098.1787, "encoder_q-layer.5": 9543.4199, "encoder_q-layer.6": 10733.2021, "encoder_q-layer.7": 12414.9238, "encoder_q-layer.8": 14313.1523, "encoder_q-layer.9": 13361.8926, "epoch": 0.73, "inbatch_neg_score": 5.0933, "inbatch_pos_score": 5.7109, "learning_rate": 8.18421052631579e-06, "loss": 3.3326, "norm_diff": 0.0705, "norm_loss": 0.0, "num_token_doc": 66.804, "num_token_overlap": 11.6815, "num_token_query": 31.3461, "num_token_union": 65.1179, "num_word_context": 202.3092, "num_word_doc": 49.8439, "num_word_query": 23.294, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22151.2283, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0938, "query_norm": 2.5663, "queue_k_norm": 2.6375, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3461, "sent_len_1": 66.804, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.8262, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 168900 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3466, "doc_norm": 2.6353, "encoder_q-embeddings": 9400.9824, "encoder_q-layer.0": 6736.6787, "encoder_q-layer.1": 7273.5093, "encoder_q-layer.10": 12977.2422, "encoder_q-layer.11": 34232.9023, "encoder_q-layer.2": 8154.7925, "encoder_q-layer.3": 8245.7363, "encoder_q-layer.4": 8668.3203, "encoder_q-layer.5": 8924.0967, "encoder_q-layer.6": 9762.2188, "encoder_q-layer.7": 11164.1318, "encoder_q-layer.8": 13040.7227, "encoder_q-layer.9": 12570.2168, "epoch": 0.73, "inbatch_neg_score": 5.096, "inbatch_pos_score": 5.7031, "learning_rate": 8.157894736842106e-06, "loss": 3.3466, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.6948, "num_token_overlap": 11.6714, "num_token_query": 31.4251, "num_token_union": 65.11, "num_word_context": 201.9314, "num_word_doc": 49.7696, "num_word_query": 23.338, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20875.9555, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0938, "query_norm": 2.5604, "queue_k_norm": 2.6377, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4251, "sent_len_1": 66.6948, "sent_len_max_0": 127.96, "sent_len_max_1": 209.095, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 169000 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.3588, "doc_norm": 2.6388, "encoder_q-embeddings": 9547.8369, "encoder_q-layer.0": 6807.5503, "encoder_q-layer.1": 7168.3901, "encoder_q-layer.10": 14293.7256, "encoder_q-layer.11": 33997.5977, "encoder_q-layer.2": 7955.7534, "encoder_q-layer.3": 8155.5654, "encoder_q-layer.4": 8755.6924, "encoder_q-layer.5": 9336.3623, "encoder_q-layer.6": 10249.9668, "encoder_q-layer.7": 11913.3525, "encoder_q-layer.8": 13736.4482, "encoder_q-layer.9": 12789.292, "epoch": 0.73, "inbatch_neg_score": 5.0964, "inbatch_pos_score": 5.7148, "learning_rate": 8.131578947368421e-06, "loss": 3.3588, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.7581, "num_token_overlap": 11.6481, "num_token_query": 31.3317, "num_token_union": 65.0819, "num_word_context": 202.0838, "num_word_doc": 49.7671, "num_word_query": 23.2614, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21460.8853, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.0977, "query_norm": 2.5587, "queue_k_norm": 2.6372, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3317, "sent_len_1": 66.7581, "sent_len_max_0": 127.9875, "sent_len_max_1": 212.2175, "stdk": 0.0496, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 169100 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3523, "doc_norm": 2.6386, "encoder_q-embeddings": 9625.9082, "encoder_q-layer.0": 6758.8184, "encoder_q-layer.1": 7154.0005, "encoder_q-layer.10": 13601.9277, "encoder_q-layer.11": 35375.9219, "encoder_q-layer.2": 8326.7217, "encoder_q-layer.3": 8712.7354, "encoder_q-layer.4": 9248.4229, "encoder_q-layer.5": 9578.9785, "encoder_q-layer.6": 10557.3076, "encoder_q-layer.7": 12224.9482, "encoder_q-layer.8": 14098.2852, "encoder_q-layer.9": 12874.5742, "epoch": 0.73, "inbatch_neg_score": 5.0995, "inbatch_pos_score": 5.7148, "learning_rate": 8.105263157894736e-06, "loss": 3.3523, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.9319, "num_token_overlap": 11.7055, "num_token_query": 31.4823, "num_token_union": 65.2323, "num_word_context": 202.3065, "num_word_doc": 49.901, "num_word_query": 23.3884, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22106.0073, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1016, "query_norm": 2.5619, "queue_k_norm": 2.6378, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4823, "sent_len_1": 66.9319, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.5475, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 169200 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.345, "doc_norm": 2.6384, "encoder_q-embeddings": 8841.2744, "encoder_q-layer.0": 6260.2476, "encoder_q-layer.1": 6710.8677, "encoder_q-layer.10": 13391.791, "encoder_q-layer.11": 33907.6484, "encoder_q-layer.2": 7483.0908, "encoder_q-layer.3": 7694.5986, "encoder_q-layer.4": 8151.5273, "encoder_q-layer.5": 8358.7246, "encoder_q-layer.6": 9529.1475, "encoder_q-layer.7": 11474.9248, "encoder_q-layer.8": 13206.2393, "encoder_q-layer.9": 12679.0977, "epoch": 0.73, "inbatch_neg_score": 5.1033, "inbatch_pos_score": 5.7188, "learning_rate": 8.078947368421053e-06, "loss": 3.345, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.8043, "num_token_overlap": 11.6582, "num_token_query": 31.2997, "num_token_union": 65.0994, "num_word_context": 202.4116, "num_word_doc": 49.8646, "num_word_query": 23.2456, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20616.6454, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1016, "query_norm": 2.5627, "queue_k_norm": 2.6382, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2997, "sent_len_1": 66.8043, "sent_len_max_0": 127.985, "sent_len_max_1": 209.2312, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 169300 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.357, "doc_norm": 2.6375, "encoder_q-embeddings": 9831.9941, "encoder_q-layer.0": 6704.8882, "encoder_q-layer.1": 7350.2334, "encoder_q-layer.10": 15024.4004, "encoder_q-layer.11": 35015.207, "encoder_q-layer.2": 8578.1318, "encoder_q-layer.3": 9231.1211, "encoder_q-layer.4": 9649.6895, "encoder_q-layer.5": 9597.3252, "encoder_q-layer.6": 10526.3271, "encoder_q-layer.7": 11909.46, "encoder_q-layer.8": 13804.8428, "encoder_q-layer.9": 13650.9746, "epoch": 0.74, "inbatch_neg_score": 5.1038, "inbatch_pos_score": 5.7188, "learning_rate": 8.052631578947368e-06, "loss": 3.357, "norm_diff": 0.071, "norm_loss": 0.0, "num_token_doc": 66.7779, "num_token_overlap": 11.6671, "num_token_query": 31.3362, "num_token_union": 65.0834, "num_word_context": 202.1098, "num_word_doc": 49.8599, "num_word_query": 23.2698, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21768.7424, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1016, "query_norm": 2.5665, "queue_k_norm": 2.6398, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3362, "sent_len_1": 66.7779, "sent_len_max_0": 127.9887, "sent_len_max_1": 206.9212, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 169400 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3441, "doc_norm": 2.6366, "encoder_q-embeddings": 9273.5898, "encoder_q-layer.0": 6575.6982, "encoder_q-layer.1": 7136.6831, "encoder_q-layer.10": 13926.2617, "encoder_q-layer.11": 35040.7031, "encoder_q-layer.2": 7972.1777, "encoder_q-layer.3": 8353.8008, "encoder_q-layer.4": 9202.4971, "encoder_q-layer.5": 9400.1191, "encoder_q-layer.6": 10422.415, "encoder_q-layer.7": 11258.2393, "encoder_q-layer.8": 13339.6836, "encoder_q-layer.9": 12901.9756, "epoch": 0.74, "inbatch_neg_score": 5.1047, "inbatch_pos_score": 5.7188, "learning_rate": 8.026315789473685e-06, "loss": 3.3441, "norm_diff": 0.0649, "norm_loss": 0.0, "num_token_doc": 66.6911, "num_token_overlap": 11.6635, "num_token_query": 31.3333, "num_token_union": 65.0917, "num_word_context": 202.2782, "num_word_doc": 49.7614, "num_word_query": 23.2701, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21581.5931, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1055, "query_norm": 2.5717, "queue_k_norm": 2.6397, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3333, "sent_len_1": 66.6911, "sent_len_max_0": 127.98, "sent_len_max_1": 209.045, "stdk": 0.0493, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 169500 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3483, "doc_norm": 2.6399, "encoder_q-embeddings": 9727.6387, "encoder_q-layer.0": 6580.9243, "encoder_q-layer.1": 7084.8311, "encoder_q-layer.10": 13960.8682, "encoder_q-layer.11": 34535.6367, "encoder_q-layer.2": 7951.4316, "encoder_q-layer.3": 8061.4043, "encoder_q-layer.4": 8733.6025, "encoder_q-layer.5": 8882.9893, "encoder_q-layer.6": 10191.166, "encoder_q-layer.7": 11511.1709, "encoder_q-layer.8": 13864.25, "encoder_q-layer.9": 13378.0576, "epoch": 0.74, "inbatch_neg_score": 5.106, "inbatch_pos_score": 5.7227, "learning_rate": 8.000000000000001e-06, "loss": 3.3483, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.7894, "num_token_overlap": 11.7284, "num_token_query": 31.4807, "num_token_union": 65.1514, "num_word_context": 202.235, "num_word_doc": 49.8775, "num_word_query": 23.3854, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21568.8697, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1055, "query_norm": 2.5673, "queue_k_norm": 2.64, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4807, "sent_len_1": 66.7894, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.05, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 169600 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3273, "doc_norm": 2.6396, "encoder_q-embeddings": 9817.2129, "encoder_q-layer.0": 7041.4062, "encoder_q-layer.1": 7610.0127, "encoder_q-layer.10": 13977.7314, "encoder_q-layer.11": 32943.875, "encoder_q-layer.2": 8526.2197, "encoder_q-layer.3": 8486.5234, "encoder_q-layer.4": 9060.7188, "encoder_q-layer.5": 9271.1836, "encoder_q-layer.6": 9813.0527, "encoder_q-layer.7": 11302.2432, "encoder_q-layer.8": 13058.3428, "encoder_q-layer.9": 12371.0156, "epoch": 0.74, "inbatch_neg_score": 5.1046, "inbatch_pos_score": 5.7344, "learning_rate": 7.973684210526316e-06, "loss": 3.3273, "norm_diff": 0.0708, "norm_loss": 0.0, "num_token_doc": 66.8501, "num_token_overlap": 11.6914, "num_token_query": 31.3995, "num_token_union": 65.1525, "num_word_context": 202.193, "num_word_doc": 49.8652, "num_word_query": 23.3272, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20774.7728, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1055, "query_norm": 2.5689, "queue_k_norm": 2.6412, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3995, "sent_len_1": 66.8501, "sent_len_max_0": 128.0, "sent_len_max_1": 209.3938, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 169700 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.335, "doc_norm": 2.6433, "encoder_q-embeddings": 10363.7607, "encoder_q-layer.0": 7174.7588, "encoder_q-layer.1": 7654.4722, "encoder_q-layer.10": 12982.7266, "encoder_q-layer.11": 33892.3984, "encoder_q-layer.2": 8520.4248, "encoder_q-layer.3": 8934.5664, "encoder_q-layer.4": 9862.3359, "encoder_q-layer.5": 9723.6094, "encoder_q-layer.6": 10370.0625, "encoder_q-layer.7": 12092.4707, "encoder_q-layer.8": 14041.1348, "encoder_q-layer.9": 12534.1396, "epoch": 0.74, "inbatch_neg_score": 5.1092, "inbatch_pos_score": 5.7227, "learning_rate": 7.947368421052633e-06, "loss": 3.335, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 66.5389, "num_token_overlap": 11.6724, "num_token_query": 31.3223, "num_token_union": 64.96, "num_word_context": 202.1864, "num_word_doc": 49.6543, "num_word_query": 23.2558, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21735.6378, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1094, "query_norm": 2.5663, "queue_k_norm": 2.6406, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3223, "sent_len_1": 66.5389, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.2138, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 169800 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3413, "doc_norm": 2.6427, "encoder_q-embeddings": 9728.8994, "encoder_q-layer.0": 7093.2056, "encoder_q-layer.1": 7375.5273, "encoder_q-layer.10": 13246.793, "encoder_q-layer.11": 34643.5391, "encoder_q-layer.2": 8125.7358, "encoder_q-layer.3": 8095.7217, "encoder_q-layer.4": 8778.5137, "encoder_q-layer.5": 8810.582, "encoder_q-layer.6": 10254.8779, "encoder_q-layer.7": 11783.3994, "encoder_q-layer.8": 14082.0918, "encoder_q-layer.9": 13065.207, "epoch": 0.74, "inbatch_neg_score": 5.1117, "inbatch_pos_score": 5.7344, "learning_rate": 7.921052631578948e-06, "loss": 3.3413, "norm_diff": 0.0731, "norm_loss": 0.0, "num_token_doc": 66.8649, "num_token_overlap": 11.7064, "num_token_query": 31.5104, "num_token_union": 65.2343, "num_word_context": 202.4927, "num_word_doc": 49.8359, "num_word_query": 23.4128, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21190.2514, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1133, "query_norm": 2.5696, "queue_k_norm": 2.6416, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5104, "sent_len_1": 66.8649, "sent_len_max_0": 127.9638, "sent_len_max_1": 213.2163, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 169900 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3362, "doc_norm": 2.6417, "encoder_q-embeddings": 10257.5127, "encoder_q-layer.0": 7037.5488, "encoder_q-layer.1": 7466.2915, "encoder_q-layer.10": 13204.1621, "encoder_q-layer.11": 32953.9883, "encoder_q-layer.2": 8379.6709, "encoder_q-layer.3": 8459.3926, "encoder_q-layer.4": 9032.5566, "encoder_q-layer.5": 9060.3965, "encoder_q-layer.6": 10470.3496, "encoder_q-layer.7": 12188.0107, "encoder_q-layer.8": 13821.4805, "encoder_q-layer.9": 12933.8496, "epoch": 0.74, "inbatch_neg_score": 5.113, "inbatch_pos_score": 5.7227, "learning_rate": 7.894736842105263e-06, "loss": 3.3362, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.8329, "num_token_overlap": 11.704, "num_token_query": 31.4549, "num_token_union": 65.1871, "num_word_context": 202.3035, "num_word_doc": 49.858, "num_word_query": 23.3505, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21021.7805, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1133, "query_norm": 2.5667, "queue_k_norm": 2.6403, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4549, "sent_len_1": 66.8329, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.0075, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0494, "stdqueue_q": 0.0, "step": 170000 }, { "dev_runtime": 30.7991, "dev_samples_per_second": 1.039, "dev_steps_per_second": 0.032, "epoch": 0.74, "step": 170000, "test_accuracy": 94.15283203125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3245335519313812, "test_doc_norm": 2.629218101501465, "test_inbatch_neg_score": 5.408053398132324, "test_inbatch_pos_score": 6.443785667419434, "test_loss": 0.3245335519313812, "test_loss_align": 1.006374478340149, "test_loss_unif": -46.39924621582031, "test_loss_unif_q@queue": -46.39924621582031, "test_norm_diff": 0.011055909097194672, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.110089302062988, "test_query_norm": 2.6402738094329834, "test_queue_k_norm": 2.640366554260254, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044331010431051254, "test_stdq": 0.044303201138973236, "test_stdqueue_k": 0.049447886645793915, "test_stdqueue_q": 0.0 }, { "dev_runtime": 30.7991, "dev_samples_per_second": 1.039, "dev_steps_per_second": 0.032, "epoch": 0.74, "eval_beir-arguana_ndcg@10": 0.4108, "eval_beir-arguana_recall@10": 0.69417, "eval_beir-arguana_recall@100": 0.95092, "eval_beir-arguana_recall@20": 0.82219, "eval_beir-avg_ndcg@10": 0.4060671666666667, "eval_beir-avg_recall@10": 0.47847875, "eval_beir-avg_recall@100": 0.6558040833333333, "eval_beir-avg_recall@20": 0.5377070833333333, "eval_beir-cqadupstack_ndcg@10": 0.29461166666666666, "eval_beir-cqadupstack_recall@10": 0.39671750000000006, "eval_beir-cqadupstack_recall@100": 0.6327808333333333, "eval_beir-cqadupstack_recall@20": 0.4634108333333334, "eval_beir-fiqa_ndcg@10": 0.26973, "eval_beir-fiqa_recall@10": 0.33533, "eval_beir-fiqa_recall@100": 0.61613, "eval_beir-fiqa_recall@20": 0.40973, "eval_beir-nfcorpus_ndcg@10": 0.32794, "eval_beir-nfcorpus_recall@10": 0.16127, "eval_beir-nfcorpus_recall@100": 0.31696, "eval_beir-nfcorpus_recall@20": 0.19944, "eval_beir-nq_ndcg@10": 0.32725, "eval_beir-nq_recall@10": 0.5197, "eval_beir-nq_recall@100": 0.84593, "eval_beir-nq_recall@20": 0.63939, "eval_beir-quora_ndcg@10": 0.79217, "eval_beir-quora_recall@10": 0.89742, "eval_beir-quora_recall@100": 0.98121, "eval_beir-quora_recall@20": 0.93636, "eval_beir-scidocs_ndcg@10": 0.1677, "eval_beir-scidocs_recall@10": 0.17703, "eval_beir-scidocs_recall@100": 0.3936, "eval_beir-scidocs_recall@20": 0.23557, "eval_beir-scifact_ndcg@10": 0.68348, "eval_beir-scifact_recall@10": 0.80567, "eval_beir-scifact_recall@100": 0.92989, "eval_beir-scifact_recall@20": 0.84578, "eval_beir-trec-covid_ndcg@10": 0.61609, "eval_beir-trec-covid_recall@10": 0.664, "eval_beir-trec-covid_recall@100": 0.4672, "eval_beir-trec-covid_recall@20": 0.611, "eval_beir-webis-touche2020_ndcg@10": 0.1709, "eval_beir-webis-touche2020_recall@10": 0.13348, "eval_beir-webis-touche2020_recall@100": 0.42342, "eval_beir-webis-touche2020_recall@20": 0.2142, "eval_senteval-avg_sts": 0.733647339651895, "eval_senteval-sickr_spearman": 0.7290570761883021, "eval_senteval-stsb_spearman": 0.7382376031154878, "step": 170000, "test_accuracy": 94.15283203125, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3245335519313812, "test_doc_norm": 2.629218101501465, "test_inbatch_neg_score": 5.408053398132324, "test_inbatch_pos_score": 6.443785667419434, "test_loss": 0.3245335519313812, "test_loss_align": 1.006374478340149, "test_loss_unif": -46.39924621582031, "test_loss_unif_q@queue": -46.39924621582031, "test_norm_diff": 0.011055909097194672, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.110089302062988, "test_query_norm": 2.6402738094329834, "test_queue_k_norm": 2.640366554260254, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.044331010431051254, "test_stdq": 0.044303201138973236, "test_stdqueue_k": 0.049447886645793915, "test_stdqueue_q": 0.0 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3403, "doc_norm": 2.6409, "encoder_q-embeddings": 9468.0723, "encoder_q-layer.0": 6615.8506, "encoder_q-layer.1": 6968.9648, "encoder_q-layer.10": 13747.5, "encoder_q-layer.11": 32747.9199, "encoder_q-layer.2": 7851.1304, "encoder_q-layer.3": 8025.1611, "encoder_q-layer.4": 8863.0996, "encoder_q-layer.5": 9634.3281, "encoder_q-layer.6": 10025.8457, "encoder_q-layer.7": 11320.7051, "encoder_q-layer.8": 13716.6318, "encoder_q-layer.9": 12476.8633, "epoch": 0.74, "inbatch_neg_score": 5.1175, "inbatch_pos_score": 5.7227, "learning_rate": 7.86842105263158e-06, "loss": 3.3403, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.8224, "num_token_overlap": 11.6752, "num_token_query": 31.4138, "num_token_union": 65.217, "num_word_context": 202.3458, "num_word_doc": 49.8446, "num_word_query": 23.3236, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20632.2437, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1172, "query_norm": 2.5649, "queue_k_norm": 2.6419, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4138, "sent_len_1": 66.8224, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.3562, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 170100 }, { "accuracy": 54.4922, "active_queue_size": 16384.0, "cl_loss": 3.3501, "doc_norm": 2.6433, "encoder_q-embeddings": 9884.5352, "encoder_q-layer.0": 6749.8008, "encoder_q-layer.1": 7276.3857, "encoder_q-layer.10": 14463.7754, "encoder_q-layer.11": 34424.375, "encoder_q-layer.2": 8187.1401, "encoder_q-layer.3": 8332.2793, "encoder_q-layer.4": 8628.6924, "encoder_q-layer.5": 8745.707, "encoder_q-layer.6": 10054.3633, "encoder_q-layer.7": 11651.6836, "encoder_q-layer.8": 14632.7012, "encoder_q-layer.9": 13592.8906, "epoch": 0.74, "inbatch_neg_score": 5.118, "inbatch_pos_score": 5.7383, "learning_rate": 7.842105263157895e-06, "loss": 3.3501, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.9295, "num_token_overlap": 11.687, "num_token_query": 31.3679, "num_token_union": 65.1862, "num_word_context": 202.3938, "num_word_doc": 49.913, "num_word_query": 23.2974, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21484.5752, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1172, "query_norm": 2.5675, "queue_k_norm": 2.6428, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3679, "sent_len_1": 66.9295, "sent_len_max_0": 127.9587, "sent_len_max_1": 210.2325, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 170200 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3557, "doc_norm": 2.6446, "encoder_q-embeddings": 9903.4385, "encoder_q-layer.0": 6795.8232, "encoder_q-layer.1": 7399.606, "encoder_q-layer.10": 13718.3828, "encoder_q-layer.11": 33970.6289, "encoder_q-layer.2": 8160.2031, "encoder_q-layer.3": 8435.7275, "encoder_q-layer.4": 8822.8955, "encoder_q-layer.5": 9037.6592, "encoder_q-layer.6": 9909.4375, "encoder_q-layer.7": 11329.0566, "encoder_q-layer.8": 13317.6787, "encoder_q-layer.9": 12674.5508, "epoch": 0.74, "inbatch_neg_score": 5.1194, "inbatch_pos_score": 5.7422, "learning_rate": 7.815789473684211e-06, "loss": 3.3557, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.8535, "num_token_overlap": 11.6402, "num_token_query": 31.3805, "num_token_union": 65.1858, "num_word_context": 202.5928, "num_word_doc": 49.8776, "num_word_query": 23.3138, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21241.4036, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1211, "query_norm": 2.5706, "queue_k_norm": 2.6436, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3805, "sent_len_1": 66.8535, "sent_len_max_0": 127.995, "sent_len_max_1": 208.77, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170300 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3415, "doc_norm": 2.6432, "encoder_q-embeddings": 9561.5283, "encoder_q-layer.0": 6835.9805, "encoder_q-layer.1": 7259.6079, "encoder_q-layer.10": 13529.0469, "encoder_q-layer.11": 34723.0938, "encoder_q-layer.2": 8176.2183, "encoder_q-layer.3": 8464.3184, "encoder_q-layer.4": 9035.1416, "encoder_q-layer.5": 9600.291, "encoder_q-layer.6": 10878.4365, "encoder_q-layer.7": 12651.1865, "encoder_q-layer.8": 14717.1807, "encoder_q-layer.9": 13218.083, "epoch": 0.74, "inbatch_neg_score": 5.124, "inbatch_pos_score": 5.7422, "learning_rate": 7.789473684210528e-06, "loss": 3.3415, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.9188, "num_token_overlap": 11.7146, "num_token_query": 31.439, "num_token_union": 65.1926, "num_word_context": 202.4718, "num_word_doc": 49.9455, "num_word_query": 23.3579, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21864.7023, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.125, "query_norm": 2.5713, "queue_k_norm": 2.6443, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.439, "sent_len_1": 66.9188, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.285, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170400 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3583, "doc_norm": 2.6428, "encoder_q-embeddings": 10835.7158, "encoder_q-layer.0": 7798.2402, "encoder_q-layer.1": 8495.4424, "encoder_q-layer.10": 14779.5898, "encoder_q-layer.11": 35333.6719, "encoder_q-layer.2": 9949.042, "encoder_q-layer.3": 10343.5029, "encoder_q-layer.4": 11113.8154, "encoder_q-layer.5": 11112.2949, "encoder_q-layer.6": 12402.2617, "encoder_q-layer.7": 13480.6523, "encoder_q-layer.8": 15199.9902, "encoder_q-layer.9": 13818.3359, "epoch": 0.74, "inbatch_neg_score": 5.1277, "inbatch_pos_score": 5.7383, "learning_rate": 7.763157894736843e-06, "loss": 3.3583, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.8296, "num_token_overlap": 11.6575, "num_token_query": 31.2379, "num_token_union": 65.0867, "num_word_context": 202.3726, "num_word_doc": 49.8962, "num_word_query": 23.1961, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23213.4993, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.125, "query_norm": 2.571, "queue_k_norm": 2.6445, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2379, "sent_len_1": 66.8296, "sent_len_max_0": 127.9562, "sent_len_max_1": 208.985, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170500 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.6441, "encoder_q-embeddings": 11877.7402, "encoder_q-layer.0": 8669.6357, "encoder_q-layer.1": 9328.6104, "encoder_q-layer.10": 15142.0205, "encoder_q-layer.11": 35761.0781, "encoder_q-layer.2": 10992.9375, "encoder_q-layer.3": 11516.8389, "encoder_q-layer.4": 12240.4189, "encoder_q-layer.5": 12451.0215, "encoder_q-layer.6": 12736.2373, "encoder_q-layer.7": 14592.8184, "encoder_q-layer.8": 16024.708, "encoder_q-layer.9": 13467.9014, "epoch": 0.74, "inbatch_neg_score": 5.13, "inbatch_pos_score": 5.7539, "learning_rate": 7.736842105263158e-06, "loss": 3.3331, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.7383, "num_token_overlap": 11.6663, "num_token_query": 31.3473, "num_token_union": 65.1122, "num_word_context": 202.2303, "num_word_doc": 49.7939, "num_word_query": 23.2816, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24019.4908, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1289, "query_norm": 2.5749, "queue_k_norm": 2.645, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3473, "sent_len_1": 66.7383, "sent_len_max_0": 127.9762, "sent_len_max_1": 207.5825, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170600 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.33, "doc_norm": 2.6449, "encoder_q-embeddings": 9769.1768, "encoder_q-layer.0": 6872.8594, "encoder_q-layer.1": 7470.188, "encoder_q-layer.10": 13873.2734, "encoder_q-layer.11": 34519.2305, "encoder_q-layer.2": 8395.2393, "encoder_q-layer.3": 8990.3799, "encoder_q-layer.4": 9690.1289, "encoder_q-layer.5": 9672.8135, "encoder_q-layer.6": 10550.9648, "encoder_q-layer.7": 12688.5479, "encoder_q-layer.8": 14071.5195, "encoder_q-layer.9": 12606.4941, "epoch": 0.74, "inbatch_neg_score": 5.1328, "inbatch_pos_score": 5.7422, "learning_rate": 7.710526315789473e-06, "loss": 3.33, "norm_diff": 0.0765, "norm_loss": 0.0, "num_token_doc": 66.8054, "num_token_overlap": 11.6899, "num_token_query": 31.3984, "num_token_union": 65.1203, "num_word_context": 202.2382, "num_word_doc": 49.8483, "num_word_query": 23.3203, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21958.0441, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1328, "query_norm": 2.5684, "queue_k_norm": 2.6452, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3984, "sent_len_1": 66.8054, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.5125, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170700 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.3444, "doc_norm": 2.6464, "encoder_q-embeddings": 18496.4316, "encoder_q-layer.0": 13270.707, "encoder_q-layer.1": 14121.7461, "encoder_q-layer.10": 30930.5879, "encoder_q-layer.11": 72244.2812, "encoder_q-layer.2": 15791.6611, "encoder_q-layer.3": 16046.2188, "encoder_q-layer.4": 17284.0078, "encoder_q-layer.5": 17984.8691, "encoder_q-layer.6": 20314.7031, "encoder_q-layer.7": 23371.6309, "encoder_q-layer.8": 27753.623, "encoder_q-layer.9": 26084.1855, "epoch": 0.74, "inbatch_neg_score": 5.1337, "inbatch_pos_score": 5.7422, "learning_rate": 7.68421052631579e-06, "loss": 3.3444, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.9447, "num_token_overlap": 11.7108, "num_token_query": 31.4626, "num_token_union": 65.281, "num_word_context": 202.4162, "num_word_doc": 49.972, "num_word_query": 23.377, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43902.789, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1328, "query_norm": 2.5736, "queue_k_norm": 2.6457, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4626, "sent_len_1": 66.9447, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.1587, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170800 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3481, "doc_norm": 2.6472, "encoder_q-embeddings": 19834.7129, "encoder_q-layer.0": 14142.2666, "encoder_q-layer.1": 15143.8955, "encoder_q-layer.10": 25792.1777, "encoder_q-layer.11": 66452.75, "encoder_q-layer.2": 16243.4316, "encoder_q-layer.3": 16299.3467, "encoder_q-layer.4": 17214.0957, "encoder_q-layer.5": 17997.75, "encoder_q-layer.6": 18374.8145, "encoder_q-layer.7": 21070.9492, "encoder_q-layer.8": 25277.0801, "encoder_q-layer.9": 23628.8574, "epoch": 0.74, "inbatch_neg_score": 5.1398, "inbatch_pos_score": 5.75, "learning_rate": 7.657894736842106e-06, "loss": 3.3481, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.8638, "num_token_overlap": 11.6583, "num_token_query": 31.3456, "num_token_union": 65.1965, "num_word_context": 202.5952, "num_word_doc": 49.8921, "num_word_query": 23.2695, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41311.1083, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1406, "query_norm": 2.5722, "queue_k_norm": 2.647, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3456, "sent_len_1": 66.8638, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.46, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 170900 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3426, "doc_norm": 2.6488, "encoder_q-embeddings": 18686.9258, "encoder_q-layer.0": 13269.8125, "encoder_q-layer.1": 14010.8877, "encoder_q-layer.10": 28969.5293, "encoder_q-layer.11": 70001.4141, "encoder_q-layer.2": 15817.6377, "encoder_q-layer.3": 16439.1836, "encoder_q-layer.4": 17408.6836, "encoder_q-layer.5": 18218.6074, "encoder_q-layer.6": 20002.9473, "encoder_q-layer.7": 23674.4531, "encoder_q-layer.8": 29442.2773, "encoder_q-layer.9": 28622.0293, "epoch": 0.74, "inbatch_neg_score": 5.1413, "inbatch_pos_score": 5.7578, "learning_rate": 7.631578947368421e-06, "loss": 3.3426, "norm_diff": 0.0747, "norm_loss": 0.0, "num_token_doc": 66.7199, "num_token_overlap": 11.6799, "num_token_query": 31.3368, "num_token_union": 65.0636, "num_word_context": 202.4885, "num_word_doc": 49.7735, "num_word_query": 23.2765, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42841.5707, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1406, "query_norm": 2.5741, "queue_k_norm": 2.6471, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3368, "sent_len_1": 66.7199, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.3975, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 171000 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3368, "doc_norm": 2.6471, "encoder_q-embeddings": 21632.5527, "encoder_q-layer.0": 16253.5859, "encoder_q-layer.1": 17928.457, "encoder_q-layer.10": 27608.8262, "encoder_q-layer.11": 69383.3906, "encoder_q-layer.2": 21149.248, "encoder_q-layer.3": 22399.5039, "encoder_q-layer.4": 24480.416, "encoder_q-layer.5": 23498.3457, "encoder_q-layer.6": 23915.5195, "encoder_q-layer.7": 25743.3789, "encoder_q-layer.8": 28279.5254, "encoder_q-layer.9": 25519.3594, "epoch": 0.74, "inbatch_neg_score": 5.1465, "inbatch_pos_score": 5.7578, "learning_rate": 7.605263157894738e-06, "loss": 3.3368, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.7569, "num_token_overlap": 11.7126, "num_token_query": 31.4129, "num_token_union": 65.0811, "num_word_context": 202.194, "num_word_doc": 49.8147, "num_word_query": 23.3292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 45801.0184, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1484, "query_norm": 2.5727, "queue_k_norm": 2.6468, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4129, "sent_len_1": 66.7569, "sent_len_max_0": 127.9537, "sent_len_max_1": 209.2125, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 171100 }, { "accuracy": 53.1738, "active_queue_size": 16384.0, "cl_loss": 3.3411, "doc_norm": 2.6467, "encoder_q-embeddings": 18758.498, "encoder_q-layer.0": 13439.7412, "encoder_q-layer.1": 14268.9053, "encoder_q-layer.10": 29978.7148, "encoder_q-layer.11": 69012.8281, "encoder_q-layer.2": 15903.5371, "encoder_q-layer.3": 16064.2363, "encoder_q-layer.4": 17356.002, "encoder_q-layer.5": 18137.7578, "encoder_q-layer.6": 20412.4551, "encoder_q-layer.7": 23215.8691, "encoder_q-layer.8": 27662.5078, "encoder_q-layer.9": 27417.9062, "epoch": 0.74, "inbatch_neg_score": 5.1479, "inbatch_pos_score": 5.7812, "learning_rate": 7.578947368421054e-06, "loss": 3.3411, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.6781, "num_token_overlap": 11.6933, "num_token_query": 31.3601, "num_token_union": 65.0119, "num_word_context": 202.2635, "num_word_doc": 49.7398, "num_word_query": 23.2912, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42462.0451, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1484, "query_norm": 2.5761, "queue_k_norm": 2.6475, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3601, "sent_len_1": 66.6781, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.5775, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 171200 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.3358, "doc_norm": 2.6495, "encoder_q-embeddings": 18927.2051, "encoder_q-layer.0": 13623.1201, "encoder_q-layer.1": 14500.123, "encoder_q-layer.10": 27239.25, "encoder_q-layer.11": 68439.9297, "encoder_q-layer.2": 15842.7471, "encoder_q-layer.3": 15975.9902, "encoder_q-layer.4": 17003.4258, "encoder_q-layer.5": 16872.4531, "encoder_q-layer.6": 19826.3438, "encoder_q-layer.7": 22292.9121, "encoder_q-layer.8": 26732.2402, "encoder_q-layer.9": 24653.6465, "epoch": 0.74, "inbatch_neg_score": 5.1521, "inbatch_pos_score": 5.7578, "learning_rate": 7.552631578947368e-06, "loss": 3.3358, "norm_diff": 0.0729, "norm_loss": 0.0, "num_token_doc": 66.743, "num_token_overlap": 11.6942, "num_token_query": 31.4826, "num_token_union": 65.1371, "num_word_context": 202.3367, "num_word_doc": 49.815, "num_word_query": 23.3859, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41918.1721, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1523, "query_norm": 2.5766, "queue_k_norm": 2.6491, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4826, "sent_len_1": 66.743, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.7025, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 171300 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3283, "doc_norm": 2.6484, "encoder_q-embeddings": 19173.7559, "encoder_q-layer.0": 14008.7939, "encoder_q-layer.1": 15056.499, "encoder_q-layer.10": 27170.457, "encoder_q-layer.11": 67205.7266, "encoder_q-layer.2": 17021.7598, "encoder_q-layer.3": 17755.3203, "encoder_q-layer.4": 18966.3086, "encoder_q-layer.5": 19142.6016, "encoder_q-layer.6": 21142.2441, "encoder_q-layer.7": 22660.5645, "encoder_q-layer.8": 27581.6328, "encoder_q-layer.9": 25161.8574, "epoch": 0.74, "inbatch_neg_score": 5.1528, "inbatch_pos_score": 5.7578, "learning_rate": 7.526315789473684e-06, "loss": 3.3283, "norm_diff": 0.0734, "norm_loss": 0.0, "num_token_doc": 66.7731, "num_token_overlap": 11.7021, "num_token_query": 31.509, "num_token_union": 65.1946, "num_word_context": 202.4298, "num_word_doc": 49.8453, "num_word_query": 23.4152, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42231.2571, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1523, "query_norm": 2.575, "queue_k_norm": 2.6489, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.509, "sent_len_1": 66.7731, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.8212, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 171400 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3474, "doc_norm": 2.6488, "encoder_q-embeddings": 18732.7148, "encoder_q-layer.0": 13154.8242, "encoder_q-layer.1": 14342.5381, "encoder_q-layer.10": 26104.3047, "encoder_q-layer.11": 67595.1953, "encoder_q-layer.2": 15830.9736, "encoder_q-layer.3": 16595.457, "encoder_q-layer.4": 17628.0527, "encoder_q-layer.5": 18152.5781, "encoder_q-layer.6": 20914.1738, "encoder_q-layer.7": 22909.084, "encoder_q-layer.8": 27420.5371, "encoder_q-layer.9": 24839.6543, "epoch": 0.74, "inbatch_neg_score": 5.1596, "inbatch_pos_score": 5.7773, "learning_rate": 7.5e-06, "loss": 3.3474, "norm_diff": 0.0711, "norm_loss": 0.0, "num_token_doc": 66.9138, "num_token_overlap": 11.6981, "num_token_query": 31.4383, "num_token_union": 65.2075, "num_word_context": 202.7046, "num_word_doc": 49.9169, "num_word_query": 23.3599, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41944.4741, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1602, "query_norm": 2.5778, "queue_k_norm": 2.6487, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4383, "sent_len_1": 66.9138, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.4212, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 171500 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.345, "doc_norm": 2.6512, "encoder_q-embeddings": 22511.4102, "encoder_q-layer.0": 16170.2363, "encoder_q-layer.1": 17891.7129, "encoder_q-layer.10": 31772.2676, "encoder_q-layer.11": 71489.8438, "encoder_q-layer.2": 19896.6797, "encoder_q-layer.3": 21754.4863, "encoder_q-layer.4": 22801.4102, "encoder_q-layer.5": 25597.2285, "encoder_q-layer.6": 26492.002, "encoder_q-layer.7": 27166.5, "encoder_q-layer.8": 32062.748, "encoder_q-layer.9": 28461.5703, "epoch": 0.74, "inbatch_neg_score": 5.163, "inbatch_pos_score": 5.7812, "learning_rate": 7.473684210526316e-06, "loss": 3.345, "norm_diff": 0.066, "norm_loss": 0.0, "num_token_doc": 66.5773, "num_token_overlap": 11.6382, "num_token_query": 31.4015, "num_token_union": 65.0679, "num_word_context": 202.2902, "num_word_doc": 49.6937, "num_word_query": 23.323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 46934.0519, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1641, "query_norm": 2.5852, "queue_k_norm": 2.6494, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4015, "sent_len_1": 66.5773, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.5037, "stdk": 0.0496, "stdq": 0.0436, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 171600 }, { "accuracy": 53.6133, "active_queue_size": 16384.0, "cl_loss": 3.3489, "doc_norm": 2.6513, "encoder_q-embeddings": 19500.0469, "encoder_q-layer.0": 13649.1396, "encoder_q-layer.1": 14654.6523, "encoder_q-layer.10": 29406.7266, "encoder_q-layer.11": 68123.7188, "encoder_q-layer.2": 16596.0586, "encoder_q-layer.3": 16814.7656, "encoder_q-layer.4": 18601.6055, "encoder_q-layer.5": 18952.4746, "encoder_q-layer.6": 22144.7539, "encoder_q-layer.7": 23429.3281, "encoder_q-layer.8": 27807.0996, "encoder_q-layer.9": 25361.8242, "epoch": 0.75, "inbatch_neg_score": 5.1661, "inbatch_pos_score": 5.7891, "learning_rate": 7.447368421052632e-06, "loss": 3.3489, "norm_diff": 0.0714, "norm_loss": 0.0, "num_token_doc": 66.7283, "num_token_overlap": 11.678, "num_token_query": 31.371, "num_token_union": 65.0533, "num_word_context": 202.31, "num_word_doc": 49.769, "num_word_query": 23.2928, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42663.7609, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1641, "query_norm": 2.5798, "queue_k_norm": 2.6499, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.371, "sent_len_1": 66.7283, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.6113, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 171700 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3606, "doc_norm": 2.6503, "encoder_q-embeddings": 19060.1855, "encoder_q-layer.0": 13479.5273, "encoder_q-layer.1": 14053.7041, "encoder_q-layer.10": 27745.9258, "encoder_q-layer.11": 68064.75, "encoder_q-layer.2": 15989.0254, "encoder_q-layer.3": 16777.668, "encoder_q-layer.4": 17609.5762, "encoder_q-layer.5": 18742.0527, "encoder_q-layer.6": 21127.5977, "encoder_q-layer.7": 23896.2266, "encoder_q-layer.8": 26806.4473, "encoder_q-layer.9": 26266.7109, "epoch": 0.75, "inbatch_neg_score": 5.1665, "inbatch_pos_score": 5.7852, "learning_rate": 7.421052631578948e-06, "loss": 3.3606, "norm_diff": 0.0701, "norm_loss": 0.0, "num_token_doc": 66.7009, "num_token_overlap": 11.6337, "num_token_query": 31.3424, "num_token_union": 65.0784, "num_word_context": 202.3464, "num_word_doc": 49.7676, "num_word_query": 23.2908, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42443.2732, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.168, "query_norm": 2.5802, "queue_k_norm": 2.6499, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3424, "sent_len_1": 66.7009, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.215, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 171800 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3379, "doc_norm": 2.6544, "encoder_q-embeddings": 19204.1895, "encoder_q-layer.0": 12964.3926, "encoder_q-layer.1": 13878.002, "encoder_q-layer.10": 26249.9277, "encoder_q-layer.11": 66337.2891, "encoder_q-layer.2": 15390.5635, "encoder_q-layer.3": 15812.5674, "encoder_q-layer.4": 17412.2988, "encoder_q-layer.5": 18446.873, "encoder_q-layer.6": 20890.2695, "encoder_q-layer.7": 22678.9141, "encoder_q-layer.8": 27258.4648, "encoder_q-layer.9": 25969.2637, "epoch": 0.75, "inbatch_neg_score": 5.1707, "inbatch_pos_score": 5.7852, "learning_rate": 7.394736842105264e-06, "loss": 3.3379, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.7119, "num_token_overlap": 11.6746, "num_token_query": 31.3546, "num_token_union": 65.0957, "num_word_context": 202.1441, "num_word_doc": 49.7586, "num_word_query": 23.2677, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41341.2546, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1719, "query_norm": 2.581, "queue_k_norm": 2.652, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3546, "sent_len_1": 66.7119, "sent_len_max_0": 127.9663, "sent_len_max_1": 208.4162, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 171900 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3458, "doc_norm": 2.6518, "encoder_q-embeddings": 19703.7812, "encoder_q-layer.0": 14236.8037, "encoder_q-layer.1": 15001.2246, "encoder_q-layer.10": 26835.6875, "encoder_q-layer.11": 65895.1016, "encoder_q-layer.2": 16897.8398, "encoder_q-layer.3": 16910.0352, "encoder_q-layer.4": 18014.0762, "encoder_q-layer.5": 17832.5, "encoder_q-layer.6": 20222.2383, "encoder_q-layer.7": 23167.7207, "encoder_q-layer.8": 26749.5078, "encoder_q-layer.9": 25035.8789, "epoch": 0.75, "inbatch_neg_score": 5.1712, "inbatch_pos_score": 5.7812, "learning_rate": 7.3684210526315784e-06, "loss": 3.3458, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.722, "num_token_overlap": 11.6661, "num_token_query": 31.3672, "num_token_union": 65.1014, "num_word_context": 202.4234, "num_word_doc": 49.8132, "num_word_query": 23.3035, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41939.043, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1719, "query_norm": 2.579, "queue_k_norm": 2.6515, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3672, "sent_len_1": 66.722, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.6975, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 172000 }, { "accuracy": 52.7344, "active_queue_size": 16384.0, "cl_loss": 3.3542, "doc_norm": 2.6548, "encoder_q-embeddings": 22143.5312, "encoder_q-layer.0": 15815.4121, "encoder_q-layer.1": 16899.1348, "encoder_q-layer.10": 28022.6367, "encoder_q-layer.11": 68062.5781, "encoder_q-layer.2": 18988.7871, "encoder_q-layer.3": 19890.084, "encoder_q-layer.4": 22244.7441, "encoder_q-layer.5": 22577.4492, "encoder_q-layer.6": 24174.6172, "encoder_q-layer.7": 26260.6016, "encoder_q-layer.8": 29277.0938, "encoder_q-layer.9": 25473.7871, "epoch": 0.75, "inbatch_neg_score": 5.1719, "inbatch_pos_score": 5.7969, "learning_rate": 7.342105263157894e-06, "loss": 3.3542, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.8392, "num_token_overlap": 11.6641, "num_token_query": 31.3329, "num_token_union": 65.1125, "num_word_context": 202.3162, "num_word_doc": 49.8708, "num_word_query": 23.2819, "postclip_grad_norm": 1.0, "preclip_grad_norm": 45013.3278, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1719, "query_norm": 2.5791, "queue_k_norm": 2.6524, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3329, "sent_len_1": 66.8392, "sent_len_max_0": 127.985, "sent_len_max_1": 208.215, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172100 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3442, "doc_norm": 2.6544, "encoder_q-embeddings": 20058.7812, "encoder_q-layer.0": 14111.5352, "encoder_q-layer.1": 15261.5615, "encoder_q-layer.10": 29682.4707, "encoder_q-layer.11": 69002.3906, "encoder_q-layer.2": 16901.5605, "encoder_q-layer.3": 17551.9336, "encoder_q-layer.4": 19738.6484, "encoder_q-layer.5": 20673.1816, "encoder_q-layer.6": 23052.5781, "encoder_q-layer.7": 25337.7969, "encoder_q-layer.8": 30007.1621, "encoder_q-layer.9": 26018.0156, "epoch": 0.75, "inbatch_neg_score": 5.1743, "inbatch_pos_score": 5.7891, "learning_rate": 7.315789473684211e-06, "loss": 3.3442, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.7042, "num_token_overlap": 11.6862, "num_token_query": 31.4642, "num_token_union": 65.1178, "num_word_context": 202.3397, "num_word_doc": 49.7588, "num_word_query": 23.3711, "postclip_grad_norm": 1.0, "preclip_grad_norm": 44693.6896, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1758, "query_norm": 2.5769, "queue_k_norm": 2.6534, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4642, "sent_len_1": 66.7042, "sent_len_max_0": 127.9213, "sent_len_max_1": 210.6387, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172200 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.3423, "doc_norm": 2.6536, "encoder_q-embeddings": 19412.2539, "encoder_q-layer.0": 13133.8867, "encoder_q-layer.1": 14291.4561, "encoder_q-layer.10": 27029.6094, "encoder_q-layer.11": 68295.1562, "encoder_q-layer.2": 16380.5508, "encoder_q-layer.3": 17146.0586, "encoder_q-layer.4": 18429.832, "encoder_q-layer.5": 18428.5801, "encoder_q-layer.6": 21405.5938, "encoder_q-layer.7": 23367.8887, "encoder_q-layer.8": 27495.9727, "encoder_q-layer.9": 24963.1934, "epoch": 0.75, "inbatch_neg_score": 5.1788, "inbatch_pos_score": 5.7852, "learning_rate": 7.289473684210527e-06, "loss": 3.3423, "norm_diff": 0.0738, "norm_loss": 0.0, "num_token_doc": 66.7303, "num_token_overlap": 11.6439, "num_token_query": 31.3271, "num_token_union": 65.1018, "num_word_context": 202.2907, "num_word_doc": 49.786, "num_word_query": 23.2525, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42631.1569, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1797, "query_norm": 2.5797, "queue_k_norm": 2.6534, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3271, "sent_len_1": 66.7303, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.1637, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 172300 }, { "accuracy": 52.7832, "active_queue_size": 16384.0, "cl_loss": 3.3328, "doc_norm": 2.6553, "encoder_q-embeddings": 18601.2832, "encoder_q-layer.0": 12926.3496, "encoder_q-layer.1": 13647.8301, "encoder_q-layer.10": 26813.4941, "encoder_q-layer.11": 66477.6406, "encoder_q-layer.2": 15383.1504, "encoder_q-layer.3": 15709.1406, "encoder_q-layer.4": 16868.7793, "encoder_q-layer.5": 17283.4883, "encoder_q-layer.6": 19138.3164, "encoder_q-layer.7": 22401.5156, "encoder_q-layer.8": 26775.0137, "encoder_q-layer.9": 24760.7324, "epoch": 0.75, "inbatch_neg_score": 5.182, "inbatch_pos_score": 5.8008, "learning_rate": 7.2631578947368426e-06, "loss": 3.3328, "norm_diff": 0.0703, "norm_loss": 0.0, "num_token_doc": 66.8339, "num_token_overlap": 11.7026, "num_token_query": 31.4913, "num_token_union": 65.2225, "num_word_context": 202.4079, "num_word_doc": 49.8837, "num_word_query": 23.4022, "postclip_grad_norm": 1.0, "preclip_grad_norm": 41713.8201, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1797, "query_norm": 2.585, "queue_k_norm": 2.6544, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4913, "sent_len_1": 66.8339, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.49, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172400 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3599, "doc_norm": 2.6545, "encoder_q-embeddings": 19151.4922, "encoder_q-layer.0": 13012.0273, "encoder_q-layer.1": 14237.4785, "encoder_q-layer.10": 27138.3652, "encoder_q-layer.11": 68644.7188, "encoder_q-layer.2": 15986.8223, "encoder_q-layer.3": 16541.5742, "encoder_q-layer.4": 17709.7188, "encoder_q-layer.5": 18549.373, "encoder_q-layer.6": 20155.873, "encoder_q-layer.7": 22514.7695, "encoder_q-layer.8": 27558.0527, "encoder_q-layer.9": 25155.7305, "epoch": 0.75, "inbatch_neg_score": 5.185, "inbatch_pos_score": 5.8008, "learning_rate": 7.236842105263158e-06, "loss": 3.3599, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.7137, "num_token_overlap": 11.6888, "num_token_query": 31.4065, "num_token_union": 65.1077, "num_word_context": 202.173, "num_word_doc": 49.7822, "num_word_query": 23.3245, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42852.7837, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.1875, "query_norm": 2.5817, "queue_k_norm": 2.6544, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4065, "sent_len_1": 66.7137, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.0025, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 172500 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.344, "doc_norm": 2.6561, "encoder_q-embeddings": 10063.6621, "encoder_q-layer.0": 7015.6479, "encoder_q-layer.1": 7517.085, "encoder_q-layer.10": 12791.4443, "encoder_q-layer.11": 33024.5, "encoder_q-layer.2": 8251.4092, "encoder_q-layer.3": 8550.6357, "encoder_q-layer.4": 9585.0557, "encoder_q-layer.5": 10061.9385, "encoder_q-layer.6": 11078.9561, "encoder_q-layer.7": 12255.5771, "encoder_q-layer.8": 13638.5176, "encoder_q-layer.9": 12193.2686, "epoch": 0.75, "inbatch_neg_score": 5.1878, "inbatch_pos_score": 5.8203, "learning_rate": 7.210526315789474e-06, "loss": 3.344, "norm_diff": 0.0712, "norm_loss": 0.0, "num_token_doc": 66.9424, "num_token_overlap": 11.6808, "num_token_query": 31.4467, "num_token_union": 65.2258, "num_word_context": 202.2393, "num_word_doc": 49.911, "num_word_query": 23.3764, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21137.4585, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1875, "query_norm": 2.5849, "queue_k_norm": 2.655, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4467, "sent_len_1": 66.9424, "sent_len_max_0": 127.9925, "sent_len_max_1": 213.4137, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172600 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3442, "doc_norm": 2.6564, "encoder_q-embeddings": 9024.4766, "encoder_q-layer.0": 6572.2104, "encoder_q-layer.1": 7155.6616, "encoder_q-layer.10": 13494.6426, "encoder_q-layer.11": 34099.7969, "encoder_q-layer.2": 7890.1558, "encoder_q-layer.3": 7887.2368, "encoder_q-layer.4": 8294.5479, "encoder_q-layer.5": 8812.2539, "encoder_q-layer.6": 9817.4824, "encoder_q-layer.7": 11448.7559, "encoder_q-layer.8": 13026.7012, "encoder_q-layer.9": 12554.6816, "epoch": 0.75, "inbatch_neg_score": 5.1909, "inbatch_pos_score": 5.8125, "learning_rate": 7.184210526315791e-06, "loss": 3.3442, "norm_diff": 0.0713, "norm_loss": 0.0, "num_token_doc": 66.8548, "num_token_overlap": 11.6592, "num_token_query": 31.3059, "num_token_union": 65.1333, "num_word_context": 202.2874, "num_word_doc": 49.8643, "num_word_query": 23.242, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20632.609, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1914, "query_norm": 2.5851, "queue_k_norm": 2.6558, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3059, "sent_len_1": 66.8548, "sent_len_max_0": 127.9938, "sent_len_max_1": 210.82, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172700 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3335, "doc_norm": 2.6599, "encoder_q-embeddings": 9095.8311, "encoder_q-layer.0": 6387.6753, "encoder_q-layer.1": 6629.1196, "encoder_q-layer.10": 13150.4121, "encoder_q-layer.11": 33809.3828, "encoder_q-layer.2": 7567.5049, "encoder_q-layer.3": 7880.7842, "encoder_q-layer.4": 8673.9121, "encoder_q-layer.5": 8665.874, "encoder_q-layer.6": 9736.665, "encoder_q-layer.7": 10924.3516, "encoder_q-layer.8": 13394.9385, "encoder_q-layer.9": 12562.4141, "epoch": 0.75, "inbatch_neg_score": 5.1919, "inbatch_pos_score": 5.8242, "learning_rate": 7.157894736842105e-06, "loss": 3.3335, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.8847, "num_token_overlap": 11.7129, "num_token_query": 31.4813, "num_token_union": 65.2083, "num_word_context": 202.3628, "num_word_doc": 49.8742, "num_word_query": 23.372, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20892.7819, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1914, "query_norm": 2.5823, "queue_k_norm": 2.6568, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4813, "sent_len_1": 66.8847, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.93, "stdk": 0.0498, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 172800 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3387, "doc_norm": 2.6589, "encoder_q-embeddings": 10149.4648, "encoder_q-layer.0": 7016.8086, "encoder_q-layer.1": 7733.5186, "encoder_q-layer.10": 14681.5098, "encoder_q-layer.11": 35633.2227, "encoder_q-layer.2": 8840.0176, "encoder_q-layer.3": 8974.5293, "encoder_q-layer.4": 9769.5518, "encoder_q-layer.5": 10206.0166, "encoder_q-layer.6": 11340.1299, "encoder_q-layer.7": 12417.1484, "encoder_q-layer.8": 14533.584, "encoder_q-layer.9": 13223.8164, "epoch": 0.75, "inbatch_neg_score": 5.1957, "inbatch_pos_score": 5.8125, "learning_rate": 7.131578947368421e-06, "loss": 3.3387, "norm_diff": 0.0704, "norm_loss": 0.0, "num_token_doc": 66.6687, "num_token_overlap": 11.6943, "num_token_query": 31.3931, "num_token_union": 65.0376, "num_word_context": 202.1241, "num_word_doc": 49.7608, "num_word_query": 23.3342, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21813.4425, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1953, "query_norm": 2.5885, "queue_k_norm": 2.6559, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3931, "sent_len_1": 66.6687, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.5062, "stdk": 0.0497, "stdq": 0.0437, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 172900 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3383, "doc_norm": 2.6577, "encoder_q-embeddings": 9315.0322, "encoder_q-layer.0": 6422.8857, "encoder_q-layer.1": 6912.0117, "encoder_q-layer.10": 14421.8672, "encoder_q-layer.11": 34349.9609, "encoder_q-layer.2": 7647.9844, "encoder_q-layer.3": 7951.7114, "encoder_q-layer.4": 8406.874, "encoder_q-layer.5": 8857.1406, "encoder_q-layer.6": 9615.1602, "encoder_q-layer.7": 10851.8408, "encoder_q-layer.8": 13545.1758, "encoder_q-layer.9": 12913.5527, "epoch": 0.75, "inbatch_neg_score": 5.2012, "inbatch_pos_score": 5.8203, "learning_rate": 7.1052631578947375e-06, "loss": 3.3383, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.8794, "num_token_overlap": 11.724, "num_token_query": 31.5087, "num_token_union": 65.2305, "num_word_context": 202.6231, "num_word_doc": 49.8731, "num_word_query": 23.4068, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21007.1912, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.1992, "query_norm": 2.5855, "queue_k_norm": 2.658, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5087, "sent_len_1": 66.8794, "sent_len_max_0": 127.9412, "sent_len_max_1": 210.6962, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 173000 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3446, "doc_norm": 2.6567, "encoder_q-embeddings": 9158.0312, "encoder_q-layer.0": 6613.3877, "encoder_q-layer.1": 6816.1318, "encoder_q-layer.10": 12950.8877, "encoder_q-layer.11": 32844.7461, "encoder_q-layer.2": 7600.9062, "encoder_q-layer.3": 7594.1943, "encoder_q-layer.4": 8184.248, "encoder_q-layer.5": 8699.6514, "encoder_q-layer.6": 9548.9209, "encoder_q-layer.7": 10794.5703, "encoder_q-layer.8": 13165.4473, "encoder_q-layer.9": 12535.4316, "epoch": 0.75, "inbatch_neg_score": 5.2017, "inbatch_pos_score": 5.8125, "learning_rate": 7.078947368421053e-06, "loss": 3.3446, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7871, "num_token_overlap": 11.6249, "num_token_query": 31.3417, "num_token_union": 65.1684, "num_word_context": 202.3788, "num_word_doc": 49.8606, "num_word_query": 23.271, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20415.8881, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2031, "query_norm": 2.5806, "queue_k_norm": 2.6582, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3417, "sent_len_1": 66.7871, "sent_len_max_0": 127.975, "sent_len_max_1": 210.685, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 173100 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3437, "doc_norm": 2.6572, "encoder_q-embeddings": 9072.6992, "encoder_q-layer.0": 6405.6646, "encoder_q-layer.1": 6729.123, "encoder_q-layer.10": 15362.833, "encoder_q-layer.11": 33546.4844, "encoder_q-layer.2": 7427.3457, "encoder_q-layer.3": 7767.3433, "encoder_q-layer.4": 8361.8184, "encoder_q-layer.5": 8646.2285, "encoder_q-layer.6": 9720.502, "encoder_q-layer.7": 11969.9561, "encoder_q-layer.8": 13585.0693, "encoder_q-layer.9": 12863.8799, "epoch": 0.75, "inbatch_neg_score": 5.2048, "inbatch_pos_score": 5.8125, "learning_rate": 7.052631578947369e-06, "loss": 3.3437, "norm_diff": 0.0772, "norm_loss": 0.0, "num_token_doc": 66.6066, "num_token_overlap": 11.633, "num_token_query": 31.3281, "num_token_union": 65.0192, "num_word_context": 202.237, "num_word_doc": 49.7046, "num_word_query": 23.2643, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20890.4593, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2031, "query_norm": 2.5801, "queue_k_norm": 2.658, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3281, "sent_len_1": 66.6066, "sent_len_max_0": 127.97, "sent_len_max_1": 208.6763, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 173200 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3422, "doc_norm": 2.6591, "encoder_q-embeddings": 9591.0615, "encoder_q-layer.0": 6580.6909, "encoder_q-layer.1": 7176.5938, "encoder_q-layer.10": 14449.3633, "encoder_q-layer.11": 34669.7422, "encoder_q-layer.2": 7588.2119, "encoder_q-layer.3": 7826.8198, "encoder_q-layer.4": 8237.0508, "encoder_q-layer.5": 8610.0166, "encoder_q-layer.6": 9663.1924, "encoder_q-layer.7": 11344.0771, "encoder_q-layer.8": 13827.7529, "encoder_q-layer.9": 12563.4648, "epoch": 0.75, "inbatch_neg_score": 5.2079, "inbatch_pos_score": 5.8242, "learning_rate": 7.026315789473685e-06, "loss": 3.3422, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.6336, "num_token_overlap": 11.6793, "num_token_query": 31.3708, "num_token_union": 65.0478, "num_word_context": 202.0081, "num_word_doc": 49.7259, "num_word_query": 23.2919, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21167.5969, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.207, "query_norm": 2.5846, "queue_k_norm": 2.6595, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3708, "sent_len_1": 66.6336, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.0037, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 173300 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3447, "doc_norm": 2.6597, "encoder_q-embeddings": 9538.7412, "encoder_q-layer.0": 6706.0093, "encoder_q-layer.1": 7101.4819, "encoder_q-layer.10": 13380.8652, "encoder_q-layer.11": 34367.3242, "encoder_q-layer.2": 8001.0044, "encoder_q-layer.3": 8608.9434, "encoder_q-layer.4": 9128.0859, "encoder_q-layer.5": 9697.0918, "encoder_q-layer.6": 10443.9883, "encoder_q-layer.7": 11546.5107, "encoder_q-layer.8": 14157.8896, "encoder_q-layer.9": 12910.9482, "epoch": 0.75, "inbatch_neg_score": 5.2104, "inbatch_pos_score": 5.8242, "learning_rate": 7.000000000000001e-06, "loss": 3.3447, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.8956, "num_token_overlap": 11.6738, "num_token_query": 31.3496, "num_token_union": 65.1747, "num_word_context": 202.4828, "num_word_doc": 49.901, "num_word_query": 23.2764, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21242.7512, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2109, "query_norm": 2.5872, "queue_k_norm": 2.6581, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3496, "sent_len_1": 66.8956, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.5588, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 173400 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3409, "doc_norm": 2.6614, "encoder_q-embeddings": 11444.4844, "encoder_q-layer.0": 8258.0645, "encoder_q-layer.1": 9784.9551, "encoder_q-layer.10": 13580.9043, "encoder_q-layer.11": 36036.2656, "encoder_q-layer.2": 10745.7324, "encoder_q-layer.3": 11009.0078, "encoder_q-layer.4": 11416.7383, "encoder_q-layer.5": 11721.6006, "encoder_q-layer.6": 11961.6104, "encoder_q-layer.7": 12464.1406, "encoder_q-layer.8": 13989.9824, "encoder_q-layer.9": 12881.1162, "epoch": 0.75, "inbatch_neg_score": 5.2139, "inbatch_pos_score": 5.8203, "learning_rate": 6.973684210526316e-06, "loss": 3.3409, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.6612, "num_token_overlap": 11.6819, "num_token_query": 31.3872, "num_token_union": 65.0552, "num_word_context": 202.2448, "num_word_doc": 49.7442, "num_word_query": 23.3083, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23466.3443, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2148, "query_norm": 2.5841, "queue_k_norm": 2.6599, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3872, "sent_len_1": 66.6612, "sent_len_max_0": 127.97, "sent_len_max_1": 208.6125, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 173500 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3365, "doc_norm": 2.6626, "encoder_q-embeddings": 9495.3301, "encoder_q-layer.0": 6974.8882, "encoder_q-layer.1": 7253.293, "encoder_q-layer.10": 13627.1221, "encoder_q-layer.11": 34821.9375, "encoder_q-layer.2": 7986.2598, "encoder_q-layer.3": 8119.5674, "encoder_q-layer.4": 8737.2061, "encoder_q-layer.5": 8876.123, "encoder_q-layer.6": 10043.5088, "encoder_q-layer.7": 11834.9365, "encoder_q-layer.8": 13289.9404, "encoder_q-layer.9": 12947.0723, "epoch": 0.75, "inbatch_neg_score": 5.2148, "inbatch_pos_score": 5.832, "learning_rate": 6.9473684210526315e-06, "loss": 3.3365, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.9238, "num_token_overlap": 11.7089, "num_token_query": 31.4032, "num_token_union": 65.2319, "num_word_context": 202.5867, "num_word_doc": 49.9676, "num_word_query": 23.313, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21849.9931, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2148, "query_norm": 2.5857, "queue_k_norm": 2.6594, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4032, "sent_len_1": 66.9238, "sent_len_max_0": 127.9925, "sent_len_max_1": 206.4563, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 173600 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3294, "doc_norm": 2.6621, "encoder_q-embeddings": 9355.8242, "encoder_q-layer.0": 6580.731, "encoder_q-layer.1": 6974.688, "encoder_q-layer.10": 13038.2881, "encoder_q-layer.11": 32659.3242, "encoder_q-layer.2": 7826.9448, "encoder_q-layer.3": 7890.0654, "encoder_q-layer.4": 8297.9941, "encoder_q-layer.5": 8781.9355, "encoder_q-layer.6": 9700.8574, "encoder_q-layer.7": 11240.1875, "encoder_q-layer.8": 13353.8359, "encoder_q-layer.9": 12391.498, "epoch": 0.75, "inbatch_neg_score": 5.2208, "inbatch_pos_score": 5.8438, "learning_rate": 6.921052631578947e-06, "loss": 3.3294, "norm_diff": 0.073, "norm_loss": 0.0, "num_token_doc": 66.8094, "num_token_overlap": 11.7011, "num_token_query": 31.3633, "num_token_union": 65.1368, "num_word_context": 202.3908, "num_word_doc": 49.8242, "num_word_query": 23.2858, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20582.0296, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2188, "query_norm": 2.5891, "queue_k_norm": 2.6613, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3633, "sent_len_1": 66.8094, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.9175, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 173700 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.361, "doc_norm": 2.6616, "encoder_q-embeddings": 9663.8945, "encoder_q-layer.0": 6967.4434, "encoder_q-layer.1": 7747.771, "encoder_q-layer.10": 13778.1787, "encoder_q-layer.11": 33989.7812, "encoder_q-layer.2": 8774.9492, "encoder_q-layer.3": 8898.0146, "encoder_q-layer.4": 9639.2754, "encoder_q-layer.5": 10311.0645, "encoder_q-layer.6": 10833.9121, "encoder_q-layer.7": 12483.9727, "encoder_q-layer.8": 14140.8516, "encoder_q-layer.9": 12806.1924, "epoch": 0.75, "inbatch_neg_score": 5.2216, "inbatch_pos_score": 5.8398, "learning_rate": 6.894736842105263e-06, "loss": 3.361, "norm_diff": 0.0734, "norm_loss": 0.0, "num_token_doc": 66.6853, "num_token_overlap": 11.5935, "num_token_query": 31.2256, "num_token_union": 65.0554, "num_word_context": 202.1998, "num_word_doc": 49.7414, "num_word_query": 23.1739, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21819.5772, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2227, "query_norm": 2.5882, "queue_k_norm": 2.6609, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2256, "sent_len_1": 66.6853, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.94, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 173800 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3527, "doc_norm": 2.6627, "encoder_q-embeddings": 10163.8994, "encoder_q-layer.0": 7280.5342, "encoder_q-layer.1": 7980.2295, "encoder_q-layer.10": 13016.2324, "encoder_q-layer.11": 33790.2891, "encoder_q-layer.2": 9117.5635, "encoder_q-layer.3": 9414.3262, "encoder_q-layer.4": 10185.1152, "encoder_q-layer.5": 10672.8926, "encoder_q-layer.6": 11386.0889, "encoder_q-layer.7": 11791.1768, "encoder_q-layer.8": 13088.2314, "encoder_q-layer.9": 12438.3037, "epoch": 0.75, "inbatch_neg_score": 5.2231, "inbatch_pos_score": 5.832, "learning_rate": 6.86842105263158e-06, "loss": 3.3527, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.7884, "num_token_overlap": 11.6601, "num_token_query": 31.3566, "num_token_union": 65.1343, "num_word_context": 202.397, "num_word_doc": 49.8591, "num_word_query": 23.2927, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21739.1142, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2227, "query_norm": 2.5855, "queue_k_norm": 2.6613, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3566, "sent_len_1": 66.7884, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.9563, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 173900 }, { "accuracy": 50.3418, "active_queue_size": 16384.0, "cl_loss": 3.3378, "doc_norm": 2.6638, "encoder_q-embeddings": 15800.9219, "encoder_q-layer.0": 11348.499, "encoder_q-layer.1": 13117.7598, "encoder_q-layer.10": 13483.9424, "encoder_q-layer.11": 34611.3203, "encoder_q-layer.2": 16852.8008, "encoder_q-layer.3": 18168.0293, "encoder_q-layer.4": 19208.9375, "encoder_q-layer.5": 19028.2969, "encoder_q-layer.6": 20106.4824, "encoder_q-layer.7": 19171.0879, "encoder_q-layer.8": 17439.2305, "encoder_q-layer.9": 13605.2988, "epoch": 0.76, "inbatch_neg_score": 5.2227, "inbatch_pos_score": 5.8398, "learning_rate": 6.842105263157896e-06, "loss": 3.3378, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7533, "num_token_overlap": 11.7205, "num_token_query": 31.4538, "num_token_union": 65.1158, "num_word_context": 202.3803, "num_word_doc": 49.8345, "num_word_query": 23.3522, "postclip_grad_norm": 1.0, "preclip_grad_norm": 28772.194, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 5.2227, "query_norm": 2.5877, "queue_k_norm": 2.6629, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4538, "sent_len_1": 66.7533, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.5163, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174000 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.661, "encoder_q-embeddings": 9827.3125, "encoder_q-layer.0": 7380.0547, "encoder_q-layer.1": 7679.7451, "encoder_q-layer.10": 14129.7881, "encoder_q-layer.11": 35169.8672, "encoder_q-layer.2": 8672.2705, "encoder_q-layer.3": 8919.8086, "encoder_q-layer.4": 9477.7734, "encoder_q-layer.5": 9982.4873, "encoder_q-layer.6": 11177.1045, "encoder_q-layer.7": 12583.0605, "encoder_q-layer.8": 14380.2529, "encoder_q-layer.9": 12603.458, "epoch": 0.76, "inbatch_neg_score": 5.2251, "inbatch_pos_score": 5.8359, "learning_rate": 6.8157894736842115e-06, "loss": 3.3508, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.6003, "num_token_overlap": 11.6803, "num_token_query": 31.3583, "num_token_union": 64.9923, "num_word_context": 202.3678, "num_word_doc": 49.6791, "num_word_query": 23.2696, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22278.7049, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2266, "query_norm": 2.5845, "queue_k_norm": 2.663, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3583, "sent_len_1": 66.6003, "sent_len_max_0": 127.9775, "sent_len_max_1": 206.6225, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174100 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3439, "doc_norm": 2.6646, "encoder_q-embeddings": 9299.4883, "encoder_q-layer.0": 6533.6357, "encoder_q-layer.1": 6860.957, "encoder_q-layer.10": 13906.0332, "encoder_q-layer.11": 33723.8789, "encoder_q-layer.2": 7550.1943, "encoder_q-layer.3": 7700.6157, "encoder_q-layer.4": 8583.0586, "encoder_q-layer.5": 8563.8594, "encoder_q-layer.6": 10001.5811, "encoder_q-layer.7": 11135.0596, "encoder_q-layer.8": 13839.6045, "encoder_q-layer.9": 12871.0801, "epoch": 0.76, "inbatch_neg_score": 5.2274, "inbatch_pos_score": 5.8438, "learning_rate": 6.7894736842105264e-06, "loss": 3.3439, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.8356, "num_token_overlap": 11.6672, "num_token_query": 31.3963, "num_token_union": 65.1706, "num_word_context": 202.5009, "num_word_doc": 49.8201, "num_word_query": 23.3161, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20893.6969, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2266, "query_norm": 2.5918, "queue_k_norm": 2.6622, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3963, "sent_len_1": 66.8356, "sent_len_max_0": 127.985, "sent_len_max_1": 210.5488, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 174200 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.341, "doc_norm": 2.6627, "encoder_q-embeddings": 8919.4727, "encoder_q-layer.0": 6663.8638, "encoder_q-layer.1": 7108.8335, "encoder_q-layer.10": 15559.4131, "encoder_q-layer.11": 34668.3867, "encoder_q-layer.2": 7975.7979, "encoder_q-layer.3": 8064.5757, "encoder_q-layer.4": 8528.9717, "encoder_q-layer.5": 9105.6826, "encoder_q-layer.6": 9992.7715, "encoder_q-layer.7": 10995.1387, "encoder_q-layer.8": 13848.4746, "encoder_q-layer.9": 12557.21, "epoch": 0.76, "inbatch_neg_score": 5.2266, "inbatch_pos_score": 5.8398, "learning_rate": 6.763157894736842e-06, "loss": 3.341, "norm_diff": 0.0762, "norm_loss": 0.0, "num_token_doc": 66.7982, "num_token_overlap": 11.7199, "num_token_query": 31.426, "num_token_union": 65.116, "num_word_context": 202.297, "num_word_doc": 49.8619, "num_word_query": 23.3408, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21540.7205, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2266, "query_norm": 2.5864, "queue_k_norm": 2.6639, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.426, "sent_len_1": 66.7982, "sent_len_max_0": 127.9675, "sent_len_max_1": 208.6887, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174300 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3389, "doc_norm": 2.6654, "encoder_q-embeddings": 5061.6846, "encoder_q-layer.0": 3636.9639, "encoder_q-layer.1": 3879.022, "encoder_q-layer.10": 6676.0771, "encoder_q-layer.11": 17029.0625, "encoder_q-layer.2": 4146.8384, "encoder_q-layer.3": 4319.0493, "encoder_q-layer.4": 4645.104, "encoder_q-layer.5": 4688.5962, "encoder_q-layer.6": 5110.9937, "encoder_q-layer.7": 5608.8398, "encoder_q-layer.8": 6833.769, "encoder_q-layer.9": 6427.123, "epoch": 0.76, "inbatch_neg_score": 5.2283, "inbatch_pos_score": 5.8438, "learning_rate": 6.736842105263158e-06, "loss": 3.3389, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.665, "num_token_overlap": 11.6657, "num_token_query": 31.3122, "num_token_union": 65.0229, "num_word_context": 201.8923, "num_word_doc": 49.7518, "num_word_query": 23.2657, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10787.3545, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2266, "query_norm": 2.5885, "queue_k_norm": 2.6642, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3122, "sent_len_1": 66.665, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.0938, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174400 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3413, "doc_norm": 2.6648, "encoder_q-embeddings": 4946.5132, "encoder_q-layer.0": 3425.2822, "encoder_q-layer.1": 3616.4905, "encoder_q-layer.10": 6643.1494, "encoder_q-layer.11": 17243.875, "encoder_q-layer.2": 3966.2705, "encoder_q-layer.3": 4105.6934, "encoder_q-layer.4": 4432.4263, "encoder_q-layer.5": 4829.0366, "encoder_q-layer.6": 5302.3149, "encoder_q-layer.7": 6021.8223, "encoder_q-layer.8": 7225.9141, "encoder_q-layer.9": 6383.6982, "epoch": 0.76, "inbatch_neg_score": 5.2314, "inbatch_pos_score": 5.832, "learning_rate": 6.710526315789474e-06, "loss": 3.3413, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.7334, "num_token_overlap": 11.7147, "num_token_query": 31.4536, "num_token_union": 65.1154, "num_word_context": 202.3991, "num_word_doc": 49.8153, "num_word_query": 23.3475, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10982.8722, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2305, "query_norm": 2.5866, "queue_k_norm": 2.6648, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4536, "sent_len_1": 66.7334, "sent_len_max_0": 127.9825, "sent_len_max_1": 206.84, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174500 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.6636, "encoder_q-embeddings": 4489.2803, "encoder_q-layer.0": 3222.8984, "encoder_q-layer.1": 3386.9617, "encoder_q-layer.10": 7301.9331, "encoder_q-layer.11": 17999.5879, "encoder_q-layer.2": 3771.2688, "encoder_q-layer.3": 3904.8435, "encoder_q-layer.4": 4117.3447, "encoder_q-layer.5": 4294.9033, "encoder_q-layer.6": 5201.5103, "encoder_q-layer.7": 5908.4395, "encoder_q-layer.8": 7105.7983, "encoder_q-layer.9": 6692.7554, "epoch": 0.76, "inbatch_neg_score": 5.2304, "inbatch_pos_score": 5.8555, "learning_rate": 6.68421052631579e-06, "loss": 3.3508, "norm_diff": 0.0718, "norm_loss": 0.0, "num_token_doc": 66.8248, "num_token_overlap": 11.6294, "num_token_query": 31.2672, "num_token_union": 65.1114, "num_word_context": 202.3562, "num_word_doc": 49.8424, "num_word_query": 23.2137, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10734.8431, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2305, "query_norm": 2.5919, "queue_k_norm": 2.6644, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2672, "sent_len_1": 66.8248, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.6687, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 174600 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.3462, "doc_norm": 2.6632, "encoder_q-embeddings": 5030.6567, "encoder_q-layer.0": 3499.0688, "encoder_q-layer.1": 3735.7268, "encoder_q-layer.10": 7358.7583, "encoder_q-layer.11": 17669.6055, "encoder_q-layer.2": 4121.0957, "encoder_q-layer.3": 4246.335, "encoder_q-layer.4": 4500.2383, "encoder_q-layer.5": 4680.5698, "encoder_q-layer.6": 5160.2456, "encoder_q-layer.7": 6093.6567, "encoder_q-layer.8": 7158.7744, "encoder_q-layer.9": 6501.3457, "epoch": 0.76, "inbatch_neg_score": 5.2337, "inbatch_pos_score": 5.8242, "learning_rate": 6.657894736842106e-06, "loss": 3.3462, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.8468, "num_token_overlap": 11.678, "num_token_query": 31.3982, "num_token_union": 65.1742, "num_word_context": 202.5664, "num_word_doc": 49.8928, "num_word_query": 23.3251, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11052.0629, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2344, "query_norm": 2.5864, "queue_k_norm": 2.6656, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3982, "sent_len_1": 66.8468, "sent_len_max_0": 127.975, "sent_len_max_1": 207.8688, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174700 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3468, "doc_norm": 2.6644, "encoder_q-embeddings": 8054.8667, "encoder_q-layer.0": 5872.7441, "encoder_q-layer.1": 6418.1523, "encoder_q-layer.10": 7412.0947, "encoder_q-layer.11": 18290.1992, "encoder_q-layer.2": 7926.3818, "encoder_q-layer.3": 7523.7642, "encoder_q-layer.4": 7504.7188, "encoder_q-layer.5": 8219.2461, "encoder_q-layer.6": 7463.7319, "encoder_q-layer.7": 8069.0874, "encoder_q-layer.8": 8396.1416, "encoder_q-layer.9": 6780.9912, "epoch": 0.76, "inbatch_neg_score": 5.2325, "inbatch_pos_score": 5.8516, "learning_rate": 6.631578947368422e-06, "loss": 3.3468, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 66.8207, "num_token_overlap": 11.6636, "num_token_query": 31.314, "num_token_union": 65.1117, "num_word_context": 202.0188, "num_word_doc": 49.8144, "num_word_query": 23.2442, "postclip_grad_norm": 1.0, "preclip_grad_norm": 13638.3927, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2305, "query_norm": 2.5874, "queue_k_norm": 2.667, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.314, "sent_len_1": 66.8207, "sent_len_max_0": 127.9712, "sent_len_max_1": 210.3787, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 174800 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.3271, "doc_norm": 2.6674, "encoder_q-embeddings": 4642.3354, "encoder_q-layer.0": 3249.5364, "encoder_q-layer.1": 3450.7017, "encoder_q-layer.10": 6957.3535, "encoder_q-layer.11": 16999.6309, "encoder_q-layer.2": 3858.7036, "encoder_q-layer.3": 3928.1118, "encoder_q-layer.4": 4188.1309, "encoder_q-layer.5": 4313.8901, "encoder_q-layer.6": 4774.2568, "encoder_q-layer.7": 5682.9243, "encoder_q-layer.8": 6805.8496, "encoder_q-layer.9": 6610.064, "epoch": 0.76, "inbatch_neg_score": 5.2352, "inbatch_pos_score": 5.8555, "learning_rate": 6.605263157894736e-06, "loss": 3.3271, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.6747, "num_token_overlap": 11.7092, "num_token_query": 31.4216, "num_token_union": 65.0958, "num_word_context": 202.1885, "num_word_doc": 49.7826, "num_word_query": 23.3363, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10471.5007, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2344, "query_norm": 2.5914, "queue_k_norm": 2.6668, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4216, "sent_len_1": 66.6747, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.5475, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 174900 }, { "accuracy": 50.0977, "active_queue_size": 16384.0, "cl_loss": 3.3381, "doc_norm": 2.6715, "encoder_q-embeddings": 4939.7803, "encoder_q-layer.0": 3537.6028, "encoder_q-layer.1": 3915.8606, "encoder_q-layer.10": 6929.999, "encoder_q-layer.11": 17427.7285, "encoder_q-layer.2": 4355.1509, "encoder_q-layer.3": 4437.4951, "encoder_q-layer.4": 4603.4302, "encoder_q-layer.5": 4814.8618, "encoder_q-layer.6": 5372.8506, "encoder_q-layer.7": 5819.7158, "encoder_q-layer.8": 7250.4912, "encoder_q-layer.9": 6732.7056, "epoch": 0.76, "inbatch_neg_score": 5.2348, "inbatch_pos_score": 5.8398, "learning_rate": 6.578947368421053e-06, "loss": 3.3381, "norm_diff": 0.0885, "norm_loss": 0.0, "num_token_doc": 66.7558, "num_token_overlap": 11.6834, "num_token_query": 31.399, "num_token_union": 65.1334, "num_word_context": 202.3918, "num_word_doc": 49.8011, "num_word_query": 23.3135, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11083.5863, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2344, "query_norm": 2.5829, "queue_k_norm": 2.666, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.399, "sent_len_1": 66.7558, "sent_len_max_0": 127.9838, "sent_len_max_1": 210.9325, "stdk": 0.0499, "stdq": 0.0427, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175000 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3478, "doc_norm": 2.6655, "encoder_q-embeddings": 4795.4492, "encoder_q-layer.0": 3369.1711, "encoder_q-layer.1": 3714.3567, "encoder_q-layer.10": 6862.2295, "encoder_q-layer.11": 16666.9316, "encoder_q-layer.2": 4192.29, "encoder_q-layer.3": 4267.7046, "encoder_q-layer.4": 4600.6953, "encoder_q-layer.5": 4940.5249, "encoder_q-layer.6": 5276.1064, "encoder_q-layer.7": 6064.5894, "encoder_q-layer.8": 7110.7754, "encoder_q-layer.9": 6313.3628, "epoch": 0.76, "inbatch_neg_score": 5.237, "inbatch_pos_score": 5.8594, "learning_rate": 6.552631578947369e-06, "loss": 3.3478, "norm_diff": 0.0764, "norm_loss": 0.0, "num_token_doc": 66.8652, "num_token_overlap": 11.7103, "num_token_query": 31.4199, "num_token_union": 65.1719, "num_word_context": 202.4397, "num_word_doc": 49.8988, "num_word_query": 23.343, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10637.6564, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2383, "query_norm": 2.5891, "queue_k_norm": 2.6672, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4199, "sent_len_1": 66.8652, "sent_len_max_0": 127.995, "sent_len_max_1": 208.4925, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175100 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3452, "doc_norm": 2.6649, "encoder_q-embeddings": 5033.8589, "encoder_q-layer.0": 3588.5015, "encoder_q-layer.1": 3754.8982, "encoder_q-layer.10": 6865.3511, "encoder_q-layer.11": 17633.9219, "encoder_q-layer.2": 4129.0703, "encoder_q-layer.3": 4179.0356, "encoder_q-layer.4": 4536.2031, "encoder_q-layer.5": 4661.1123, "encoder_q-layer.6": 5099.7295, "encoder_q-layer.7": 5596.3174, "encoder_q-layer.8": 6782.354, "encoder_q-layer.9": 6474.9878, "epoch": 0.76, "inbatch_neg_score": 5.2382, "inbatch_pos_score": 5.8477, "learning_rate": 6.526315789473685e-06, "loss": 3.3452, "norm_diff": 0.0752, "norm_loss": 0.0, "num_token_doc": 66.8013, "num_token_overlap": 11.6684, "num_token_query": 31.3984, "num_token_union": 65.1701, "num_word_context": 202.5024, "num_word_doc": 49.8644, "num_word_query": 23.3406, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10865.6877, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2383, "query_norm": 2.5897, "queue_k_norm": 2.6673, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3984, "sent_len_1": 66.8013, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.8025, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175200 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3399, "doc_norm": 2.6669, "encoder_q-embeddings": 4987.8921, "encoder_q-layer.0": 3652.3591, "encoder_q-layer.1": 3858.8958, "encoder_q-layer.10": 6587.0269, "encoder_q-layer.11": 16675.4961, "encoder_q-layer.2": 4323.1094, "encoder_q-layer.3": 4327.7031, "encoder_q-layer.4": 4752.6084, "encoder_q-layer.5": 4911.3813, "encoder_q-layer.6": 5281.9712, "encoder_q-layer.7": 5868.3438, "encoder_q-layer.8": 6752.2568, "encoder_q-layer.9": 6253.7349, "epoch": 0.76, "inbatch_neg_score": 5.2397, "inbatch_pos_score": 5.8438, "learning_rate": 6.5000000000000004e-06, "loss": 3.3399, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7442, "num_token_overlap": 11.6903, "num_token_query": 31.4335, "num_token_union": 65.1333, "num_word_context": 202.3173, "num_word_doc": 49.8028, "num_word_query": 23.3543, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10672.0157, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2383, "query_norm": 2.5873, "queue_k_norm": 2.6664, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4335, "sent_len_1": 66.7442, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.5325, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 175300 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3412, "doc_norm": 2.6675, "encoder_q-embeddings": 4958.7754, "encoder_q-layer.0": 3652.7085, "encoder_q-layer.1": 4038.5396, "encoder_q-layer.10": 6454.6919, "encoder_q-layer.11": 16612.2441, "encoder_q-layer.2": 4501.3979, "encoder_q-layer.3": 4606.3535, "encoder_q-layer.4": 5073.2588, "encoder_q-layer.5": 5221.3711, "encoder_q-layer.6": 6156.6899, "encoder_q-layer.7": 6293.9658, "encoder_q-layer.8": 6742.6865, "encoder_q-layer.9": 6293.0215, "epoch": 0.76, "inbatch_neg_score": 5.241, "inbatch_pos_score": 5.8594, "learning_rate": 6.473684210526316e-06, "loss": 3.3412, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.8099, "num_token_overlap": 11.687, "num_token_query": 31.4535, "num_token_union": 65.1806, "num_word_context": 202.2054, "num_word_doc": 49.8813, "num_word_query": 23.3688, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10789.1194, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2422, "query_norm": 2.5931, "queue_k_norm": 2.6684, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4535, "sent_len_1": 66.8099, "sent_len_max_0": 127.9775, "sent_len_max_1": 207.6738, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175400 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3379, "doc_norm": 2.6679, "encoder_q-embeddings": 4741.606, "encoder_q-layer.0": 3334.0007, "encoder_q-layer.1": 3477.9907, "encoder_q-layer.10": 7363.5059, "encoder_q-layer.11": 17443.9375, "encoder_q-layer.2": 3965.6091, "encoder_q-layer.3": 4039.7217, "encoder_q-layer.4": 4476.9609, "encoder_q-layer.5": 4414.8291, "encoder_q-layer.6": 5090.7817, "encoder_q-layer.7": 5835.6953, "encoder_q-layer.8": 6803.9331, "encoder_q-layer.9": 6381.3608, "epoch": 0.76, "inbatch_neg_score": 5.2415, "inbatch_pos_score": 5.8555, "learning_rate": 6.447368421052632e-06, "loss": 3.3379, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.8077, "num_token_overlap": 11.6701, "num_token_query": 31.4131, "num_token_union": 65.1716, "num_word_context": 202.5051, "num_word_doc": 49.8652, "num_word_query": 23.3304, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10824.5934, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2422, "query_norm": 2.591, "queue_k_norm": 2.6696, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4131, "sent_len_1": 66.8077, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.29, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 175500 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.338, "doc_norm": 2.6686, "encoder_q-embeddings": 4941.8833, "encoder_q-layer.0": 3518.7844, "encoder_q-layer.1": 3740.3247, "encoder_q-layer.10": 7193.8311, "encoder_q-layer.11": 17367.0527, "encoder_q-layer.2": 4251.335, "encoder_q-layer.3": 4358.7241, "encoder_q-layer.4": 4663.5005, "encoder_q-layer.5": 4798.4023, "encoder_q-layer.6": 5458.2441, "encoder_q-layer.7": 6125.2236, "encoder_q-layer.8": 6920.5811, "encoder_q-layer.9": 6574.624, "epoch": 0.76, "inbatch_neg_score": 5.2453, "inbatch_pos_score": 5.8594, "learning_rate": 6.421052631578947e-06, "loss": 3.338, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.686, "num_token_overlap": 11.6701, "num_token_query": 31.3681, "num_token_union": 65.087, "num_word_context": 202.5144, "num_word_doc": 49.7658, "num_word_query": 23.2958, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10922.4294, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2461, "query_norm": 2.5944, "queue_k_norm": 2.6688, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3681, "sent_len_1": 66.686, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.5525, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175600 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3421, "doc_norm": 2.6687, "encoder_q-embeddings": 4897.168, "encoder_q-layer.0": 3553.874, "encoder_q-layer.1": 3782.9705, "encoder_q-layer.10": 6715.2095, "encoder_q-layer.11": 17232.1484, "encoder_q-layer.2": 4247.1079, "encoder_q-layer.3": 4426.5044, "encoder_q-layer.4": 4624.4053, "encoder_q-layer.5": 4739.2007, "encoder_q-layer.6": 5365.1421, "encoder_q-layer.7": 5798.9946, "encoder_q-layer.8": 6825.1719, "encoder_q-layer.9": 6430.6953, "epoch": 0.76, "inbatch_neg_score": 5.2478, "inbatch_pos_score": 5.8594, "learning_rate": 6.394736842105263e-06, "loss": 3.3421, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.7386, "num_token_overlap": 11.7003, "num_token_query": 31.4063, "num_token_union": 65.1148, "num_word_context": 202.0985, "num_word_doc": 49.7971, "num_word_query": 23.3292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10787.2557, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.25, "query_norm": 2.5913, "queue_k_norm": 2.6685, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4063, "sent_len_1": 66.7386, "sent_len_max_0": 127.975, "sent_len_max_1": 208.8537, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175700 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3435, "doc_norm": 2.6695, "encoder_q-embeddings": 4859.0098, "encoder_q-layer.0": 3503.134, "encoder_q-layer.1": 3713.9653, "encoder_q-layer.10": 7566.4419, "encoder_q-layer.11": 17716.2891, "encoder_q-layer.2": 4151.4546, "encoder_q-layer.3": 4177.1265, "encoder_q-layer.4": 4587.0415, "encoder_q-layer.5": 4835.3848, "encoder_q-layer.6": 5139.7729, "encoder_q-layer.7": 5912.4746, "encoder_q-layer.8": 6984.1133, "encoder_q-layer.9": 6649.9287, "epoch": 0.76, "inbatch_neg_score": 5.2505, "inbatch_pos_score": 5.8672, "learning_rate": 6.368421052631579e-06, "loss": 3.3435, "norm_diff": 0.0764, "norm_loss": 0.0, "num_token_doc": 66.8983, "num_token_overlap": 11.7147, "num_token_query": 31.4399, "num_token_union": 65.1778, "num_word_context": 202.6021, "num_word_doc": 49.8967, "num_word_query": 23.3535, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10900.5174, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.25, "query_norm": 2.593, "queue_k_norm": 2.6695, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4399, "sent_len_1": 66.8983, "sent_len_max_0": 127.98, "sent_len_max_1": 208.9663, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 175800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3427, "doc_norm": 2.6709, "encoder_q-embeddings": 4566.7695, "encoder_q-layer.0": 3362.6782, "encoder_q-layer.1": 3505.8267, "encoder_q-layer.10": 6749.5415, "encoder_q-layer.11": 17001.5703, "encoder_q-layer.2": 3879.1089, "encoder_q-layer.3": 4027.6636, "encoder_q-layer.4": 4411.2461, "encoder_q-layer.5": 4518.376, "encoder_q-layer.6": 4934.1533, "encoder_q-layer.7": 5485.0269, "encoder_q-layer.8": 6828.3262, "encoder_q-layer.9": 6111.2295, "epoch": 0.76, "inbatch_neg_score": 5.2533, "inbatch_pos_score": 5.8672, "learning_rate": 6.342105263157895e-06, "loss": 3.3427, "norm_diff": 0.0783, "norm_loss": 0.0, "num_token_doc": 66.7899, "num_token_overlap": 11.6766, "num_token_query": 31.3881, "num_token_union": 65.1205, "num_word_context": 202.2648, "num_word_doc": 49.8279, "num_word_query": 23.3125, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10430.2849, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2539, "query_norm": 2.5926, "queue_k_norm": 2.6687, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3881, "sent_len_1": 66.7899, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.3363, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 175900 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3327, "doc_norm": 2.6677, "encoder_q-embeddings": 4724.3877, "encoder_q-layer.0": 3241.5703, "encoder_q-layer.1": 3498.5264, "encoder_q-layer.10": 7134.6865, "encoder_q-layer.11": 17782.8164, "encoder_q-layer.2": 3962.0159, "encoder_q-layer.3": 4100.499, "encoder_q-layer.4": 4457.9268, "encoder_q-layer.5": 4770.7383, "encoder_q-layer.6": 5371.3335, "encoder_q-layer.7": 6089.1963, "encoder_q-layer.8": 7444.2002, "encoder_q-layer.9": 6803.5239, "epoch": 0.76, "inbatch_neg_score": 5.2584, "inbatch_pos_score": 5.8672, "learning_rate": 6.315789473684211e-06, "loss": 3.3327, "norm_diff": 0.0748, "norm_loss": 0.0, "num_token_doc": 66.7917, "num_token_overlap": 11.6944, "num_token_query": 31.5008, "num_token_union": 65.2103, "num_word_context": 202.3288, "num_word_doc": 49.8554, "num_word_query": 23.4113, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11201.3521, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2578, "query_norm": 2.5929, "queue_k_norm": 2.6712, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5008, "sent_len_1": 66.7917, "sent_len_max_0": 127.9562, "sent_len_max_1": 206.8313, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 176000 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3489, "doc_norm": 2.6695, "encoder_q-embeddings": 4876.8623, "encoder_q-layer.0": 3434.8191, "encoder_q-layer.1": 3569.6641, "encoder_q-layer.10": 7367.4458, "encoder_q-layer.11": 18344.0527, "encoder_q-layer.2": 3863.3088, "encoder_q-layer.3": 4111.2856, "encoder_q-layer.4": 4363.1709, "encoder_q-layer.5": 4558.3218, "encoder_q-layer.6": 5108.7725, "encoder_q-layer.7": 5859.917, "encoder_q-layer.8": 7010.6211, "encoder_q-layer.9": 6559.9512, "epoch": 0.76, "inbatch_neg_score": 5.2602, "inbatch_pos_score": 5.8711, "learning_rate": 6.289473684210527e-06, "loss": 3.3489, "norm_diff": 0.0755, "norm_loss": 0.0, "num_token_doc": 66.817, "num_token_overlap": 11.6715, "num_token_query": 31.3972, "num_token_union": 65.1697, "num_word_context": 202.3645, "num_word_doc": 49.8457, "num_word_query": 23.3243, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10994.4329, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2617, "query_norm": 2.5941, "queue_k_norm": 2.6697, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3972, "sent_len_1": 66.817, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.9275, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 176100 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3203, "doc_norm": 2.6693, "encoder_q-embeddings": 4797.2871, "encoder_q-layer.0": 3346.8035, "encoder_q-layer.1": 3528.6472, "encoder_q-layer.10": 6784.9722, "encoder_q-layer.11": 17650.0488, "encoder_q-layer.2": 3925.2129, "encoder_q-layer.3": 4083.3389, "encoder_q-layer.4": 4311.7476, "encoder_q-layer.5": 4554.3896, "encoder_q-layer.6": 5267.2329, "encoder_q-layer.7": 5870.3535, "encoder_q-layer.8": 7181.2598, "encoder_q-layer.9": 6552.2466, "epoch": 0.76, "inbatch_neg_score": 5.2653, "inbatch_pos_score": 5.8828, "learning_rate": 6.263157894736843e-06, "loss": 3.3203, "norm_diff": 0.0683, "norm_loss": 0.0, "num_token_doc": 66.7133, "num_token_overlap": 11.677, "num_token_query": 31.3733, "num_token_union": 65.0897, "num_word_context": 201.866, "num_word_doc": 49.7747, "num_word_query": 23.3017, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11099.0946, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2656, "query_norm": 2.601, "queue_k_norm": 2.6705, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3733, "sent_len_1": 66.7133, "sent_len_max_0": 127.9625, "sent_len_max_1": 209.3938, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176200 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3378, "doc_norm": 2.6705, "encoder_q-embeddings": 4809.4111, "encoder_q-layer.0": 3349.9458, "encoder_q-layer.1": 3608.7764, "encoder_q-layer.10": 6631.7612, "encoder_q-layer.11": 16419.7031, "encoder_q-layer.2": 4161.3862, "encoder_q-layer.3": 4145.9937, "encoder_q-layer.4": 4489.1025, "encoder_q-layer.5": 4649.7358, "encoder_q-layer.6": 5251.0132, "encoder_q-layer.7": 5939.0742, "encoder_q-layer.8": 7053.71, "encoder_q-layer.9": 6660.6289, "epoch": 0.77, "inbatch_neg_score": 5.2659, "inbatch_pos_score": 5.8906, "learning_rate": 6.236842105263159e-06, "loss": 3.3378, "norm_diff": 0.0715, "norm_loss": 0.0, "num_token_doc": 66.6793, "num_token_overlap": 11.6804, "num_token_query": 31.4158, "num_token_union": 65.1083, "num_word_context": 202.2579, "num_word_doc": 49.75, "num_word_query": 23.329, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10518.2923, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.2656, "query_norm": 2.599, "queue_k_norm": 2.67, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4158, "sent_len_1": 66.6793, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7388, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 176300 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3315, "doc_norm": 2.67, "encoder_q-embeddings": 9773.8418, "encoder_q-layer.0": 6972.5664, "encoder_q-layer.1": 7406.1792, "encoder_q-layer.10": 16836.0137, "encoder_q-layer.11": 34673.082, "encoder_q-layer.2": 8051.8257, "encoder_q-layer.3": 8269.8652, "encoder_q-layer.4": 9039.498, "encoder_q-layer.5": 9229.582, "encoder_q-layer.6": 10335.7363, "encoder_q-layer.7": 11603.1152, "encoder_q-layer.8": 13708.8486, "encoder_q-layer.9": 13327.4727, "epoch": 0.77, "inbatch_neg_score": 5.2693, "inbatch_pos_score": 5.8672, "learning_rate": 6.2105263157894745e-06, "loss": 3.3315, "norm_diff": 0.0735, "norm_loss": 0.0, "num_token_doc": 66.8269, "num_token_overlap": 11.6718, "num_token_query": 31.3701, "num_token_union": 65.1429, "num_word_context": 202.2285, "num_word_doc": 49.8476, "num_word_query": 23.2895, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21904.3118, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2695, "query_norm": 2.5965, "queue_k_norm": 2.6714, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3701, "sent_len_1": 66.8269, "sent_len_max_0": 127.95, "sent_len_max_1": 210.07, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176400 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.336, "doc_norm": 2.6719, "encoder_q-embeddings": 9818.2861, "encoder_q-layer.0": 6867.7852, "encoder_q-layer.1": 7568.8252, "encoder_q-layer.10": 14814.5703, "encoder_q-layer.11": 35433.8164, "encoder_q-layer.2": 8395.1602, "encoder_q-layer.3": 8372.7305, "encoder_q-layer.4": 8803.1465, "encoder_q-layer.5": 8774.0732, "encoder_q-layer.6": 9872.8262, "encoder_q-layer.7": 11503.7617, "encoder_q-layer.8": 14478.8779, "encoder_q-layer.9": 13497.4297, "epoch": 0.77, "inbatch_neg_score": 5.2747, "inbatch_pos_score": 5.9023, "learning_rate": 6.184210526315789e-06, "loss": 3.336, "norm_diff": 0.0697, "norm_loss": 0.0, "num_token_doc": 66.8758, "num_token_overlap": 11.6833, "num_token_query": 31.3599, "num_token_union": 65.1516, "num_word_context": 202.2655, "num_word_doc": 49.867, "num_word_query": 23.2964, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21899.1054, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2734, "query_norm": 2.6023, "queue_k_norm": 2.6725, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3599, "sent_len_1": 66.8758, "sent_len_max_0": 127.9575, "sent_len_max_1": 210.1575, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176500 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3371, "doc_norm": 2.6721, "encoder_q-embeddings": 9390.832, "encoder_q-layer.0": 6674.834, "encoder_q-layer.1": 7201.2095, "encoder_q-layer.10": 14931.4502, "encoder_q-layer.11": 34798.3164, "encoder_q-layer.2": 7935.4258, "encoder_q-layer.3": 8278.373, "encoder_q-layer.4": 8602.0225, "encoder_q-layer.5": 8945.8145, "encoder_q-layer.6": 10312.249, "encoder_q-layer.7": 11875.4551, "encoder_q-layer.8": 14754.6055, "encoder_q-layer.9": 13568.0596, "epoch": 0.77, "inbatch_neg_score": 5.2765, "inbatch_pos_score": 5.8867, "learning_rate": 6.157894736842105e-06, "loss": 3.3371, "norm_diff": 0.0687, "norm_loss": 0.0, "num_token_doc": 66.672, "num_token_overlap": 11.6709, "num_token_query": 31.3406, "num_token_union": 65.0421, "num_word_context": 202.1007, "num_word_doc": 49.7488, "num_word_query": 23.2775, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22015.2257, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2773, "query_norm": 2.6033, "queue_k_norm": 2.6724, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3406, "sent_len_1": 66.672, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.6025, "stdk": 0.0495, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176600 }, { "accuracy": 53.3203, "active_queue_size": 16384.0, "cl_loss": 3.3403, "doc_norm": 2.6722, "encoder_q-embeddings": 9039.9971, "encoder_q-layer.0": 6320.8486, "encoder_q-layer.1": 6875.0972, "encoder_q-layer.10": 12829.6562, "encoder_q-layer.11": 33223.2734, "encoder_q-layer.2": 7643.6353, "encoder_q-layer.3": 7890.4004, "encoder_q-layer.4": 8449.6094, "encoder_q-layer.5": 8796.7773, "encoder_q-layer.6": 9894.7197, "encoder_q-layer.7": 11143.6826, "encoder_q-layer.8": 13604.8359, "encoder_q-layer.9": 13053.5947, "epoch": 0.77, "inbatch_neg_score": 5.2802, "inbatch_pos_score": 5.9102, "learning_rate": 6.131578947368422e-06, "loss": 3.3403, "norm_diff": 0.073, "norm_loss": 0.0, "num_token_doc": 66.7598, "num_token_overlap": 11.6616, "num_token_query": 31.427, "num_token_union": 65.1219, "num_word_context": 202.1775, "num_word_doc": 49.8023, "num_word_query": 23.3395, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20743.9315, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2812, "query_norm": 2.5993, "queue_k_norm": 2.6728, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.427, "sent_len_1": 66.7598, "sent_len_max_0": 127.975, "sent_len_max_1": 211.4075, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176700 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3374, "doc_norm": 2.6717, "encoder_q-embeddings": 10277.0469, "encoder_q-layer.0": 7249.562, "encoder_q-layer.1": 7839.6929, "encoder_q-layer.10": 14401.1279, "encoder_q-layer.11": 36048.6523, "encoder_q-layer.2": 8898.6816, "encoder_q-layer.3": 8947.0684, "encoder_q-layer.4": 9492.6172, "encoder_q-layer.5": 9335.7109, "encoder_q-layer.6": 10410.7793, "encoder_q-layer.7": 11561.6279, "encoder_q-layer.8": 14571.0059, "encoder_q-layer.9": 13301.9453, "epoch": 0.77, "inbatch_neg_score": 5.2837, "inbatch_pos_score": 5.8984, "learning_rate": 6.105263157894737e-06, "loss": 3.3374, "norm_diff": 0.0678, "norm_loss": 0.0, "num_token_doc": 66.8071, "num_token_overlap": 11.7136, "num_token_query": 31.4607, "num_token_union": 65.1415, "num_word_context": 202.1915, "num_word_doc": 49.8665, "num_word_query": 23.3715, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22725.7882, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2812, "query_norm": 2.6039, "queue_k_norm": 2.6733, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4607, "sent_len_1": 66.8071, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.2612, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176800 }, { "accuracy": 50.8789, "active_queue_size": 16384.0, "cl_loss": 3.3521, "doc_norm": 2.675, "encoder_q-embeddings": 12095.1006, "encoder_q-layer.0": 8912.2295, "encoder_q-layer.1": 9438.7822, "encoder_q-layer.10": 14380.9912, "encoder_q-layer.11": 33457.5938, "encoder_q-layer.2": 10875.5029, "encoder_q-layer.3": 11151.8926, "encoder_q-layer.4": 11705.5791, "encoder_q-layer.5": 12028.6641, "encoder_q-layer.6": 13407.4648, "encoder_q-layer.7": 13922.5039, "encoder_q-layer.8": 15178.6816, "encoder_q-layer.9": 13333.2539, "epoch": 0.77, "inbatch_neg_score": 5.2864, "inbatch_pos_score": 5.9062, "learning_rate": 6.078947368421053e-06, "loss": 3.3521, "norm_diff": 0.0719, "norm_loss": 0.0, "num_token_doc": 66.7353, "num_token_overlap": 11.6708, "num_token_query": 31.2692, "num_token_union": 65.0219, "num_word_context": 202.3071, "num_word_doc": 49.8096, "num_word_query": 23.2143, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23166.8393, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2852, "query_norm": 2.6031, "queue_k_norm": 2.6737, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2692, "sent_len_1": 66.7353, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.6775, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 176900 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3366, "doc_norm": 2.6747, "encoder_q-embeddings": 9911.8701, "encoder_q-layer.0": 6805.0088, "encoder_q-layer.1": 7312.9976, "encoder_q-layer.10": 13991.9971, "encoder_q-layer.11": 34103.8984, "encoder_q-layer.2": 8275.2314, "encoder_q-layer.3": 8687.5283, "encoder_q-layer.4": 9372.3252, "encoder_q-layer.5": 9629.4678, "encoder_q-layer.6": 10894.6143, "encoder_q-layer.7": 11726.9707, "encoder_q-layer.8": 14819.8594, "encoder_q-layer.9": 13015.4482, "epoch": 0.77, "inbatch_neg_score": 5.2896, "inbatch_pos_score": 5.9102, "learning_rate": 6.0526315789473685e-06, "loss": 3.3366, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.8378, "num_token_overlap": 11.6805, "num_token_query": 31.3197, "num_token_union": 65.1312, "num_word_context": 202.3218, "num_word_doc": 49.8303, "num_word_query": 23.2648, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21632.8146, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2891, "query_norm": 2.6024, "queue_k_norm": 2.6743, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3197, "sent_len_1": 66.8378, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.975, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177000 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3428, "doc_norm": 2.675, "encoder_q-embeddings": 9916.3857, "encoder_q-layer.0": 6869.0586, "encoder_q-layer.1": 7400.0986, "encoder_q-layer.10": 14143.8086, "encoder_q-layer.11": 35030.668, "encoder_q-layer.2": 8312.6523, "encoder_q-layer.3": 8494.1396, "encoder_q-layer.4": 9182.541, "encoder_q-layer.5": 9453.0371, "encoder_q-layer.6": 10999.9883, "encoder_q-layer.7": 11883.2539, "encoder_q-layer.8": 14159.4922, "encoder_q-layer.9": 13091.125, "epoch": 0.77, "inbatch_neg_score": 5.2914, "inbatch_pos_score": 5.9023, "learning_rate": 6.026315789473685e-06, "loss": 3.3428, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.7982, "num_token_overlap": 11.6976, "num_token_query": 31.4077, "num_token_union": 65.1392, "num_word_context": 202.4104, "num_word_doc": 49.8382, "num_word_query": 23.3148, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22065.6999, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.293, "query_norm": 2.6014, "queue_k_norm": 2.676, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4077, "sent_len_1": 66.7982, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.2125, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 177100 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3446, "doc_norm": 2.6773, "encoder_q-embeddings": 9553.5029, "encoder_q-layer.0": 7121.3804, "encoder_q-layer.1": 7521.3955, "encoder_q-layer.10": 14850.3652, "encoder_q-layer.11": 36441.0664, "encoder_q-layer.2": 8294.7598, "encoder_q-layer.3": 8550.0635, "encoder_q-layer.4": 9044.0566, "encoder_q-layer.5": 9260.2021, "encoder_q-layer.6": 10574.0225, "encoder_q-layer.7": 12033.1299, "encoder_q-layer.8": 14503.876, "encoder_q-layer.9": 13260.8047, "epoch": 0.77, "inbatch_neg_score": 5.2958, "inbatch_pos_score": 5.918, "learning_rate": 6e-06, "loss": 3.3446, "norm_diff": 0.0751, "norm_loss": 0.0, "num_token_doc": 66.8583, "num_token_overlap": 11.6908, "num_token_query": 31.3728, "num_token_union": 65.1775, "num_word_context": 202.6718, "num_word_doc": 49.8905, "num_word_query": 23.3163, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22740.4579, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.2969, "query_norm": 2.6023, "queue_k_norm": 2.6767, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3728, "sent_len_1": 66.8583, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.2837, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 177200 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3384, "doc_norm": 2.6729, "encoder_q-embeddings": 9659.8916, "encoder_q-layer.0": 6682.6533, "encoder_q-layer.1": 7443.729, "encoder_q-layer.10": 16573.9102, "encoder_q-layer.11": 36186.332, "encoder_q-layer.2": 7995.0103, "encoder_q-layer.3": 7993.7495, "encoder_q-layer.4": 8429.3213, "encoder_q-layer.5": 8701.291, "encoder_q-layer.6": 9935.0352, "encoder_q-layer.7": 11492.417, "encoder_q-layer.8": 14779.666, "encoder_q-layer.9": 14468.1055, "epoch": 0.77, "inbatch_neg_score": 5.3004, "inbatch_pos_score": 5.9102, "learning_rate": 5.973684210526316e-06, "loss": 3.3384, "norm_diff": 0.0667, "norm_loss": 0.0, "num_token_doc": 66.6549, "num_token_overlap": 11.675, "num_token_query": 31.4439, "num_token_union": 65.1083, "num_word_context": 202.141, "num_word_doc": 49.7407, "num_word_query": 23.3571, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21991.1733, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3008, "query_norm": 2.6062, "queue_k_norm": 2.6758, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4439, "sent_len_1": 66.6549, "sent_len_max_0": 127.9575, "sent_len_max_1": 209.0525, "stdk": 0.0493, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177300 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3403, "doc_norm": 2.6723, "encoder_q-embeddings": 9612.3057, "encoder_q-layer.0": 6844.6841, "encoder_q-layer.1": 7192.6724, "encoder_q-layer.10": 14350.2646, "encoder_q-layer.11": 35893.7773, "encoder_q-layer.2": 8074.2969, "encoder_q-layer.3": 8263.75, "encoder_q-layer.4": 8905.1152, "encoder_q-layer.5": 9090.4785, "encoder_q-layer.6": 10444.8105, "encoder_q-layer.7": 12137.8164, "encoder_q-layer.8": 13712.5625, "encoder_q-layer.9": 13187.0645, "epoch": 0.77, "inbatch_neg_score": 5.3025, "inbatch_pos_score": 5.9102, "learning_rate": 5.947368421052632e-06, "loss": 3.3403, "norm_diff": 0.07, "norm_loss": 0.0, "num_token_doc": 66.7081, "num_token_overlap": 11.6891, "num_token_query": 31.4218, "num_token_union": 65.0997, "num_word_context": 202.3671, "num_word_doc": 49.791, "num_word_query": 23.3586, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22154.2217, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3047, "query_norm": 2.6023, "queue_k_norm": 2.6763, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4218, "sent_len_1": 66.7081, "sent_len_max_0": 127.99, "sent_len_max_1": 209.3338, "stdk": 0.0492, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177400 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3295, "doc_norm": 2.6781, "encoder_q-embeddings": 10946.2949, "encoder_q-layer.0": 7909.5342, "encoder_q-layer.1": 8533.5146, "encoder_q-layer.10": 13695.2402, "encoder_q-layer.11": 34471.4219, "encoder_q-layer.2": 8908.6387, "encoder_q-layer.3": 9102.0879, "encoder_q-layer.4": 9556.2305, "encoder_q-layer.5": 9816.3135, "encoder_q-layer.6": 10385.0029, "encoder_q-layer.7": 11214.6133, "encoder_q-layer.8": 13265.498, "encoder_q-layer.9": 12867.3848, "epoch": 0.77, "inbatch_neg_score": 5.3025, "inbatch_pos_score": 5.918, "learning_rate": 5.921052631578948e-06, "loss": 3.3295, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.7907, "num_token_overlap": 11.7206, "num_token_query": 31.4078, "num_token_union": 65.1005, "num_word_context": 202.1824, "num_word_doc": 49.8474, "num_word_query": 23.3296, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21541.9556, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3047, "query_norm": 2.6027, "queue_k_norm": 2.6767, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4078, "sent_len_1": 66.7907, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.595, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177500 }, { "accuracy": 53.8086, "active_queue_size": 16384.0, "cl_loss": 3.3416, "doc_norm": 2.6785, "encoder_q-embeddings": 9815.6621, "encoder_q-layer.0": 7113.729, "encoder_q-layer.1": 7756.4458, "encoder_q-layer.10": 13935.9502, "encoder_q-layer.11": 34900.8984, "encoder_q-layer.2": 8932.9971, "encoder_q-layer.3": 9109.2686, "encoder_q-layer.4": 9790.7471, "encoder_q-layer.5": 9915.0127, "encoder_q-layer.6": 11458.5117, "encoder_q-layer.7": 12954.0459, "encoder_q-layer.8": 14700.9834, "encoder_q-layer.9": 12983.1592, "epoch": 0.77, "inbatch_neg_score": 5.3034, "inbatch_pos_score": 5.9297, "learning_rate": 5.8947368421052634e-06, "loss": 3.3416, "norm_diff": 0.0712, "norm_loss": 0.0, "num_token_doc": 66.6659, "num_token_overlap": 11.6739, "num_token_query": 31.4347, "num_token_union": 65.0965, "num_word_context": 202.3182, "num_word_doc": 49.7499, "num_word_query": 23.3538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22537.5611, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3047, "query_norm": 2.6073, "queue_k_norm": 2.6774, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4347, "sent_len_1": 66.6659, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.705, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177600 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3419, "doc_norm": 2.6766, "encoder_q-embeddings": 9351.3662, "encoder_q-layer.0": 6760.0225, "encoder_q-layer.1": 7255.7227, "encoder_q-layer.10": 13802.3164, "encoder_q-layer.11": 32737.0879, "encoder_q-layer.2": 7749.2397, "encoder_q-layer.3": 8021.478, "encoder_q-layer.4": 8557.8906, "encoder_q-layer.5": 8770.1152, "encoder_q-layer.6": 9883.7324, "encoder_q-layer.7": 11509.0664, "encoder_q-layer.8": 13554.2568, "encoder_q-layer.9": 12117.2236, "epoch": 0.77, "inbatch_neg_score": 5.3043, "inbatch_pos_score": 5.9297, "learning_rate": 5.868421052631579e-06, "loss": 3.3419, "norm_diff": 0.0746, "norm_loss": 0.0, "num_token_doc": 66.6244, "num_token_overlap": 11.6908, "num_token_query": 31.3657, "num_token_union": 65.0142, "num_word_context": 202.1152, "num_word_doc": 49.6988, "num_word_query": 23.3026, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20475.2467, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3047, "query_norm": 2.602, "queue_k_norm": 2.6775, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3657, "sent_len_1": 66.6244, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.8388, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 177700 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3489, "doc_norm": 2.6792, "encoder_q-embeddings": 10330.4746, "encoder_q-layer.0": 7052.0278, "encoder_q-layer.1": 7417.3159, "encoder_q-layer.10": 13596.9375, "encoder_q-layer.11": 33572.2148, "encoder_q-layer.2": 8167.7847, "encoder_q-layer.3": 8518.043, "encoder_q-layer.4": 9150.8418, "encoder_q-layer.5": 9269.2979, "encoder_q-layer.6": 10571.4082, "encoder_q-layer.7": 12309.7324, "encoder_q-layer.8": 13579.7695, "encoder_q-layer.9": 12822.1934, "epoch": 0.77, "inbatch_neg_score": 5.3068, "inbatch_pos_score": 5.9219, "learning_rate": 5.842105263157895e-06, "loss": 3.3489, "norm_diff": 0.0757, "norm_loss": 0.0, "num_token_doc": 66.7673, "num_token_overlap": 11.7002, "num_token_query": 31.4284, "num_token_union": 65.0929, "num_word_context": 202.4263, "num_word_doc": 49.8098, "num_word_query": 23.3598, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21358.43, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3086, "query_norm": 2.6035, "queue_k_norm": 2.6784, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4284, "sent_len_1": 66.7673, "sent_len_max_0": 127.9925, "sent_len_max_1": 210.6163, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177800 }, { "accuracy": 49.6582, "active_queue_size": 16384.0, "cl_loss": 3.3434, "doc_norm": 2.6785, "encoder_q-embeddings": 10785.5791, "encoder_q-layer.0": 7693.3237, "encoder_q-layer.1": 8407.4131, "encoder_q-layer.10": 13655.708, "encoder_q-layer.11": 34316.5586, "encoder_q-layer.2": 9659.5879, "encoder_q-layer.3": 10085.6846, "encoder_q-layer.4": 10972.9336, "encoder_q-layer.5": 11401.0547, "encoder_q-layer.6": 12268.916, "encoder_q-layer.7": 13014.709, "encoder_q-layer.8": 14377.4717, "encoder_q-layer.9": 12701.2188, "epoch": 0.77, "inbatch_neg_score": 5.3072, "inbatch_pos_score": 5.9141, "learning_rate": 5.815789473684211e-06, "loss": 3.3434, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.7528, "num_token_overlap": 11.66, "num_token_query": 31.3811, "num_token_union": 65.1218, "num_word_context": 202.2376, "num_word_doc": 49.8111, "num_word_query": 23.321, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22444.9315, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3086, "query_norm": 2.6031, "queue_k_norm": 2.6796, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3811, "sent_len_1": 66.7528, "sent_len_max_0": 127.995, "sent_len_max_1": 208.0637, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 177900 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3351, "doc_norm": 2.6794, "encoder_q-embeddings": 9644.2451, "encoder_q-layer.0": 6793.8101, "encoder_q-layer.1": 7343.939, "encoder_q-layer.10": 13585.4023, "encoder_q-layer.11": 34279.2109, "encoder_q-layer.2": 8160.6914, "encoder_q-layer.3": 8184.5571, "encoder_q-layer.4": 8896.7207, "encoder_q-layer.5": 9268.4375, "encoder_q-layer.6": 10286.5439, "encoder_q-layer.7": 11129.999, "encoder_q-layer.8": 13411.6064, "encoder_q-layer.9": 12276.8623, "epoch": 0.77, "inbatch_neg_score": 5.3089, "inbatch_pos_score": 5.9219, "learning_rate": 5.789473684210527e-06, "loss": 3.3351, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.7769, "num_token_overlap": 11.722, "num_token_query": 31.5185, "num_token_union": 65.2049, "num_word_context": 202.4553, "num_word_doc": 49.8375, "num_word_query": 23.4132, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21427.6793, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3086, "query_norm": 2.5993, "queue_k_norm": 2.68, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5185, "sent_len_1": 66.7769, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.75, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178000 }, { "accuracy": 50.293, "active_queue_size": 16384.0, "cl_loss": 3.3349, "doc_norm": 2.682, "encoder_q-embeddings": 9787.4551, "encoder_q-layer.0": 7043.0015, "encoder_q-layer.1": 7606.1094, "encoder_q-layer.10": 15392.1631, "encoder_q-layer.11": 35510.3086, "encoder_q-layer.2": 8680.2715, "encoder_q-layer.3": 8869.2559, "encoder_q-layer.4": 9618.0098, "encoder_q-layer.5": 10040.5508, "encoder_q-layer.6": 11496.958, "encoder_q-layer.7": 12910.623, "encoder_q-layer.8": 15363.0322, "encoder_q-layer.9": 14159.4326, "epoch": 0.77, "inbatch_neg_score": 5.3121, "inbatch_pos_score": 5.9219, "learning_rate": 5.7631578947368425e-06, "loss": 3.3349, "norm_diff": 0.0816, "norm_loss": 0.0, "num_token_doc": 66.7433, "num_token_overlap": 11.688, "num_token_query": 31.4196, "num_token_union": 65.0966, "num_word_context": 202.4071, "num_word_doc": 49.7965, "num_word_query": 23.3383, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22409.3624, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.3125, "query_norm": 2.6004, "queue_k_norm": 2.6804, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4196, "sent_len_1": 66.7433, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.5462, "stdk": 0.0497, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178100 }, { "accuracy": 53.8574, "active_queue_size": 16384.0, "cl_loss": 3.3439, "doc_norm": 2.6856, "encoder_q-embeddings": 4911.8096, "encoder_q-layer.0": 3393.7932, "encoder_q-layer.1": 3574.1504, "encoder_q-layer.10": 7176.6523, "encoder_q-layer.11": 17362.1348, "encoder_q-layer.2": 3933.0696, "encoder_q-layer.3": 4088.7939, "encoder_q-layer.4": 4489.5498, "encoder_q-layer.5": 4576.7363, "encoder_q-layer.6": 5018.1069, "encoder_q-layer.7": 5664.9062, "encoder_q-layer.8": 6816.2725, "encoder_q-layer.9": 6382.3076, "epoch": 0.77, "inbatch_neg_score": 5.3121, "inbatch_pos_score": 5.9414, "learning_rate": 5.7368421052631575e-06, "loss": 3.3439, "norm_diff": 0.0821, "norm_loss": 0.0, "num_token_doc": 66.7723, "num_token_overlap": 11.6898, "num_token_query": 31.3802, "num_token_union": 65.1322, "num_word_context": 202.32, "num_word_doc": 49.856, "num_word_query": 23.3137, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10831.4378, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3125, "query_norm": 2.6035, "queue_k_norm": 2.6806, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3802, "sent_len_1": 66.7723, "sent_len_max_0": 127.995, "sent_len_max_1": 206.5838, "stdk": 0.0499, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178200 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3413, "doc_norm": 2.6795, "encoder_q-embeddings": 5101.6602, "encoder_q-layer.0": 3509.2583, "encoder_q-layer.1": 3756.429, "encoder_q-layer.10": 6830.3657, "encoder_q-layer.11": 16919.2891, "encoder_q-layer.2": 4294.8813, "encoder_q-layer.3": 4616.5225, "encoder_q-layer.4": 5086.0283, "encoder_q-layer.5": 5241.187, "encoder_q-layer.6": 5658.4512, "encoder_q-layer.7": 6114.2012, "encoder_q-layer.8": 7398.8188, "encoder_q-layer.9": 6367.2212, "epoch": 0.77, "inbatch_neg_score": 5.314, "inbatch_pos_score": 5.9258, "learning_rate": 5.710526315789474e-06, "loss": 3.3413, "norm_diff": 0.0767, "norm_loss": 0.0, "num_token_doc": 66.6922, "num_token_overlap": 11.6653, "num_token_query": 31.3805, "num_token_union": 65.0773, "num_word_context": 202.186, "num_word_doc": 49.7471, "num_word_query": 23.3123, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11009.3165, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3125, "query_norm": 2.6028, "queue_k_norm": 2.6806, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3805, "sent_len_1": 66.6922, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.8212, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178300 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3422, "doc_norm": 2.6805, "encoder_q-embeddings": 4825.041, "encoder_q-layer.0": 3319.5032, "encoder_q-layer.1": 3512.6389, "encoder_q-layer.10": 6360.2202, "encoder_q-layer.11": 17231.5254, "encoder_q-layer.2": 3975.3577, "encoder_q-layer.3": 4088.0811, "encoder_q-layer.4": 4466.1123, "encoder_q-layer.5": 4374.394, "encoder_q-layer.6": 4927.9038, "encoder_q-layer.7": 5846.4893, "encoder_q-layer.8": 6910.9648, "encoder_q-layer.9": 6260.3901, "epoch": 0.77, "inbatch_neg_score": 5.3149, "inbatch_pos_score": 5.9258, "learning_rate": 5.68421052631579e-06, "loss": 3.3422, "norm_diff": 0.0783, "norm_loss": 0.0, "num_token_doc": 66.9416, "num_token_overlap": 11.6882, "num_token_query": 31.3704, "num_token_union": 65.1869, "num_word_context": 202.4183, "num_word_doc": 49.9354, "num_word_query": 23.291, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10698.3011, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3164, "query_norm": 2.6022, "queue_k_norm": 2.681, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3704, "sent_len_1": 66.9416, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.6975, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178400 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3281, "doc_norm": 2.6804, "encoder_q-embeddings": 4822.8042, "encoder_q-layer.0": 3349.1108, "encoder_q-layer.1": 3624.0054, "encoder_q-layer.10": 6811.1089, "encoder_q-layer.11": 16848.1582, "encoder_q-layer.2": 3999.4878, "encoder_q-layer.3": 4232.0967, "encoder_q-layer.4": 4598.6226, "encoder_q-layer.5": 4835.1494, "encoder_q-layer.6": 5143.6685, "encoder_q-layer.7": 5591.0166, "encoder_q-layer.8": 6378.394, "encoder_q-layer.9": 6287.377, "epoch": 0.77, "inbatch_neg_score": 5.3139, "inbatch_pos_score": 5.9336, "learning_rate": 5.657894736842106e-06, "loss": 3.3281, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.7746, "num_token_overlap": 11.6763, "num_token_query": 31.3599, "num_token_union": 65.1117, "num_word_context": 202.106, "num_word_doc": 49.7969, "num_word_query": 23.2904, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10558.0869, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3125, "query_norm": 2.6061, "queue_k_norm": 2.6814, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3599, "sent_len_1": 66.7746, "sent_len_max_0": 127.98, "sent_len_max_1": 210.1625, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178500 }, { "accuracy": 52.2949, "active_queue_size": 16384.0, "cl_loss": 3.3525, "doc_norm": 2.6809, "encoder_q-embeddings": 4837.7129, "encoder_q-layer.0": 3408.2437, "encoder_q-layer.1": 3851.2861, "encoder_q-layer.10": 6497.6855, "encoder_q-layer.11": 16765.627, "encoder_q-layer.2": 4282.0229, "encoder_q-layer.3": 4209.8965, "encoder_q-layer.4": 4405.9038, "encoder_q-layer.5": 4434.9058, "encoder_q-layer.6": 4854.1499, "encoder_q-layer.7": 5450.1641, "encoder_q-layer.8": 6760.7261, "encoder_q-layer.9": 6393.668, "epoch": 0.77, "inbatch_neg_score": 5.3136, "inbatch_pos_score": 5.9375, "learning_rate": 5.631578947368421e-06, "loss": 3.3525, "norm_diff": 0.0765, "norm_loss": 0.0, "num_token_doc": 66.794, "num_token_overlap": 11.6565, "num_token_query": 31.3602, "num_token_union": 65.1319, "num_word_context": 202.2333, "num_word_doc": 49.8304, "num_word_query": 23.2826, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10681.8245, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3125, "query_norm": 2.6044, "queue_k_norm": 2.6817, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3602, "sent_len_1": 66.794, "sent_len_max_0": 127.985, "sent_len_max_1": 208.215, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178600 }, { "accuracy": 53.6133, "active_queue_size": 16384.0, "cl_loss": 3.3487, "doc_norm": 2.6848, "encoder_q-embeddings": 4677.4458, "encoder_q-layer.0": 3289.95, "encoder_q-layer.1": 3568.021, "encoder_q-layer.10": 7070.9834, "encoder_q-layer.11": 17224.2012, "encoder_q-layer.2": 3975.7886, "encoder_q-layer.3": 4018.5818, "encoder_q-layer.4": 4283.6978, "encoder_q-layer.5": 4419.7441, "encoder_q-layer.6": 4770.605, "encoder_q-layer.7": 5634.3813, "encoder_q-layer.8": 6893.48, "encoder_q-layer.9": 6456.3208, "epoch": 0.78, "inbatch_neg_score": 5.3149, "inbatch_pos_score": 5.9375, "learning_rate": 5.6052631578947374e-06, "loss": 3.3487, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.6231, "num_token_overlap": 11.6664, "num_token_query": 31.408, "num_token_union": 65.0906, "num_word_context": 202.2887, "num_word_doc": 49.7074, "num_word_query": 23.3333, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10700.6168, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3164, "query_norm": 2.6069, "queue_k_norm": 2.6819, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.408, "sent_len_1": 66.6231, "sent_len_max_0": 127.9488, "sent_len_max_1": 208.0075, "stdk": 0.0498, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178700 }, { "accuracy": 53.1738, "active_queue_size": 16384.0, "cl_loss": 3.3391, "doc_norm": 2.6839, "encoder_q-embeddings": 4841.5356, "encoder_q-layer.0": 3348.6692, "encoder_q-layer.1": 3611.499, "encoder_q-layer.10": 6896.3584, "encoder_q-layer.11": 17693.8145, "encoder_q-layer.2": 4033.0559, "encoder_q-layer.3": 4101.0991, "encoder_q-layer.4": 4394.2129, "encoder_q-layer.5": 4569.6904, "encoder_q-layer.6": 5235.3071, "encoder_q-layer.7": 5954.6006, "encoder_q-layer.8": 7256.313, "encoder_q-layer.9": 6512.4404, "epoch": 0.78, "inbatch_neg_score": 5.3165, "inbatch_pos_score": 5.9492, "learning_rate": 5.578947368421053e-06, "loss": 3.3391, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.6718, "num_token_overlap": 11.6731, "num_token_query": 31.448, "num_token_union": 65.1092, "num_word_context": 202.4142, "num_word_doc": 49.7564, "num_word_query": 23.3659, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11139.2699, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3164, "query_norm": 2.6065, "queue_k_norm": 2.6823, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.448, "sent_len_1": 66.6718, "sent_len_max_0": 127.975, "sent_len_max_1": 208.6325, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178800 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3403, "doc_norm": 2.6841, "encoder_q-embeddings": 4681.9795, "encoder_q-layer.0": 3315.2417, "encoder_q-layer.1": 3568.8149, "encoder_q-layer.10": 7342.2681, "encoder_q-layer.11": 17965.9082, "encoder_q-layer.2": 3864.4028, "encoder_q-layer.3": 4016.0298, "encoder_q-layer.4": 4295.7939, "encoder_q-layer.5": 4408.3257, "encoder_q-layer.6": 5078.8921, "encoder_q-layer.7": 5788.5352, "encoder_q-layer.8": 7307.9946, "encoder_q-layer.9": 6878.0112, "epoch": 0.78, "inbatch_neg_score": 5.3163, "inbatch_pos_score": 5.9375, "learning_rate": 5.552631578947369e-06, "loss": 3.3403, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.7927, "num_token_overlap": 11.6547, "num_token_query": 31.3264, "num_token_union": 65.1402, "num_word_context": 202.5414, "num_word_doc": 49.8572, "num_word_query": 23.2741, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10816.6409, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3164, "query_norm": 2.6048, "queue_k_norm": 2.6826, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3264, "sent_len_1": 66.7927, "sent_len_max_0": 127.9737, "sent_len_max_1": 209.1725, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 178900 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3357, "doc_norm": 2.6831, "encoder_q-embeddings": 4776.1128, "encoder_q-layer.0": 3323.7944, "encoder_q-layer.1": 3528.7346, "encoder_q-layer.10": 6377.6934, "encoder_q-layer.11": 17252.5898, "encoder_q-layer.2": 3987.7993, "encoder_q-layer.3": 4093.1389, "encoder_q-layer.4": 4487.605, "encoder_q-layer.5": 4698.0703, "encoder_q-layer.6": 4937.6318, "encoder_q-layer.7": 5649.0117, "encoder_q-layer.8": 6742.4312, "encoder_q-layer.9": 6398.9819, "epoch": 0.78, "inbatch_neg_score": 5.3187, "inbatch_pos_score": 5.9297, "learning_rate": 5.526315789473684e-06, "loss": 3.3357, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.8913, "num_token_overlap": 11.7016, "num_token_query": 31.4786, "num_token_union": 65.227, "num_word_context": 202.6957, "num_word_doc": 49.9439, "num_word_query": 23.3998, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10691.7693, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3203, "query_norm": 2.6035, "queue_k_norm": 2.6831, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4786, "sent_len_1": 66.8913, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.5975, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179000 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3322, "doc_norm": 2.6795, "encoder_q-embeddings": 5637.731, "encoder_q-layer.0": 3867.2151, "encoder_q-layer.1": 4267.8477, "encoder_q-layer.10": 6691.188, "encoder_q-layer.11": 17305.0762, "encoder_q-layer.2": 4804.5576, "encoder_q-layer.3": 4983.0918, "encoder_q-layer.4": 5513.9023, "encoder_q-layer.5": 5473.3486, "encoder_q-layer.6": 6117.835, "encoder_q-layer.7": 6441.7812, "encoder_q-layer.8": 6989.8052, "encoder_q-layer.9": 6544.9854, "epoch": 0.78, "inbatch_neg_score": 5.3213, "inbatch_pos_score": 5.9336, "learning_rate": 5.500000000000001e-06, "loss": 3.3322, "norm_diff": 0.0748, "norm_loss": 0.0, "num_token_doc": 66.837, "num_token_overlap": 11.7127, "num_token_query": 31.4567, "num_token_union": 65.2342, "num_word_context": 202.1866, "num_word_doc": 49.8613, "num_word_query": 23.3683, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11282.9889, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3203, "query_norm": 2.6047, "queue_k_norm": 2.6828, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4567, "sent_len_1": 66.837, "sent_len_max_0": 127.9638, "sent_len_max_1": 207.1788, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179100 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.3384, "doc_norm": 2.6831, "encoder_q-embeddings": 4676.9419, "encoder_q-layer.0": 3291.7136, "encoder_q-layer.1": 3427.9568, "encoder_q-layer.10": 7143.7822, "encoder_q-layer.11": 17907.3555, "encoder_q-layer.2": 3851.6663, "encoder_q-layer.3": 3910.552, "encoder_q-layer.4": 4410.3442, "encoder_q-layer.5": 4575.6943, "encoder_q-layer.6": 4857.4697, "encoder_q-layer.7": 5913.8477, "encoder_q-layer.8": 7178.6367, "encoder_q-layer.9": 6429.625, "epoch": 0.78, "inbatch_neg_score": 5.3214, "inbatch_pos_score": 5.9414, "learning_rate": 5.4736842105263165e-06, "loss": 3.3384, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.844, "num_token_overlap": 11.6955, "num_token_query": 31.338, "num_token_union": 65.1144, "num_word_context": 202.121, "num_word_doc": 49.8631, "num_word_query": 23.2651, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10830.641, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3203, "query_norm": 2.6058, "queue_k_norm": 2.6832, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.338, "sent_len_1": 66.844, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.0813, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179200 }, { "accuracy": 52.9785, "active_queue_size": 16384.0, "cl_loss": 3.334, "doc_norm": 2.6859, "encoder_q-embeddings": 4978.52, "encoder_q-layer.0": 3541.2832, "encoder_q-layer.1": 3801.1592, "encoder_q-layer.10": 6435.1367, "encoder_q-layer.11": 16735.4512, "encoder_q-layer.2": 4269.0698, "encoder_q-layer.3": 4431.52, "encoder_q-layer.4": 4873.6133, "encoder_q-layer.5": 4776.3945, "encoder_q-layer.6": 5190.1948, "encoder_q-layer.7": 5562.9834, "encoder_q-layer.8": 6557.6284, "encoder_q-layer.9": 6341.9067, "epoch": 0.78, "inbatch_neg_score": 5.3242, "inbatch_pos_score": 5.9453, "learning_rate": 5.4473684210526315e-06, "loss": 3.334, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.8135, "num_token_overlap": 11.6793, "num_token_query": 31.3627, "num_token_union": 65.1487, "num_word_context": 202.5649, "num_word_doc": 49.8694, "num_word_query": 23.292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10585.7804, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3242, "query_norm": 2.6065, "queue_k_norm": 2.6848, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3627, "sent_len_1": 66.8135, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.7575, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 179300 }, { "accuracy": 53.2715, "active_queue_size": 16384.0, "cl_loss": 3.3261, "doc_norm": 2.684, "encoder_q-embeddings": 4532.7285, "encoder_q-layer.0": 3255.9141, "encoder_q-layer.1": 3423.6653, "encoder_q-layer.10": 6562.1162, "encoder_q-layer.11": 17807.1055, "encoder_q-layer.2": 3905.644, "encoder_q-layer.3": 4012.0359, "encoder_q-layer.4": 4355.0278, "encoder_q-layer.5": 4370.3242, "encoder_q-layer.6": 5350.3037, "encoder_q-layer.7": 5715.3618, "encoder_q-layer.8": 6804.2129, "encoder_q-layer.9": 6273.6162, "epoch": 0.78, "inbatch_neg_score": 5.3259, "inbatch_pos_score": 5.9453, "learning_rate": 5.421052631578947e-06, "loss": 3.3261, "norm_diff": 0.0772, "norm_loss": 0.0, "num_token_doc": 66.6589, "num_token_overlap": 11.6603, "num_token_query": 31.3603, "num_token_union": 65.0336, "num_word_context": 201.9107, "num_word_doc": 49.7353, "num_word_query": 23.2895, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10952.5259, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3281, "query_norm": 2.6068, "queue_k_norm": 2.6839, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3603, "sent_len_1": 66.6589, "sent_len_max_0": 127.9938, "sent_len_max_1": 209.4263, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179400 }, { "accuracy": 53.8086, "active_queue_size": 16384.0, "cl_loss": 3.3345, "doc_norm": 2.6856, "encoder_q-embeddings": 2224.0867, "encoder_q-layer.0": 1610.5996, "encoder_q-layer.1": 1704.3359, "encoder_q-layer.10": 3490.9192, "encoder_q-layer.11": 8578.8564, "encoder_q-layer.2": 1935.6042, "encoder_q-layer.3": 1985.276, "encoder_q-layer.4": 2184.5596, "encoder_q-layer.5": 2373.7903, "encoder_q-layer.6": 2611.2139, "encoder_q-layer.7": 2965.5574, "encoder_q-layer.8": 3364.6606, "encoder_q-layer.9": 3130.822, "epoch": 0.78, "inbatch_neg_score": 5.3263, "inbatch_pos_score": 5.9609, "learning_rate": 5.394736842105263e-06, "loss": 3.3345, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.8195, "num_token_overlap": 11.6975, "num_token_query": 31.4624, "num_token_union": 65.1633, "num_word_context": 202.1077, "num_word_doc": 49.8537, "num_word_query": 23.3724, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5351.5821, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3281, "query_norm": 2.6047, "queue_k_norm": 2.6846, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4624, "sent_len_1": 66.8195, "sent_len_max_0": 127.99, "sent_len_max_1": 211.2363, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179500 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3266, "doc_norm": 2.6868, "encoder_q-embeddings": 2278.8376, "encoder_q-layer.0": 1695.9875, "encoder_q-layer.1": 1797.297, "encoder_q-layer.10": 3476.9609, "encoder_q-layer.11": 8587.4512, "encoder_q-layer.2": 1976.2797, "encoder_q-layer.3": 2009.1436, "encoder_q-layer.4": 2204.7458, "encoder_q-layer.5": 2214.5405, "encoder_q-layer.6": 2524.1755, "encoder_q-layer.7": 2976.9775, "encoder_q-layer.8": 3403.4541, "encoder_q-layer.9": 3185.9792, "epoch": 0.78, "inbatch_neg_score": 5.3276, "inbatch_pos_score": 5.9531, "learning_rate": 5.36842105263158e-06, "loss": 3.3266, "norm_diff": 0.0778, "norm_loss": 0.0, "num_token_doc": 66.8569, "num_token_overlap": 11.7041, "num_token_query": 31.4581, "num_token_union": 65.1743, "num_word_context": 202.2451, "num_word_doc": 49.8686, "num_word_query": 23.3683, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5380.051, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3281, "query_norm": 2.609, "queue_k_norm": 2.6854, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4581, "sent_len_1": 66.8569, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.93, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 179600 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.323, "doc_norm": 2.6806, "encoder_q-embeddings": 2408.8811, "encoder_q-layer.0": 1655.3248, "encoder_q-layer.1": 1775.8344, "encoder_q-layer.10": 3331.1328, "encoder_q-layer.11": 8544.5254, "encoder_q-layer.2": 1996.1429, "encoder_q-layer.3": 1989.8074, "encoder_q-layer.4": 2175.3857, "encoder_q-layer.5": 2293.636, "encoder_q-layer.6": 2586.2678, "encoder_q-layer.7": 2883.9053, "encoder_q-layer.8": 3444.7422, "encoder_q-layer.9": 3325.3, "epoch": 0.78, "inbatch_neg_score": 5.3303, "inbatch_pos_score": 5.9375, "learning_rate": 5.342105263157895e-06, "loss": 3.323, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.7859, "num_token_overlap": 11.6934, "num_token_query": 31.426, "num_token_union": 65.1493, "num_word_context": 202.2328, "num_word_doc": 49.8549, "num_word_query": 23.3373, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5355.3866, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3281, "query_norm": 2.6065, "queue_k_norm": 2.6848, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.426, "sent_len_1": 66.7859, "sent_len_max_0": 127.9675, "sent_len_max_1": 209.1538, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179700 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3369, "doc_norm": 2.6844, "encoder_q-embeddings": 2356.3752, "encoder_q-layer.0": 1711.3662, "encoder_q-layer.1": 1875.1951, "encoder_q-layer.10": 3435.1436, "encoder_q-layer.11": 8421.6035, "encoder_q-layer.2": 2062.1172, "encoder_q-layer.3": 2083.6692, "encoder_q-layer.4": 2277.9429, "encoder_q-layer.5": 2453.6802, "encoder_q-layer.6": 2628.1567, "encoder_q-layer.7": 2853.4482, "encoder_q-layer.8": 3405.6506, "encoder_q-layer.9": 3118.7061, "epoch": 0.78, "inbatch_neg_score": 5.33, "inbatch_pos_score": 5.9492, "learning_rate": 5.315789473684211e-06, "loss": 3.3369, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.7492, "num_token_overlap": 11.7204, "num_token_query": 31.4246, "num_token_union": 65.0936, "num_word_context": 202.3187, "num_word_doc": 49.8155, "num_word_query": 23.3407, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5319.2662, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.332, "query_norm": 2.6063, "queue_k_norm": 2.6851, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4246, "sent_len_1": 66.7492, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.9512, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 179800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3332, "doc_norm": 2.6846, "encoder_q-embeddings": 2346.8323, "encoder_q-layer.0": 1630.8079, "encoder_q-layer.1": 1757.9998, "encoder_q-layer.10": 3343.8506, "encoder_q-layer.11": 8505.6006, "encoder_q-layer.2": 2038.3506, "encoder_q-layer.3": 2088.46, "encoder_q-layer.4": 2222.2915, "encoder_q-layer.5": 2394.7563, "encoder_q-layer.6": 2679.9844, "encoder_q-layer.7": 2885.6145, "encoder_q-layer.8": 3545.8147, "encoder_q-layer.9": 3277.959, "epoch": 0.78, "inbatch_neg_score": 5.3343, "inbatch_pos_score": 5.9531, "learning_rate": 5.289473684210526e-06, "loss": 3.3332, "norm_diff": 0.0769, "norm_loss": 0.0, "num_token_doc": 66.6815, "num_token_overlap": 11.6835, "num_token_query": 31.4439, "num_token_union": 65.1005, "num_word_context": 202.355, "num_word_doc": 49.7701, "num_word_query": 23.3516, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5372.0254, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3359, "query_norm": 2.6077, "queue_k_norm": 2.6861, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4439, "sent_len_1": 66.6815, "sent_len_max_0": 127.9725, "sent_len_max_1": 208.1937, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 179900 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3601, "doc_norm": 2.6854, "encoder_q-embeddings": 2406.7009, "encoder_q-layer.0": 1744.4213, "encoder_q-layer.1": 1861.3756, "encoder_q-layer.10": 3703.7754, "encoder_q-layer.11": 9028.2031, "encoder_q-layer.2": 2115.1956, "encoder_q-layer.3": 2167.5596, "encoder_q-layer.4": 2419.7188, "encoder_q-layer.5": 2486.5188, "encoder_q-layer.6": 2846.5669, "encoder_q-layer.7": 3049.4509, "encoder_q-layer.8": 3552.6667, "encoder_q-layer.9": 3343.1304, "epoch": 0.78, "inbatch_neg_score": 5.3347, "inbatch_pos_score": 5.9453, "learning_rate": 5.263157894736842e-06, "loss": 3.3601, "norm_diff": 0.08, "norm_loss": 0.0, "num_token_doc": 66.6143, "num_token_overlap": 11.6322, "num_token_query": 31.3197, "num_token_union": 65.0052, "num_word_context": 202.2239, "num_word_doc": 49.7036, "num_word_query": 23.2678, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5564.8645, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3359, "query_norm": 2.6055, "queue_k_norm": 2.6848, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3197, "sent_len_1": 66.6143, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.3475, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 180000 }, { "dev_runtime": 28.4781, "dev_samples_per_second": 1.124, "dev_steps_per_second": 0.035, "epoch": 0.78, "step": 180000, "test_accuracy": 94.384765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3201791048049927, "test_doc_norm": 2.672759532928467, "test_inbatch_neg_score": 5.627645492553711, "test_inbatch_pos_score": 6.655579566955566, "test_loss": 0.3201791048049927, "test_loss_align": 0.9945827722549438, "test_loss_unif": -50.874786376953125, "test_loss_unif_q@queue": -50.874786376953125, "test_norm_diff": 0.0030812472105026245, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.330533981323242, "test_query_norm": 2.674452543258667, "test_queue_k_norm": 2.6848952770233154, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04436445236206055, "test_stdq": 0.04391000419855118, "test_stdqueue_k": 0.04954461380839348, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.4781, "dev_samples_per_second": 1.124, "dev_steps_per_second": 0.035, "epoch": 0.78, "eval_beir-arguana_ndcg@10": 0.40719, "eval_beir-arguana_recall@10": 0.69061, "eval_beir-arguana_recall@100": 0.94523, "eval_beir-arguana_recall@20": 0.81437, "eval_beir-avg_ndcg@10": 0.40124325, "eval_beir-avg_recall@10": 0.47440625000000003, "eval_beir-avg_recall@100": 0.6532188333333334, "eval_beir-avg_recall@20": 0.5383650833333333, "eval_beir-cqadupstack_ndcg@10": 0.2941725, "eval_beir-cqadupstack_recall@10": 0.3960425, "eval_beir-cqadupstack_recall@100": 0.6339583333333335, "eval_beir-cqadupstack_recall@20": 0.46739083333333337, "eval_beir-fiqa_ndcg@10": 0.27363, "eval_beir-fiqa_recall@10": 0.34224, "eval_beir-fiqa_recall@100": 0.6112, "eval_beir-fiqa_recall@20": 0.4171, "eval_beir-nfcorpus_ndcg@10": 0.32679, "eval_beir-nfcorpus_recall@10": 0.15969, "eval_beir-nfcorpus_recall@100": 0.31433, "eval_beir-nfcorpus_recall@20": 0.20464, "eval_beir-nq_ndcg@10": 0.32732, "eval_beir-nq_recall@10": 0.5197, "eval_beir-nq_recall@100": 0.84198, "eval_beir-nq_recall@20": 0.63772, "eval_beir-quora_ndcg@10": 0.7966, "eval_beir-quora_recall@10": 0.89985, "eval_beir-quora_recall@100": 0.98207, "eval_beir-quora_recall@20": 0.93916, "eval_beir-scidocs_ndcg@10": 0.16454, "eval_beir-scidocs_recall@10": 0.17262, "eval_beir-scidocs_recall@100": 0.38945, "eval_beir-scidocs_recall@20": 0.23377, "eval_beir-scifact_ndcg@10": 0.67774, "eval_beir-scifact_recall@10": 0.809, "eval_beir-scifact_recall@100": 0.92989, "eval_beir-scifact_recall@20": 0.84811, "eval_beir-trec-covid_ndcg@10": 0.57372, "eval_beir-trec-covid_recall@10": 0.626, "eval_beir-trec-covid_recall@100": 0.4566, "eval_beir-trec-covid_recall@20": 0.605, "eval_beir-webis-touche2020_ndcg@10": 0.17073, "eval_beir-webis-touche2020_recall@10": 0.12831, "eval_beir-webis-touche2020_recall@100": 0.42748, "eval_beir-webis-touche2020_recall@20": 0.21639, "eval_senteval-avg_sts": 0.7340177366426724, "eval_senteval-sickr_spearman": 0.7247049300558932, "eval_senteval-stsb_spearman": 0.7433305432294517, "step": 180000, "test_accuracy": 94.384765625, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3201791048049927, "test_doc_norm": 2.672759532928467, "test_inbatch_neg_score": 5.627645492553711, "test_inbatch_pos_score": 6.655579566955566, "test_loss": 0.3201791048049927, "test_loss_align": 0.9945827722549438, "test_loss_unif": -50.874786376953125, "test_loss_unif_q@queue": -50.874786376953125, "test_norm_diff": 0.0030812472105026245, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.330533981323242, "test_query_norm": 2.674452543258667, "test_queue_k_norm": 2.6848952770233154, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04436445236206055, "test_stdq": 0.04391000419855118, "test_stdqueue_k": 0.04954461380839348, "test_stdqueue_q": 0.0 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3541, "doc_norm": 2.6851, "encoder_q-embeddings": 2529.4219, "encoder_q-layer.0": 1804.5048, "encoder_q-layer.1": 1947.327, "encoder_q-layer.10": 3410.8669, "encoder_q-layer.11": 8835.6357, "encoder_q-layer.2": 2196.3252, "encoder_q-layer.3": 2259.0403, "encoder_q-layer.4": 2467.0735, "encoder_q-layer.5": 2635.2122, "encoder_q-layer.6": 2978.3401, "encoder_q-layer.7": 3226.5544, "encoder_q-layer.8": 3647.4783, "encoder_q-layer.9": 3311.8015, "epoch": 0.78, "inbatch_neg_score": 5.34, "inbatch_pos_score": 5.9414, "learning_rate": 5.236842105263158e-06, "loss": 3.3541, "norm_diff": 0.0756, "norm_loss": 0.0, "num_token_doc": 66.8533, "num_token_overlap": 11.6502, "num_token_query": 31.3016, "num_token_union": 65.1562, "num_word_context": 202.562, "num_word_doc": 49.8769, "num_word_query": 23.2476, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5647.4891, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3398, "query_norm": 2.6096, "queue_k_norm": 2.6852, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3016, "sent_len_1": 66.8533, "sent_len_max_0": 127.9813, "sent_len_max_1": 208.5725, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 180100 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3402, "doc_norm": 2.6864, "encoder_q-embeddings": 2379.5403, "encoder_q-layer.0": 1648.3422, "encoder_q-layer.1": 1766.1116, "encoder_q-layer.10": 4037.0642, "encoder_q-layer.11": 8670.9873, "encoder_q-layer.2": 1971.4406, "encoder_q-layer.3": 1993.0681, "encoder_q-layer.4": 2155.0271, "encoder_q-layer.5": 2242.8552, "encoder_q-layer.6": 2560.4988, "encoder_q-layer.7": 2720.5918, "encoder_q-layer.8": 3467.7712, "encoder_q-layer.9": 3211.0991, "epoch": 0.78, "inbatch_neg_score": 5.342, "inbatch_pos_score": 5.9531, "learning_rate": 5.210526315789474e-06, "loss": 3.3402, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7532, "num_token_overlap": 11.6918, "num_token_query": 31.4077, "num_token_union": 65.1083, "num_word_context": 202.4255, "num_word_doc": 49.7993, "num_word_query": 23.3381, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5336.8884, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3398, "query_norm": 2.6103, "queue_k_norm": 2.6864, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4077, "sent_len_1": 66.7532, "sent_len_max_0": 127.9338, "sent_len_max_1": 209.7337, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 180200 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3518, "doc_norm": 2.6864, "encoder_q-embeddings": 2610.2537, "encoder_q-layer.0": 1833.1647, "encoder_q-layer.1": 2028.9666, "encoder_q-layer.10": 3762.2515, "encoder_q-layer.11": 8604.0996, "encoder_q-layer.2": 2190.0754, "encoder_q-layer.3": 2265.6753, "encoder_q-layer.4": 2447.7739, "encoder_q-layer.5": 2514.1194, "encoder_q-layer.6": 2717.9668, "encoder_q-layer.7": 3164.3896, "encoder_q-layer.8": 3436.0547, "encoder_q-layer.9": 3169.0916, "epoch": 0.78, "inbatch_neg_score": 5.3439, "inbatch_pos_score": 5.957, "learning_rate": 5.18421052631579e-06, "loss": 3.3518, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.9141, "num_token_overlap": 11.6691, "num_token_query": 31.3312, "num_token_union": 65.1634, "num_word_context": 202.4759, "num_word_doc": 49.9081, "num_word_query": 23.2649, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5502.1855, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3438, "query_norm": 2.6077, "queue_k_norm": 2.6854, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3312, "sent_len_1": 66.9141, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.4863, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 180300 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3364, "doc_norm": 2.6899, "encoder_q-embeddings": 2357.6907, "encoder_q-layer.0": 1708.282, "encoder_q-layer.1": 1783.15, "encoder_q-layer.10": 3601.5195, "encoder_q-layer.11": 9006.5918, "encoder_q-layer.2": 1980.2319, "encoder_q-layer.3": 2030.0968, "encoder_q-layer.4": 2243.8472, "encoder_q-layer.5": 2356.3569, "encoder_q-layer.6": 2664.6274, "encoder_q-layer.7": 3019.603, "encoder_q-layer.8": 3475.9189, "encoder_q-layer.9": 3380.4324, "epoch": 0.78, "inbatch_neg_score": 5.3427, "inbatch_pos_score": 5.9688, "learning_rate": 5.1578947368421055e-06, "loss": 3.3364, "norm_diff": 0.075, "norm_loss": 0.0, "num_token_doc": 66.8143, "num_token_overlap": 11.7196, "num_token_query": 31.436, "num_token_union": 65.125, "num_word_context": 202.1883, "num_word_doc": 49.84, "num_word_query": 23.3478, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5543.3972, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3438, "query_norm": 2.6149, "queue_k_norm": 2.6874, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.436, "sent_len_1": 66.8143, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.1912, "stdk": 0.0498, "stdq": 0.0436, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 180400 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3522, "doc_norm": 2.6838, "encoder_q-embeddings": 2224.7009, "encoder_q-layer.0": 1627.1556, "encoder_q-layer.1": 1735.9351, "encoder_q-layer.10": 3542.0847, "encoder_q-layer.11": 8468.0273, "encoder_q-layer.2": 1942.8934, "encoder_q-layer.3": 2037.4559, "encoder_q-layer.4": 2190.3572, "encoder_q-layer.5": 2223.1648, "encoder_q-layer.6": 2514.2993, "encoder_q-layer.7": 3039.5269, "encoder_q-layer.8": 3355.0671, "encoder_q-layer.9": 3185.8459, "epoch": 0.78, "inbatch_neg_score": 5.3446, "inbatch_pos_score": 5.9609, "learning_rate": 5.131578947368421e-06, "loss": 3.3522, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.7848, "num_token_overlap": 11.6415, "num_token_query": 31.2856, "num_token_union": 65.1051, "num_word_context": 202.3383, "num_word_doc": 49.8409, "num_word_query": 23.2421, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5271.6558, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3477, "query_norm": 2.6095, "queue_k_norm": 2.6865, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2856, "sent_len_1": 66.7848, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.1438, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 180500 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.6869, "encoder_q-embeddings": 2356.0464, "encoder_q-layer.0": 1659.4312, "encoder_q-layer.1": 1753.4438, "encoder_q-layer.10": 3282.6021, "encoder_q-layer.11": 8703.8252, "encoder_q-layer.2": 1948.3511, "encoder_q-layer.3": 2005.7174, "encoder_q-layer.4": 2166.7808, "encoder_q-layer.5": 2217.3989, "encoder_q-layer.6": 2482.3281, "encoder_q-layer.7": 2757.5376, "encoder_q-layer.8": 3357.49, "encoder_q-layer.9": 3157.4478, "epoch": 0.78, "inbatch_neg_score": 5.3487, "inbatch_pos_score": 5.9609, "learning_rate": 5.105263157894737e-06, "loss": 3.3331, "norm_diff": 0.0742, "norm_loss": 0.0, "num_token_doc": 66.7286, "num_token_overlap": 11.7087, "num_token_query": 31.4331, "num_token_union": 65.1074, "num_word_context": 202.1726, "num_word_doc": 49.759, "num_word_query": 23.3357, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5355.9441, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3516, "query_norm": 2.6127, "queue_k_norm": 2.6885, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4331, "sent_len_1": 66.7286, "sent_len_max_0": 127.99, "sent_len_max_1": 209.845, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 180600 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3483, "doc_norm": 2.6889, "encoder_q-embeddings": 2556.9092, "encoder_q-layer.0": 1795.9814, "encoder_q-layer.1": 1921.5422, "encoder_q-layer.10": 3466.7722, "encoder_q-layer.11": 8552.8457, "encoder_q-layer.2": 2142.1865, "encoder_q-layer.3": 2266.1389, "encoder_q-layer.4": 2452.3438, "encoder_q-layer.5": 2471.1936, "encoder_q-layer.6": 2732.9038, "encoder_q-layer.7": 2980.8481, "encoder_q-layer.8": 3347.3765, "encoder_q-layer.9": 3116.6973, "epoch": 0.78, "inbatch_neg_score": 5.3531, "inbatch_pos_score": 5.957, "learning_rate": 5.078947368421053e-06, "loss": 3.3483, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.8376, "num_token_overlap": 11.6833, "num_token_query": 31.4639, "num_token_union": 65.2421, "num_word_context": 202.6054, "num_word_doc": 49.9002, "num_word_query": 23.3745, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5461.8239, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3516, "query_norm": 2.6094, "queue_k_norm": 2.6879, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4639, "sent_len_1": 66.8376, "sent_len_max_0": 127.9925, "sent_len_max_1": 207.7275, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 180700 }, { "accuracy": 52.7344, "active_queue_size": 16384.0, "cl_loss": 3.3389, "doc_norm": 2.6898, "encoder_q-embeddings": 2401.0002, "encoder_q-layer.0": 1694.2386, "encoder_q-layer.1": 1855.3308, "encoder_q-layer.10": 3618.8665, "encoder_q-layer.11": 8956.6006, "encoder_q-layer.2": 2142.7495, "encoder_q-layer.3": 2154.4536, "encoder_q-layer.4": 2383.5098, "encoder_q-layer.5": 2404.3892, "encoder_q-layer.6": 2674.9629, "encoder_q-layer.7": 3127.4727, "encoder_q-layer.8": 3559.2161, "encoder_q-layer.9": 3344.2649, "epoch": 0.78, "inbatch_neg_score": 5.3559, "inbatch_pos_score": 5.9648, "learning_rate": 5.052631578947369e-06, "loss": 3.3389, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.7725, "num_token_overlap": 11.6274, "num_token_query": 31.1946, "num_token_union": 65.0349, "num_word_context": 202.0854, "num_word_doc": 49.8317, "num_word_query": 23.1735, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5606.6677, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3555, "query_norm": 2.6115, "queue_k_norm": 2.687, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.1946, "sent_len_1": 66.7725, "sent_len_max_0": 127.99, "sent_len_max_1": 211.2, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 180800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3351, "doc_norm": 2.6897, "encoder_q-embeddings": 2946.6643, "encoder_q-layer.0": 2146.5125, "encoder_q-layer.1": 2445.1794, "encoder_q-layer.10": 3413.03, "encoder_q-layer.11": 8634.7344, "encoder_q-layer.2": 2850.5586, "encoder_q-layer.3": 2986.1487, "encoder_q-layer.4": 3372.1323, "encoder_q-layer.5": 3471.6875, "encoder_q-layer.6": 3906.9917, "encoder_q-layer.7": 4380.2031, "encoder_q-layer.8": 4731.127, "encoder_q-layer.9": 3848.2886, "epoch": 0.78, "inbatch_neg_score": 5.3594, "inbatch_pos_score": 5.9766, "learning_rate": 5.026315789473685e-06, "loss": 3.3351, "norm_diff": 0.0723, "norm_loss": 0.0, "num_token_doc": 66.6066, "num_token_overlap": 11.6648, "num_token_query": 31.4039, "num_token_union": 65.0635, "num_word_context": 202.3953, "num_word_doc": 49.7018, "num_word_query": 23.3227, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6322.647, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3594, "query_norm": 2.6174, "queue_k_norm": 2.6886, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4039, "sent_len_1": 66.6066, "sent_len_max_0": 127.9963, "sent_len_max_1": 208.7788, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 180900 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3336, "doc_norm": 2.6864, "encoder_q-embeddings": 2567.6123, "encoder_q-layer.0": 1808.2974, "encoder_q-layer.1": 1971.7477, "encoder_q-layer.10": 3840.1377, "encoder_q-layer.11": 8868.3115, "encoder_q-layer.2": 2143.5815, "encoder_q-layer.3": 2213.8184, "encoder_q-layer.4": 2407.6045, "encoder_q-layer.5": 2383.5623, "encoder_q-layer.6": 2612.0273, "encoder_q-layer.7": 2914.2339, "encoder_q-layer.8": 3537.9946, "encoder_q-layer.9": 3248.8572, "epoch": 0.79, "inbatch_neg_score": 5.3609, "inbatch_pos_score": 5.9688, "learning_rate": 5e-06, "loss": 3.3336, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.6573, "num_token_overlap": 11.6478, "num_token_query": 31.3471, "num_token_union": 65.0845, "num_word_context": 202.2016, "num_word_doc": 49.7576, "num_word_query": 23.2761, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5517.3895, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3594, "query_norm": 2.6137, "queue_k_norm": 2.6888, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3471, "sent_len_1": 66.6573, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.4338, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181000 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3334, "doc_norm": 2.6906, "encoder_q-embeddings": 2672.9385, "encoder_q-layer.0": 1863.5442, "encoder_q-layer.1": 2046.3186, "encoder_q-layer.10": 3413.2849, "encoder_q-layer.11": 8660.6934, "encoder_q-layer.2": 2393.4897, "encoder_q-layer.3": 2459.8491, "encoder_q-layer.4": 2691.7139, "encoder_q-layer.5": 2805.8491, "encoder_q-layer.6": 2958.0598, "encoder_q-layer.7": 3267.9653, "encoder_q-layer.8": 3610.6245, "encoder_q-layer.9": 3282.0586, "epoch": 0.79, "inbatch_neg_score": 5.3616, "inbatch_pos_score": 5.9805, "learning_rate": 4.973684210526316e-06, "loss": 3.3334, "norm_diff": 0.0707, "norm_loss": 0.0, "num_token_doc": 66.5986, "num_token_overlap": 11.6504, "num_token_query": 31.3285, "num_token_union": 65.0173, "num_word_context": 202.4549, "num_word_doc": 49.6732, "num_word_query": 23.2499, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5685.1131, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3594, "query_norm": 2.62, "queue_k_norm": 2.689, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3285, "sent_len_1": 66.5986, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.3, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181100 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3326, "doc_norm": 2.6902, "encoder_q-embeddings": 2428.6604, "encoder_q-layer.0": 1679.8029, "encoder_q-layer.1": 1803.9257, "encoder_q-layer.10": 3528.4976, "encoder_q-layer.11": 8965.3838, "encoder_q-layer.2": 2016.2255, "encoder_q-layer.3": 2060.2871, "encoder_q-layer.4": 2186.9558, "encoder_q-layer.5": 2203.761, "encoder_q-layer.6": 2495.0681, "encoder_q-layer.7": 2864.8108, "encoder_q-layer.8": 3443.1467, "encoder_q-layer.9": 3296.0605, "epoch": 0.79, "inbatch_neg_score": 5.3658, "inbatch_pos_score": 5.9844, "learning_rate": 4.947368421052632e-06, "loss": 3.3326, "norm_diff": 0.0747, "norm_loss": 0.0, "num_token_doc": 66.6644, "num_token_overlap": 11.6741, "num_token_query": 31.3604, "num_token_union": 65.0193, "num_word_context": 202.1478, "num_word_doc": 49.7441, "num_word_query": 23.2951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5542.7171, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3672, "query_norm": 2.6155, "queue_k_norm": 2.6897, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3604, "sent_len_1": 66.6644, "sent_len_max_0": 127.99, "sent_len_max_1": 208.83, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181200 }, { "accuracy": 53.7109, "active_queue_size": 16384.0, "cl_loss": 3.3346, "doc_norm": 2.6881, "encoder_q-embeddings": 2282.2815, "encoder_q-layer.0": 1707.7507, "encoder_q-layer.1": 1851.6094, "encoder_q-layer.10": 3472.7454, "encoder_q-layer.11": 8291.1855, "encoder_q-layer.2": 2026.4681, "encoder_q-layer.3": 2101.0942, "encoder_q-layer.4": 2320.2705, "encoder_q-layer.5": 2285.6594, "encoder_q-layer.6": 2507.7749, "encoder_q-layer.7": 2786.4966, "encoder_q-layer.8": 3385.9817, "encoder_q-layer.9": 3177.0193, "epoch": 0.79, "inbatch_neg_score": 5.3649, "inbatch_pos_score": 6.0, "learning_rate": 4.921052631578948e-06, "loss": 3.3346, "norm_diff": 0.0726, "norm_loss": 0.0, "num_token_doc": 66.7647, "num_token_overlap": 11.7065, "num_token_query": 31.4729, "num_token_union": 65.1735, "num_word_context": 202.3862, "num_word_doc": 49.8207, "num_word_query": 23.3887, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5216.3983, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.3672, "query_norm": 2.6155, "queue_k_norm": 2.69, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4729, "sent_len_1": 66.7647, "sent_len_max_0": 127.9675, "sent_len_max_1": 210.4988, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181300 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3344, "doc_norm": 2.6899, "encoder_q-embeddings": 2300.5566, "encoder_q-layer.0": 1618.3533, "encoder_q-layer.1": 1763.9669, "encoder_q-layer.10": 3763.1145, "encoder_q-layer.11": 9023.8154, "encoder_q-layer.2": 1975.7208, "encoder_q-layer.3": 1990.4539, "encoder_q-layer.4": 2113.324, "encoder_q-layer.5": 2213.9243, "encoder_q-layer.6": 2520.8149, "encoder_q-layer.7": 2892.4734, "encoder_q-layer.8": 3340.8533, "encoder_q-layer.9": 3225.4934, "epoch": 0.79, "inbatch_neg_score": 5.3673, "inbatch_pos_score": 5.9805, "learning_rate": 4.894736842105263e-06, "loss": 3.3344, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.755, "num_token_overlap": 11.6931, "num_token_query": 31.385, "num_token_union": 65.0997, "num_word_context": 202.1844, "num_word_doc": 49.801, "num_word_query": 23.3114, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5483.4312, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3672, "query_norm": 2.6128, "queue_k_norm": 2.69, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.385, "sent_len_1": 66.755, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.155, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181400 }, { "accuracy": 48.9746, "active_queue_size": 16384.0, "cl_loss": 3.3446, "doc_norm": 2.6887, "encoder_q-embeddings": 6225.2568, "encoder_q-layer.0": 4628.0239, "encoder_q-layer.1": 5315.0186, "encoder_q-layer.10": 6725.6274, "encoder_q-layer.11": 17489.877, "encoder_q-layer.2": 6248.1514, "encoder_q-layer.3": 6532.3862, "encoder_q-layer.4": 6847.874, "encoder_q-layer.5": 6697.7871, "encoder_q-layer.6": 7094.2832, "encoder_q-layer.7": 7378.6294, "encoder_q-layer.8": 7895.8525, "encoder_q-layer.9": 6648.0767, "epoch": 0.79, "inbatch_neg_score": 5.3724, "inbatch_pos_score": 5.9688, "learning_rate": 4.868421052631579e-06, "loss": 3.3446, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.6941, "num_token_overlap": 11.6309, "num_token_query": 31.2275, "num_token_union": 65.0084, "num_word_context": 202.1247, "num_word_doc": 49.7568, "num_word_query": 23.1783, "postclip_grad_norm": 1.0, "preclip_grad_norm": 12260.5738, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3711, "query_norm": 2.6112, "queue_k_norm": 2.6913, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2275, "sent_len_1": 66.6941, "sent_len_max_0": 127.99, "sent_len_max_1": 209.33, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181500 }, { "accuracy": 53.5645, "active_queue_size": 16384.0, "cl_loss": 3.3242, "doc_norm": 2.6921, "encoder_q-embeddings": 4530.146, "encoder_q-layer.0": 3240.5769, "encoder_q-layer.1": 3445.2649, "encoder_q-layer.10": 6843.436, "encoder_q-layer.11": 16748.3594, "encoder_q-layer.2": 3845.8091, "encoder_q-layer.3": 4085.0933, "encoder_q-layer.4": 4422.2563, "encoder_q-layer.5": 4317.4297, "encoder_q-layer.6": 5027.5615, "encoder_q-layer.7": 5733.5791, "encoder_q-layer.8": 6813.9282, "encoder_q-layer.9": 6336.8276, "epoch": 0.79, "inbatch_neg_score": 5.3736, "inbatch_pos_score": 6.0117, "learning_rate": 4.842105263157895e-06, "loss": 3.3242, "norm_diff": 0.0692, "norm_loss": 0.0, "num_token_doc": 66.8824, "num_token_overlap": 11.6391, "num_token_query": 31.332, "num_token_union": 65.1963, "num_word_context": 202.3399, "num_word_doc": 49.9014, "num_word_query": 23.2727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10494.1955, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.375, "query_norm": 2.623, "queue_k_norm": 2.6916, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.332, "sent_len_1": 66.8824, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.1775, "stdk": 0.0496, "stdq": 0.0437, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181600 }, { "accuracy": 53.7598, "active_queue_size": 16384.0, "cl_loss": 3.3355, "doc_norm": 2.6947, "encoder_q-embeddings": 4694.0815, "encoder_q-layer.0": 3281.0642, "encoder_q-layer.1": 3520.5571, "encoder_q-layer.10": 6828.6152, "encoder_q-layer.11": 17518.0137, "encoder_q-layer.2": 3987.7336, "encoder_q-layer.3": 4081.3167, "encoder_q-layer.4": 4385.5186, "encoder_q-layer.5": 4528.0894, "encoder_q-layer.6": 5008.5132, "encoder_q-layer.7": 5917.1382, "encoder_q-layer.8": 6817.5044, "encoder_q-layer.9": 6543.2754, "epoch": 0.79, "inbatch_neg_score": 5.3747, "inbatch_pos_score": 6.0078, "learning_rate": 4.815789473684211e-06, "loss": 3.3355, "norm_diff": 0.0754, "norm_loss": 0.0, "num_token_doc": 66.7254, "num_token_overlap": 11.6653, "num_token_query": 31.2875, "num_token_union": 65.0341, "num_word_context": 201.9535, "num_word_doc": 49.7915, "num_word_query": 23.2419, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10820.4189, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.375, "query_norm": 2.6193, "queue_k_norm": 2.691, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2875, "sent_len_1": 66.7254, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.9775, "stdk": 0.0498, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 181700 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.3322, "doc_norm": 2.6893, "encoder_q-embeddings": 4865.4209, "encoder_q-layer.0": 3418.5869, "encoder_q-layer.1": 3583.1719, "encoder_q-layer.10": 7955.0967, "encoder_q-layer.11": 18640.8535, "encoder_q-layer.2": 4106.104, "encoder_q-layer.3": 4270.1465, "encoder_q-layer.4": 4533.6519, "encoder_q-layer.5": 4843.6191, "encoder_q-layer.6": 5343.4189, "encoder_q-layer.7": 5791.438, "encoder_q-layer.8": 7452.1104, "encoder_q-layer.9": 7260.3491, "epoch": 0.79, "inbatch_neg_score": 5.3788, "inbatch_pos_score": 5.9922, "learning_rate": 4.789473684210526e-06, "loss": 3.3322, "norm_diff": 0.0706, "norm_loss": 0.0, "num_token_doc": 66.7298, "num_token_overlap": 11.6671, "num_token_query": 31.3463, "num_token_union": 65.0793, "num_word_context": 202.1314, "num_word_doc": 49.7781, "num_word_query": 23.273, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11206.928, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3789, "query_norm": 2.6187, "queue_k_norm": 2.6916, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3463, "sent_len_1": 66.7298, "sent_len_max_0": 127.99, "sent_len_max_1": 208.93, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181800 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3421, "doc_norm": 2.6924, "encoder_q-embeddings": 4535.9746, "encoder_q-layer.0": 3214.3352, "encoder_q-layer.1": 3515.1721, "encoder_q-layer.10": 6654.249, "encoder_q-layer.11": 16887.3379, "encoder_q-layer.2": 3882.5684, "encoder_q-layer.3": 4014.4578, "encoder_q-layer.4": 4202.6978, "encoder_q-layer.5": 4272.2383, "encoder_q-layer.6": 4731.1201, "encoder_q-layer.7": 5479.8242, "encoder_q-layer.8": 6358.6909, "encoder_q-layer.9": 6006.479, "epoch": 0.79, "inbatch_neg_score": 5.3793, "inbatch_pos_score": 6.0, "learning_rate": 4.763157894736842e-06, "loss": 3.3421, "norm_diff": 0.0762, "norm_loss": 0.0, "num_token_doc": 66.7492, "num_token_overlap": 11.6622, "num_token_query": 31.3739, "num_token_union": 65.1202, "num_word_context": 202.273, "num_word_doc": 49.7884, "num_word_query": 23.3033, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10472.6307, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3789, "query_norm": 2.6162, "queue_k_norm": 2.6932, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3739, "sent_len_1": 66.7492, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.1425, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 181900 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3378, "doc_norm": 2.6945, "encoder_q-embeddings": 4741.7988, "encoder_q-layer.0": 3281.3694, "encoder_q-layer.1": 3513.1982, "encoder_q-layer.10": 6730.2324, "encoder_q-layer.11": 17123.4062, "encoder_q-layer.2": 3889.3047, "encoder_q-layer.3": 3968.5735, "encoder_q-layer.4": 4096.9268, "encoder_q-layer.5": 4215.6938, "encoder_q-layer.6": 4709.2676, "encoder_q-layer.7": 5542.8354, "encoder_q-layer.8": 7050.8511, "encoder_q-layer.9": 6550.4854, "epoch": 0.79, "inbatch_neg_score": 5.3805, "inbatch_pos_score": 5.9922, "learning_rate": 4.736842105263159e-06, "loss": 3.3378, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.9654, "num_token_overlap": 11.6718, "num_token_query": 31.26, "num_token_union": 65.1331, "num_word_context": 202.352, "num_word_doc": 49.9649, "num_word_query": 23.1949, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10631.4796, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3789, "query_norm": 2.6185, "queue_k_norm": 2.6916, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.26, "sent_len_1": 66.9654, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.1937, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 182000 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.3282, "doc_norm": 2.6933, "encoder_q-embeddings": 4876.6123, "encoder_q-layer.0": 3361.7485, "encoder_q-layer.1": 3567.5696, "encoder_q-layer.10": 6965.7178, "encoder_q-layer.11": 17370.502, "encoder_q-layer.2": 3931.1392, "encoder_q-layer.3": 3980.8127, "encoder_q-layer.4": 4236.6665, "encoder_q-layer.5": 4451.0068, "encoder_q-layer.6": 4943.7515, "encoder_q-layer.7": 5783.7764, "encoder_q-layer.8": 6854.8438, "encoder_q-layer.9": 6199.061, "epoch": 0.79, "inbatch_neg_score": 5.384, "inbatch_pos_score": 5.9844, "learning_rate": 4.7105263157894736e-06, "loss": 3.3282, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.7792, "num_token_overlap": 11.7062, "num_token_query": 31.4128, "num_token_union": 65.1327, "num_word_context": 202.2357, "num_word_doc": 49.8048, "num_word_query": 23.3354, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10849.9777, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3828, "query_norm": 2.6173, "queue_k_norm": 2.6923, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4128, "sent_len_1": 66.7792, "sent_len_max_0": 127.98, "sent_len_max_1": 211.31, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 182100 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3414, "doc_norm": 2.6911, "encoder_q-embeddings": 4537.0547, "encoder_q-layer.0": 3357.4932, "encoder_q-layer.1": 3650.9905, "encoder_q-layer.10": 6548.2798, "encoder_q-layer.11": 17115.9473, "encoder_q-layer.2": 4152.6021, "encoder_q-layer.3": 4129.5234, "encoder_q-layer.4": 4611.0762, "encoder_q-layer.5": 5045.3516, "encoder_q-layer.6": 5422.3281, "encoder_q-layer.7": 6056.9644, "encoder_q-layer.8": 6905.0781, "encoder_q-layer.9": 6212.6138, "epoch": 0.79, "inbatch_neg_score": 5.3847, "inbatch_pos_score": 5.9922, "learning_rate": 4.684210526315789e-06, "loss": 3.3414, "norm_diff": 0.0766, "norm_loss": 0.0, "num_token_doc": 66.8708, "num_token_overlap": 11.7163, "num_token_query": 31.399, "num_token_union": 65.1325, "num_word_context": 202.3107, "num_word_doc": 49.8868, "num_word_query": 23.3244, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10650.0248, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3828, "query_norm": 2.6145, "queue_k_norm": 2.6939, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.399, "sent_len_1": 66.8708, "sent_len_max_0": 127.955, "sent_len_max_1": 209.65, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182200 }, { "accuracy": 49.4629, "active_queue_size": 16384.0, "cl_loss": 3.3609, "doc_norm": 2.6951, "encoder_q-embeddings": 5308.4463, "encoder_q-layer.0": 3774.0103, "encoder_q-layer.1": 4266.04, "encoder_q-layer.10": 6797.8521, "encoder_q-layer.11": 17524.2148, "encoder_q-layer.2": 4963.3672, "encoder_q-layer.3": 5245.3599, "encoder_q-layer.4": 5982.4883, "encoder_q-layer.5": 5730.8125, "encoder_q-layer.6": 6266.7529, "encoder_q-layer.7": 6740.5288, "encoder_q-layer.8": 7342.4062, "encoder_q-layer.9": 6498.9312, "epoch": 0.79, "inbatch_neg_score": 5.3888, "inbatch_pos_score": 5.9805, "learning_rate": 4.657894736842105e-06, "loss": 3.3609, "norm_diff": 0.0828, "norm_loss": 0.0, "num_token_doc": 66.7506, "num_token_overlap": 11.6548, "num_token_query": 31.2588, "num_token_union": 65.0376, "num_word_context": 202.3543, "num_word_doc": 49.8054, "num_word_query": 23.2179, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11613.9551, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3867, "query_norm": 2.6123, "queue_k_norm": 2.6932, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2588, "sent_len_1": 66.7506, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.0375, "stdk": 0.0496, "stdq": 0.0427, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 182300 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3276, "doc_norm": 2.6945, "encoder_q-embeddings": 4535.9043, "encoder_q-layer.0": 3240.6394, "encoder_q-layer.1": 3444.3101, "encoder_q-layer.10": 6777.646, "encoder_q-layer.11": 17239.3711, "encoder_q-layer.2": 3877.0955, "encoder_q-layer.3": 4005.6953, "encoder_q-layer.4": 4301.9482, "encoder_q-layer.5": 4565.3701, "encoder_q-layer.6": 5017.168, "encoder_q-layer.7": 5702.4048, "encoder_q-layer.8": 6913.8066, "encoder_q-layer.9": 6674.4077, "epoch": 0.79, "inbatch_neg_score": 5.3863, "inbatch_pos_score": 6.0, "learning_rate": 4.631578947368422e-06, "loss": 3.3276, "norm_diff": 0.0783, "norm_loss": 0.0, "num_token_doc": 66.6739, "num_token_overlap": 11.7042, "num_token_query": 31.4355, "num_token_union": 65.0543, "num_word_context": 202.1929, "num_word_doc": 49.7351, "num_word_query": 23.3521, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10746.2298, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3867, "query_norm": 2.6162, "queue_k_norm": 2.6942, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4355, "sent_len_1": 66.6739, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.52, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182400 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3398, "doc_norm": 2.6921, "encoder_q-embeddings": 4700.8477, "encoder_q-layer.0": 3319.4443, "encoder_q-layer.1": 3490.1306, "encoder_q-layer.10": 6806.2964, "encoder_q-layer.11": 16895.9707, "encoder_q-layer.2": 3899.8499, "encoder_q-layer.3": 3997.1565, "encoder_q-layer.4": 4281.2227, "encoder_q-layer.5": 4369.3521, "encoder_q-layer.6": 4869.8989, "encoder_q-layer.7": 5726.9014, "encoder_q-layer.8": 6858.1914, "encoder_q-layer.9": 6405.0171, "epoch": 0.79, "inbatch_neg_score": 5.39, "inbatch_pos_score": 5.9961, "learning_rate": 4.605263157894737e-06, "loss": 3.3398, "norm_diff": 0.0728, "norm_loss": 0.0, "num_token_doc": 66.7465, "num_token_overlap": 11.6926, "num_token_query": 31.3732, "num_token_union": 65.1159, "num_word_context": 202.3293, "num_word_doc": 49.78, "num_word_query": 23.2964, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10352.8574, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3906, "query_norm": 2.6193, "queue_k_norm": 2.6954, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3732, "sent_len_1": 66.7465, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.5263, "stdk": 0.0493, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182500 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3344, "doc_norm": 2.6956, "encoder_q-embeddings": 5021.3735, "encoder_q-layer.0": 3502.8213, "encoder_q-layer.1": 3824.7737, "encoder_q-layer.10": 7482.314, "encoder_q-layer.11": 18387.9531, "encoder_q-layer.2": 4200.1006, "encoder_q-layer.3": 4194.2759, "encoder_q-layer.4": 4546.9688, "encoder_q-layer.5": 4503.4209, "encoder_q-layer.6": 5058.0293, "encoder_q-layer.7": 5860.2598, "encoder_q-layer.8": 6933.4712, "encoder_q-layer.9": 6736.2578, "epoch": 0.79, "inbatch_neg_score": 5.391, "inbatch_pos_score": 6.0039, "learning_rate": 4.578947368421053e-06, "loss": 3.3344, "norm_diff": 0.0764, "norm_loss": 0.0, "num_token_doc": 66.6956, "num_token_overlap": 11.6401, "num_token_query": 31.284, "num_token_union": 65.0855, "num_word_context": 202.2857, "num_word_doc": 49.7508, "num_word_query": 23.2247, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11430.2812, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3906, "query_norm": 2.6191, "queue_k_norm": 2.6952, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.284, "sent_len_1": 66.6956, "sent_len_max_0": 127.99, "sent_len_max_1": 207.2212, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182600 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3354, "doc_norm": 2.6923, "encoder_q-embeddings": 4774.5029, "encoder_q-layer.0": 3385.2913, "encoder_q-layer.1": 3625.5864, "encoder_q-layer.10": 6710.1426, "encoder_q-layer.11": 17836.9004, "encoder_q-layer.2": 4033.2625, "encoder_q-layer.3": 4072.2036, "encoder_q-layer.4": 4409.5547, "encoder_q-layer.5": 4574.5542, "encoder_q-layer.6": 5180.105, "encoder_q-layer.7": 5719.7803, "encoder_q-layer.8": 7002.6963, "encoder_q-layer.9": 6351.189, "epoch": 0.79, "inbatch_neg_score": 5.3927, "inbatch_pos_score": 6.0078, "learning_rate": 4.5526315789473685e-06, "loss": 3.3354, "norm_diff": 0.075, "norm_loss": 0.0, "num_token_doc": 66.6959, "num_token_overlap": 11.6833, "num_token_query": 31.3932, "num_token_union": 65.0595, "num_word_context": 202.3532, "num_word_doc": 49.7827, "num_word_query": 23.3167, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11122.4409, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3945, "query_norm": 2.6173, "queue_k_norm": 2.695, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3932, "sent_len_1": 66.6959, "sent_len_max_0": 127.955, "sent_len_max_1": 207.8063, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182700 }, { "accuracy": 49.1699, "active_queue_size": 16384.0, "cl_loss": 3.3456, "doc_norm": 2.6939, "encoder_q-embeddings": 4841.561, "encoder_q-layer.0": 3477.532, "encoder_q-layer.1": 3750.9087, "encoder_q-layer.10": 7832.8174, "encoder_q-layer.11": 18097.9141, "encoder_q-layer.2": 4074.2002, "encoder_q-layer.3": 4095.5032, "encoder_q-layer.4": 4375.2358, "encoder_q-layer.5": 4660.8799, "encoder_q-layer.6": 5274.5894, "encoder_q-layer.7": 6063.7256, "encoder_q-layer.8": 7184.9561, "encoder_q-layer.9": 7224.8901, "epoch": 0.79, "inbatch_neg_score": 5.3948, "inbatch_pos_score": 5.9883, "learning_rate": 4.526315789473685e-06, "loss": 3.3456, "norm_diff": 0.08, "norm_loss": 0.0, "num_token_doc": 66.706, "num_token_overlap": 11.6572, "num_token_query": 31.3176, "num_token_union": 65.0471, "num_word_context": 202.323, "num_word_doc": 49.8102, "num_word_query": 23.2622, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11058.4612, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3945, "query_norm": 2.6139, "queue_k_norm": 2.6953, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3176, "sent_len_1": 66.706, "sent_len_max_0": 127.985, "sent_len_max_1": 208.6975, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 182800 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.3293, "doc_norm": 2.6938, "encoder_q-embeddings": 4927.895, "encoder_q-layer.0": 3456.9492, "encoder_q-layer.1": 3699.6174, "encoder_q-layer.10": 8347.5078, "encoder_q-layer.11": 18665.8047, "encoder_q-layer.2": 4108.3794, "encoder_q-layer.3": 4190.0879, "encoder_q-layer.4": 4493.8408, "encoder_q-layer.5": 4883.9341, "encoder_q-layer.6": 5371.9966, "encoder_q-layer.7": 6206.541, "encoder_q-layer.8": 7804.0425, "encoder_q-layer.9": 7007.8564, "epoch": 0.79, "inbatch_neg_score": 5.3978, "inbatch_pos_score": 5.9961, "learning_rate": 4.5e-06, "loss": 3.3293, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.9581, "num_token_overlap": 11.7096, "num_token_query": 31.3693, "num_token_union": 65.1913, "num_word_context": 202.5057, "num_word_doc": 49.9814, "num_word_query": 23.301, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11586.3448, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3945, "query_norm": 2.6158, "queue_k_norm": 2.6953, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3693, "sent_len_1": 66.9581, "sent_len_max_0": 128.0, "sent_len_max_1": 208.0788, "stdk": 0.0493, "stdq": 0.0429, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 182900 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3339, "doc_norm": 2.6948, "encoder_q-embeddings": 4881.7275, "encoder_q-layer.0": 3324.4353, "encoder_q-layer.1": 3595.0249, "encoder_q-layer.10": 6990.2241, "encoder_q-layer.11": 18110.3164, "encoder_q-layer.2": 4006.488, "encoder_q-layer.3": 4104.7476, "encoder_q-layer.4": 4361.0513, "encoder_q-layer.5": 4557.6738, "encoder_q-layer.6": 5416.8193, "encoder_q-layer.7": 6087.2651, "encoder_q-layer.8": 6927.5698, "encoder_q-layer.9": 6633.9014, "epoch": 0.79, "inbatch_neg_score": 5.3992, "inbatch_pos_score": 6.0039, "learning_rate": 4.473684210526316e-06, "loss": 3.3339, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.7111, "num_token_overlap": 11.6652, "num_token_query": 31.4295, "num_token_union": 65.1324, "num_word_context": 202.5364, "num_word_doc": 49.8009, "num_word_query": 23.3547, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11330.6703, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3984, "query_norm": 2.6164, "queue_k_norm": 2.6962, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4295, "sent_len_1": 66.7111, "sent_len_max_0": 127.9988, "sent_len_max_1": 207.5813, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183000 }, { "accuracy": 49.707, "active_queue_size": 16384.0, "cl_loss": 3.3366, "doc_norm": 2.6964, "encoder_q-embeddings": 4949.7319, "encoder_q-layer.0": 3444.8137, "encoder_q-layer.1": 3616.8057, "encoder_q-layer.10": 7406.4619, "encoder_q-layer.11": 18008.4336, "encoder_q-layer.2": 4100.0649, "encoder_q-layer.3": 4309.9097, "encoder_q-layer.4": 4631.9697, "encoder_q-layer.5": 4830.5127, "encoder_q-layer.6": 5451.5928, "encoder_q-layer.7": 5955.4097, "encoder_q-layer.8": 6940.1602, "encoder_q-layer.9": 6484.6973, "epoch": 0.79, "inbatch_neg_score": 5.3981, "inbatch_pos_score": 6.0, "learning_rate": 4.447368421052632e-06, "loss": 3.3366, "norm_diff": 0.0773, "norm_loss": 0.0, "num_token_doc": 66.8226, "num_token_overlap": 11.699, "num_token_query": 31.4391, "num_token_union": 65.1743, "num_word_context": 202.2386, "num_word_doc": 49.8451, "num_word_query": 23.3536, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11261.9473, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3984, "query_norm": 2.6191, "queue_k_norm": 2.6978, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4391, "sent_len_1": 66.8226, "sent_len_max_0": 127.975, "sent_len_max_1": 209.8537, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 183100 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3445, "doc_norm": 2.6957, "encoder_q-embeddings": 5135.0239, "encoder_q-layer.0": 3599.79, "encoder_q-layer.1": 3748.2529, "encoder_q-layer.10": 7734.9287, "encoder_q-layer.11": 18478.9688, "encoder_q-layer.2": 4158.9639, "encoder_q-layer.3": 4203.3091, "encoder_q-layer.4": 4526.1919, "encoder_q-layer.5": 4687.2905, "encoder_q-layer.6": 5435.2803, "encoder_q-layer.7": 5976.8193, "encoder_q-layer.8": 7126.5552, "encoder_q-layer.9": 6736.9644, "epoch": 0.79, "inbatch_neg_score": 5.4003, "inbatch_pos_score": 6.0117, "learning_rate": 4.4210526315789476e-06, "loss": 3.3445, "norm_diff": 0.0722, "norm_loss": 0.0, "num_token_doc": 66.8636, "num_token_overlap": 11.6808, "num_token_query": 31.4172, "num_token_union": 65.2203, "num_word_context": 202.5045, "num_word_doc": 49.8927, "num_word_query": 23.3317, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11356.2144, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.3984, "query_norm": 2.6236, "queue_k_norm": 2.6973, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4172, "sent_len_1": 66.8636, "sent_len_max_0": 127.98, "sent_len_max_1": 208.6987, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183200 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.327, "doc_norm": 2.6967, "encoder_q-embeddings": 2392.2451, "encoder_q-layer.0": 1671.7555, "encoder_q-layer.1": 1864.1342, "encoder_q-layer.10": 3557.5989, "encoder_q-layer.11": 9379.0869, "encoder_q-layer.2": 1993.3119, "encoder_q-layer.3": 2024.1439, "encoder_q-layer.4": 2152.4575, "encoder_q-layer.5": 2247.6992, "encoder_q-layer.6": 2582.7217, "encoder_q-layer.7": 2915.2737, "encoder_q-layer.8": 3551.4768, "encoder_q-layer.9": 3375.1714, "epoch": 0.8, "inbatch_neg_score": 5.4022, "inbatch_pos_score": 6.0078, "learning_rate": 4.394736842105263e-06, "loss": 3.327, "norm_diff": 0.0749, "norm_loss": 0.0, "num_token_doc": 66.7874, "num_token_overlap": 11.6511, "num_token_query": 31.2723, "num_token_union": 65.1076, "num_word_context": 202.1877, "num_word_doc": 49.808, "num_word_query": 23.2092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5710.8045, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4023, "query_norm": 2.6219, "queue_k_norm": 2.6976, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2723, "sent_len_1": 66.7874, "sent_len_max_0": 127.975, "sent_len_max_1": 209.96, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183300 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.341, "doc_norm": 2.6974, "encoder_q-embeddings": 2451.5793, "encoder_q-layer.0": 1660.9462, "encoder_q-layer.1": 1797.1194, "encoder_q-layer.10": 3467.4431, "encoder_q-layer.11": 8797.8145, "encoder_q-layer.2": 1978.9081, "encoder_q-layer.3": 2101.1228, "encoder_q-layer.4": 2246.6829, "encoder_q-layer.5": 2431.8865, "encoder_q-layer.6": 2655.439, "encoder_q-layer.7": 2858.2295, "encoder_q-layer.8": 3387.438, "encoder_q-layer.9": 3180.7888, "epoch": 0.8, "inbatch_neg_score": 5.4032, "inbatch_pos_score": 6.0039, "learning_rate": 4.368421052631579e-06, "loss": 3.341, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.7939, "num_token_overlap": 11.6782, "num_token_query": 31.3369, "num_token_union": 65.1364, "num_word_context": 202.2313, "num_word_doc": 49.8485, "num_word_query": 23.2711, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5425.7217, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4023, "query_norm": 2.6165, "queue_k_norm": 2.6988, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3369, "sent_len_1": 66.7939, "sent_len_max_0": 128.0, "sent_len_max_1": 207.16, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 183400 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3363, "doc_norm": 2.697, "encoder_q-embeddings": 2440.5264, "encoder_q-layer.0": 1686.3319, "encoder_q-layer.1": 1830.8904, "encoder_q-layer.10": 3325.4558, "encoder_q-layer.11": 8299.2383, "encoder_q-layer.2": 1995.4004, "encoder_q-layer.3": 2134.218, "encoder_q-layer.4": 2267.4131, "encoder_q-layer.5": 2432.7922, "encoder_q-layer.6": 2642.5, "encoder_q-layer.7": 3077.9172, "encoder_q-layer.8": 3441.9382, "encoder_q-layer.9": 3026.2878, "epoch": 0.8, "inbatch_neg_score": 5.4055, "inbatch_pos_score": 6.0195, "learning_rate": 4.342105263157895e-06, "loss": 3.3363, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.8591, "num_token_overlap": 11.697, "num_token_query": 31.4031, "num_token_union": 65.1383, "num_word_context": 202.2573, "num_word_doc": 49.8661, "num_word_query": 23.3292, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5272.8305, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4062, "query_norm": 2.6199, "queue_k_norm": 2.6982, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4031, "sent_len_1": 66.8591, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.695, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183500 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3372, "doc_norm": 2.6995, "encoder_q-embeddings": 2259.1831, "encoder_q-layer.0": 1645.3732, "encoder_q-layer.1": 1778.219, "encoder_q-layer.10": 3509.1912, "encoder_q-layer.11": 8760.6387, "encoder_q-layer.2": 1946.349, "encoder_q-layer.3": 1971.2643, "encoder_q-layer.4": 2083.7122, "encoder_q-layer.5": 2166.2859, "encoder_q-layer.6": 2412.2893, "encoder_q-layer.7": 2843.1406, "encoder_q-layer.8": 3428.1558, "encoder_q-layer.9": 3069.4414, "epoch": 0.8, "inbatch_neg_score": 5.4065, "inbatch_pos_score": 6.0312, "learning_rate": 4.315789473684211e-06, "loss": 3.3372, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 66.6597, "num_token_overlap": 11.6751, "num_token_query": 31.4064, "num_token_union": 65.0889, "num_word_context": 202.1233, "num_word_doc": 49.7471, "num_word_query": 23.3141, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5396.3799, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4062, "query_norm": 2.6224, "queue_k_norm": 2.6988, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4064, "sent_len_1": 66.6597, "sent_len_max_0": 127.975, "sent_len_max_1": 207.365, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183600 }, { "accuracy": 50.5371, "active_queue_size": 16384.0, "cl_loss": 3.3483, "doc_norm": 2.6993, "encoder_q-embeddings": 2431.2256, "encoder_q-layer.0": 1768.7133, "encoder_q-layer.1": 1867.7358, "encoder_q-layer.10": 3715.4539, "encoder_q-layer.11": 9203.585, "encoder_q-layer.2": 2099.8533, "encoder_q-layer.3": 2203.6541, "encoder_q-layer.4": 2342.4609, "encoder_q-layer.5": 2448.5369, "encoder_q-layer.6": 2738.165, "encoder_q-layer.7": 3124.207, "encoder_q-layer.8": 3705.2268, "encoder_q-layer.9": 3362.6646, "epoch": 0.8, "inbatch_neg_score": 5.4084, "inbatch_pos_score": 6.0156, "learning_rate": 4.289473684210527e-06, "loss": 3.3483, "norm_diff": 0.076, "norm_loss": 0.0, "num_token_doc": 66.6933, "num_token_overlap": 11.6395, "num_token_query": 31.2857, "num_token_union": 65.0603, "num_word_context": 202.0302, "num_word_doc": 49.7666, "num_word_query": 23.232, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5764.2226, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4062, "query_norm": 2.6233, "queue_k_norm": 2.6985, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2857, "sent_len_1": 66.6933, "sent_len_max_0": 127.9475, "sent_len_max_1": 209.6188, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183700 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3436, "doc_norm": 2.6992, "encoder_q-embeddings": 2363.9546, "encoder_q-layer.0": 1704.6021, "encoder_q-layer.1": 1901.6473, "encoder_q-layer.10": 3446.6987, "encoder_q-layer.11": 8933.5078, "encoder_q-layer.2": 2166.7656, "encoder_q-layer.3": 2225.4932, "encoder_q-layer.4": 2301.4043, "encoder_q-layer.5": 2405.083, "encoder_q-layer.6": 2626.9248, "encoder_q-layer.7": 2921.1575, "encoder_q-layer.8": 3384.7944, "encoder_q-layer.9": 3187.5159, "epoch": 0.8, "inbatch_neg_score": 5.4061, "inbatch_pos_score": 6.0195, "learning_rate": 4.2631578947368425e-06, "loss": 3.3436, "norm_diff": 0.0777, "norm_loss": 0.0, "num_token_doc": 66.7481, "num_token_overlap": 11.6791, "num_token_query": 31.3569, "num_token_union": 65.1392, "num_word_context": 202.3636, "num_word_doc": 49.8067, "num_word_query": 23.2772, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5541.6205, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4062, "query_norm": 2.6215, "queue_k_norm": 2.6988, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3569, "sent_len_1": 66.7481, "sent_len_max_0": 127.9663, "sent_len_max_1": 209.4913, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 183800 }, { "accuracy": 53.6133, "active_queue_size": 16384.0, "cl_loss": 3.3384, "doc_norm": 2.7027, "encoder_q-embeddings": 2358.844, "encoder_q-layer.0": 1638.0425, "encoder_q-layer.1": 1732.6738, "encoder_q-layer.10": 3247.0256, "encoder_q-layer.11": 8562.9121, "encoder_q-layer.2": 1928.6277, "encoder_q-layer.3": 1988.9343, "encoder_q-layer.4": 2185.1001, "encoder_q-layer.5": 2271.5361, "encoder_q-layer.6": 2699.064, "encoder_q-layer.7": 2993.8086, "encoder_q-layer.8": 3507.751, "encoder_q-layer.9": 3260.5642, "epoch": 0.8, "inbatch_neg_score": 5.4113, "inbatch_pos_score": 6.0391, "learning_rate": 4.2368421052631575e-06, "loss": 3.3384, "norm_diff": 0.077, "norm_loss": 0.0, "num_token_doc": 66.8071, "num_token_overlap": 11.6888, "num_token_query": 31.3852, "num_token_union": 65.1313, "num_word_context": 202.3843, "num_word_doc": 49.8589, "num_word_query": 23.3079, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5348.5545, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4102, "query_norm": 2.6257, "queue_k_norm": 2.7006, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3852, "sent_len_1": 66.8071, "sent_len_max_0": 127.9938, "sent_len_max_1": 208.0137, "stdk": 0.0498, "stdq": 0.0435, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 183900 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3347, "doc_norm": 2.701, "encoder_q-embeddings": 4695.2568, "encoder_q-layer.0": 3677.8286, "encoder_q-layer.1": 3912.6135, "encoder_q-layer.10": 3599.9607, "encoder_q-layer.11": 9011.709, "encoder_q-layer.2": 4198.2104, "encoder_q-layer.3": 3890.1458, "encoder_q-layer.4": 3966.9778, "encoder_q-layer.5": 3891.9819, "encoder_q-layer.6": 3825.5989, "encoder_q-layer.7": 3885.196, "encoder_q-layer.8": 4153.9863, "encoder_q-layer.9": 3467.0444, "epoch": 0.8, "inbatch_neg_score": 5.4127, "inbatch_pos_score": 6.0234, "learning_rate": 4.210526315789474e-06, "loss": 3.3347, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.6773, "num_token_overlap": 11.6748, "num_token_query": 31.3648, "num_token_union": 65.0797, "num_word_context": 202.3546, "num_word_doc": 49.8185, "num_word_query": 23.3039, "postclip_grad_norm": 1.0, "preclip_grad_norm": 7159.3616, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4141, "query_norm": 2.6233, "queue_k_norm": 2.7005, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3648, "sent_len_1": 66.6773, "sent_len_max_0": 127.98, "sent_len_max_1": 205.7688, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184000 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3423, "doc_norm": 2.7016, "encoder_q-embeddings": 2351.3701, "encoder_q-layer.0": 1666.7325, "encoder_q-layer.1": 1780.0935, "encoder_q-layer.10": 3733.5747, "encoder_q-layer.11": 8945.5957, "encoder_q-layer.2": 1972.2217, "encoder_q-layer.3": 2049.6558, "encoder_q-layer.4": 2135.5483, "encoder_q-layer.5": 2251.3875, "encoder_q-layer.6": 2520.8186, "encoder_q-layer.7": 3037.9204, "encoder_q-layer.8": 3474.9897, "encoder_q-layer.9": 3234.7776, "epoch": 0.8, "inbatch_neg_score": 5.4127, "inbatch_pos_score": 6.0312, "learning_rate": 4.18421052631579e-06, "loss": 3.3423, "norm_diff": 0.0747, "norm_loss": 0.0, "num_token_doc": 66.7316, "num_token_overlap": 11.7062, "num_token_query": 31.5351, "num_token_union": 65.2056, "num_word_context": 202.2824, "num_word_doc": 49.7838, "num_word_query": 23.4226, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5478.1411, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4141, "query_norm": 2.6269, "queue_k_norm": 2.701, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5351, "sent_len_1": 66.7316, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.7262, "stdk": 0.0497, "stdq": 0.0435, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 184100 }, { "accuracy": 49.5117, "active_queue_size": 16384.0, "cl_loss": 3.3506, "doc_norm": 2.7, "encoder_q-embeddings": 2500.1028, "encoder_q-layer.0": 1721.2832, "encoder_q-layer.1": 1859.1913, "encoder_q-layer.10": 4068.2354, "encoder_q-layer.11": 9563.5684, "encoder_q-layer.2": 2084.2029, "encoder_q-layer.3": 2175.4146, "encoder_q-layer.4": 2350.3301, "encoder_q-layer.5": 2426.3533, "encoder_q-layer.6": 2729.1401, "encoder_q-layer.7": 3079.6719, "encoder_q-layer.8": 3787.1396, "encoder_q-layer.9": 3511.7942, "epoch": 0.8, "inbatch_neg_score": 5.4163, "inbatch_pos_score": 6.0195, "learning_rate": 4.157894736842106e-06, "loss": 3.3506, "norm_diff": 0.0803, "norm_loss": 0.0, "num_token_doc": 66.7102, "num_token_overlap": 11.6593, "num_token_query": 31.3342, "num_token_union": 65.0575, "num_word_context": 202.1758, "num_word_doc": 49.7931, "num_word_query": 23.2578, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5930.0487, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.418, "query_norm": 2.6197, "queue_k_norm": 2.7004, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3342, "sent_len_1": 66.7102, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.2425, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184200 }, { "accuracy": 50.3906, "active_queue_size": 16384.0, "cl_loss": 3.3469, "doc_norm": 2.7001, "encoder_q-embeddings": 2340.7175, "encoder_q-layer.0": 1656.814, "encoder_q-layer.1": 1745.5446, "encoder_q-layer.10": 3657.1855, "encoder_q-layer.11": 8879.6836, "encoder_q-layer.2": 1985.5583, "encoder_q-layer.3": 1980.4136, "encoder_q-layer.4": 2168.3062, "encoder_q-layer.5": 2218.1562, "encoder_q-layer.6": 2529.2761, "encoder_q-layer.7": 2965.293, "encoder_q-layer.8": 3444.751, "encoder_q-layer.9": 3222.1587, "epoch": 0.8, "inbatch_neg_score": 5.4181, "inbatch_pos_score": 6.0234, "learning_rate": 4.131578947368421e-06, "loss": 3.3469, "norm_diff": 0.0789, "norm_loss": 0.0, "num_token_doc": 66.655, "num_token_overlap": 11.6389, "num_token_query": 31.3061, "num_token_union": 65.0618, "num_word_context": 202.2579, "num_word_doc": 49.7541, "num_word_query": 23.248, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5461.1978, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.418, "query_norm": 2.6212, "queue_k_norm": 2.7008, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3061, "sent_len_1": 66.655, "sent_len_max_0": 127.985, "sent_len_max_1": 207.9175, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184300 }, { "accuracy": 50.9277, "active_queue_size": 16384.0, "cl_loss": 3.3311, "doc_norm": 2.7021, "encoder_q-embeddings": 2343.8169, "encoder_q-layer.0": 1667.3232, "encoder_q-layer.1": 1771.9148, "encoder_q-layer.10": 3465.8757, "encoder_q-layer.11": 8968.6504, "encoder_q-layer.2": 1955.0636, "encoder_q-layer.3": 1960.7156, "encoder_q-layer.4": 2077.1968, "encoder_q-layer.5": 2177.1587, "encoder_q-layer.6": 2497.9822, "encoder_q-layer.7": 2825.3706, "encoder_q-layer.8": 3293.9492, "encoder_q-layer.9": 3189.7197, "epoch": 0.8, "inbatch_neg_score": 5.4191, "inbatch_pos_score": 6.0391, "learning_rate": 4.105263157894737e-06, "loss": 3.3311, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.6931, "num_token_overlap": 11.6615, "num_token_query": 31.3355, "num_token_union": 65.084, "num_word_context": 202.1797, "num_word_doc": 49.7526, "num_word_query": 23.2621, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5488.6087, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.418, "query_norm": 2.6241, "queue_k_norm": 2.701, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3355, "sent_len_1": 66.6931, "sent_len_max_0": 127.9975, "sent_len_max_1": 206.0575, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184400 }, { "accuracy": 49.4141, "active_queue_size": 16384.0, "cl_loss": 3.3371, "doc_norm": 2.6992, "encoder_q-embeddings": 2559.3069, "encoder_q-layer.0": 1838.0884, "encoder_q-layer.1": 1993.5178, "encoder_q-layer.10": 3420.573, "encoder_q-layer.11": 8957.8564, "encoder_q-layer.2": 2227.8115, "encoder_q-layer.3": 2281.7808, "encoder_q-layer.4": 2416.0874, "encoder_q-layer.5": 2576.2141, "encoder_q-layer.6": 2780.6001, "encoder_q-layer.7": 2982.5364, "encoder_q-layer.8": 3663.1323, "encoder_q-layer.9": 3333.9229, "epoch": 0.8, "inbatch_neg_score": 5.4199, "inbatch_pos_score": 6.0234, "learning_rate": 4.078947368421053e-06, "loss": 3.3371, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.8899, "num_token_overlap": 11.6785, "num_token_query": 31.335, "num_token_union": 65.1437, "num_word_context": 202.0193, "num_word_doc": 49.8794, "num_word_query": 23.2665, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5653.3742, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4219, "query_norm": 2.6217, "queue_k_norm": 2.7012, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.335, "sent_len_1": 66.8899, "sent_len_max_0": 127.97, "sent_len_max_1": 210.0675, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184500 }, { "accuracy": 53.1738, "active_queue_size": 16384.0, "cl_loss": 3.3497, "doc_norm": 2.7013, "encoder_q-embeddings": 2229.2651, "encoder_q-layer.0": 1582.1562, "encoder_q-layer.1": 1710.006, "encoder_q-layer.10": 3387.3669, "encoder_q-layer.11": 8749.7617, "encoder_q-layer.2": 1916.3916, "encoder_q-layer.3": 1951.0996, "encoder_q-layer.4": 2120.9004, "encoder_q-layer.5": 2181.8442, "encoder_q-layer.6": 2366.449, "encoder_q-layer.7": 2740.9111, "encoder_q-layer.8": 3342.9631, "encoder_q-layer.9": 3110.9119, "epoch": 0.8, "inbatch_neg_score": 5.4207, "inbatch_pos_score": 6.0469, "learning_rate": 4.052631578947368e-06, "loss": 3.3497, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.607, "num_token_overlap": 11.6808, "num_token_query": 31.4164, "num_token_union": 65.0372, "num_word_context": 202.1304, "num_word_doc": 49.6934, "num_word_query": 23.3324, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5315.5931, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4219, "query_norm": 2.624, "queue_k_norm": 2.7014, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4164, "sent_len_1": 66.607, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.2, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184600 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3351, "doc_norm": 2.7021, "encoder_q-embeddings": 3454.1748, "encoder_q-layer.0": 2495.2944, "encoder_q-layer.1": 2885.0195, "encoder_q-layer.10": 3925.7402, "encoder_q-layer.11": 8992.5703, "encoder_q-layer.2": 3282.9473, "encoder_q-layer.3": 3518.7129, "encoder_q-layer.4": 3891.6367, "encoder_q-layer.5": 4139.9229, "encoder_q-layer.6": 4156.9355, "encoder_q-layer.7": 3938.0669, "encoder_q-layer.8": 3985.8645, "encoder_q-layer.9": 3500.7671, "epoch": 0.8, "inbatch_neg_score": 5.4249, "inbatch_pos_score": 6.0312, "learning_rate": 4.026315789473684e-06, "loss": 3.3351, "norm_diff": 0.0782, "norm_loss": 0.0, "num_token_doc": 66.7288, "num_token_overlap": 11.6779, "num_token_query": 31.3482, "num_token_union": 65.0463, "num_word_context": 202.4504, "num_word_doc": 49.7941, "num_word_query": 23.2947, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6498.6578, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4219, "query_norm": 2.624, "queue_k_norm": 2.7016, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3482, "sent_len_1": 66.7288, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.69, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184700 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3346, "doc_norm": 2.7009, "encoder_q-embeddings": 2283.9712, "encoder_q-layer.0": 1631.0958, "encoder_q-layer.1": 1708.9048, "encoder_q-layer.10": 3801.6558, "encoder_q-layer.11": 8622.5254, "encoder_q-layer.2": 1953.0764, "encoder_q-layer.3": 1995.3938, "encoder_q-layer.4": 2115.582, "encoder_q-layer.5": 2176.718, "encoder_q-layer.6": 2485.7861, "encoder_q-layer.7": 2830.4753, "encoder_q-layer.8": 3267.9968, "encoder_q-layer.9": 3080.8711, "epoch": 0.8, "inbatch_neg_score": 5.4242, "inbatch_pos_score": 6.0391, "learning_rate": 4.000000000000001e-06, "loss": 3.3346, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.9806, "num_token_overlap": 11.691, "num_token_query": 31.3596, "num_token_union": 65.2101, "num_word_context": 202.3641, "num_word_doc": 49.9529, "num_word_query": 23.284, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5282.7223, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4258, "query_norm": 2.6228, "queue_k_norm": 2.7013, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3596, "sent_len_1": 66.9806, "sent_len_max_0": 127.9513, "sent_len_max_1": 208.76, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 184800 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3311, "doc_norm": 2.7011, "encoder_q-embeddings": 2394.0178, "encoder_q-layer.0": 1659.4702, "encoder_q-layer.1": 1812.6582, "encoder_q-layer.10": 3696.7209, "encoder_q-layer.11": 8724.3887, "encoder_q-layer.2": 2065.3994, "encoder_q-layer.3": 2123.2395, "encoder_q-layer.4": 2299.3005, "encoder_q-layer.5": 2388.7417, "encoder_q-layer.6": 2645.8811, "encoder_q-layer.7": 3059.4465, "encoder_q-layer.8": 3473.7043, "encoder_q-layer.9": 3118.6467, "epoch": 0.8, "inbatch_neg_score": 5.4276, "inbatch_pos_score": 6.0312, "learning_rate": 3.9736842105263165e-06, "loss": 3.3311, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.7779, "num_token_overlap": 11.6966, "num_token_query": 31.3827, "num_token_union": 65.079, "num_word_context": 202.1055, "num_word_doc": 49.809, "num_word_query": 23.3152, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5462.1993, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4258, "query_norm": 2.6196, "queue_k_norm": 2.7021, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3827, "sent_len_1": 66.7779, "sent_len_max_0": 127.9762, "sent_len_max_1": 210.02, "stdk": 0.0494, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 184900 }, { "accuracy": 53.2715, "active_queue_size": 16384.0, "cl_loss": 3.348, "doc_norm": 2.7019, "encoder_q-embeddings": 2210.2463, "encoder_q-layer.0": 1617.1384, "encoder_q-layer.1": 1718.5251, "encoder_q-layer.10": 3350.4487, "encoder_q-layer.11": 8506.8555, "encoder_q-layer.2": 1907.59, "encoder_q-layer.3": 1955.1211, "encoder_q-layer.4": 2099.3936, "encoder_q-layer.5": 2216.739, "encoder_q-layer.6": 2442.5986, "encoder_q-layer.7": 2791.1931, "encoder_q-layer.8": 3285.6694, "encoder_q-layer.9": 3095.1936, "epoch": 0.8, "inbatch_neg_score": 5.4287, "inbatch_pos_score": 6.0586, "learning_rate": 3.9473684210526315e-06, "loss": 3.348, "norm_diff": 0.0745, "norm_loss": 0.0, "num_token_doc": 66.7928, "num_token_overlap": 11.674, "num_token_query": 31.3292, "num_token_union": 65.1001, "num_word_context": 202.3839, "num_word_doc": 49.8355, "num_word_query": 23.262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5215.0623, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.4297, "query_norm": 2.6274, "queue_k_norm": 2.7021, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3292, "sent_len_1": 66.7928, "sent_len_max_0": 127.995, "sent_len_max_1": 207.915, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185000 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3334, "doc_norm": 2.7002, "encoder_q-embeddings": 2368.3159, "encoder_q-layer.0": 1685.9341, "encoder_q-layer.1": 1813.3138, "encoder_q-layer.10": 3620.1726, "encoder_q-layer.11": 9260.7236, "encoder_q-layer.2": 2037.5913, "encoder_q-layer.3": 2111.719, "encoder_q-layer.4": 2375.1736, "encoder_q-layer.5": 2432.1433, "encoder_q-layer.6": 2703.1702, "encoder_q-layer.7": 3018.2351, "encoder_q-layer.8": 3665.3745, "encoder_q-layer.9": 3379.748, "epoch": 0.8, "inbatch_neg_score": 5.4294, "inbatch_pos_score": 6.0391, "learning_rate": 3.921052631578947e-06, "loss": 3.3334, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.5932, "num_token_overlap": 11.6505, "num_token_query": 31.369, "num_token_union": 65.0038, "num_word_context": 202.2728, "num_word_doc": 49.6828, "num_word_query": 23.3131, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5648.8551, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4297, "query_norm": 2.6231, "queue_k_norm": 2.7022, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.369, "sent_len_1": 66.5932, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.8, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185100 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3389, "doc_norm": 2.7015, "encoder_q-embeddings": 2463.0344, "encoder_q-layer.0": 1740.8815, "encoder_q-layer.1": 1844.7401, "encoder_q-layer.10": 3800.4646, "encoder_q-layer.11": 8910.582, "encoder_q-layer.2": 2026.2953, "encoder_q-layer.3": 2028.6959, "encoder_q-layer.4": 2097.0852, "encoder_q-layer.5": 2171.8428, "encoder_q-layer.6": 2442.999, "encoder_q-layer.7": 2821.3318, "encoder_q-layer.8": 3539.1348, "encoder_q-layer.9": 3287.155, "epoch": 0.8, "inbatch_neg_score": 5.4327, "inbatch_pos_score": 6.0352, "learning_rate": 3.894736842105264e-06, "loss": 3.3389, "norm_diff": 0.0795, "norm_loss": 0.0, "num_token_doc": 66.6943, "num_token_overlap": 11.6771, "num_token_query": 31.3523, "num_token_union": 65.0377, "num_word_context": 202.2349, "num_word_doc": 49.7518, "num_word_query": 23.2917, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5485.096, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4297, "query_norm": 2.622, "queue_k_norm": 2.7031, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3523, "sent_len_1": 66.6943, "sent_len_max_0": 127.9988, "sent_len_max_1": 209.2875, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185200 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3367, "doc_norm": 2.7016, "encoder_q-embeddings": 4802.8911, "encoder_q-layer.0": 3477.8879, "encoder_q-layer.1": 3886.9551, "encoder_q-layer.10": 6454.0371, "encoder_q-layer.11": 16858.9102, "encoder_q-layer.2": 4259.5835, "encoder_q-layer.3": 4341.9209, "encoder_q-layer.4": 4571.1753, "encoder_q-layer.5": 4816.2998, "encoder_q-layer.6": 5375.666, "encoder_q-layer.7": 5840.7158, "encoder_q-layer.8": 6830.3574, "encoder_q-layer.9": 5987.7749, "epoch": 0.8, "inbatch_neg_score": 5.4332, "inbatch_pos_score": 6.0547, "learning_rate": 3.868421052631579e-06, "loss": 3.3367, "norm_diff": 0.0739, "norm_loss": 0.0, "num_token_doc": 66.7045, "num_token_overlap": 11.6683, "num_token_query": 31.3715, "num_token_union": 65.0819, "num_word_context": 202.1454, "num_word_doc": 49.7514, "num_word_query": 23.2945, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10680.3783, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4297, "query_norm": 2.6276, "queue_k_norm": 2.7036, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3715, "sent_len_1": 66.7045, "sent_len_max_0": 127.985, "sent_len_max_1": 209.1863, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185300 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3341, "doc_norm": 2.7028, "encoder_q-embeddings": 4854.9814, "encoder_q-layer.0": 3461.3855, "encoder_q-layer.1": 3742.5439, "encoder_q-layer.10": 6867.3286, "encoder_q-layer.11": 17710.3906, "encoder_q-layer.2": 4202.7788, "encoder_q-layer.3": 4330.7012, "encoder_q-layer.4": 4814.7026, "encoder_q-layer.5": 5168.1216, "encoder_q-layer.6": 5651.7207, "encoder_q-layer.7": 6112.1328, "encoder_q-layer.8": 7396.146, "encoder_q-layer.9": 6700.271, "epoch": 0.8, "inbatch_neg_score": 5.4342, "inbatch_pos_score": 6.0469, "learning_rate": 3.842105263157895e-06, "loss": 3.3341, "norm_diff": 0.0791, "norm_loss": 0.0, "num_token_doc": 66.8101, "num_token_overlap": 11.6822, "num_token_query": 31.3554, "num_token_union": 65.1193, "num_word_context": 202.3797, "num_word_doc": 49.8264, "num_word_query": 23.2687, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11107.4791, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4336, "query_norm": 2.6237, "queue_k_norm": 2.7037, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3554, "sent_len_1": 66.8101, "sent_len_max_0": 127.995, "sent_len_max_1": 208.045, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185400 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3383, "doc_norm": 2.7027, "encoder_q-embeddings": 4918.4434, "encoder_q-layer.0": 3404.5256, "encoder_q-layer.1": 3626.8081, "encoder_q-layer.10": 7574.4019, "encoder_q-layer.11": 18324.1113, "encoder_q-layer.2": 4038.554, "encoder_q-layer.3": 4224.2744, "encoder_q-layer.4": 4569.5405, "encoder_q-layer.5": 4758.9321, "encoder_q-layer.6": 5203.7524, "encoder_q-layer.7": 5829.2563, "encoder_q-layer.8": 7037.0615, "encoder_q-layer.9": 6709.2158, "epoch": 0.8, "inbatch_neg_score": 5.4325, "inbatch_pos_score": 6.0508, "learning_rate": 3.8157894736842105e-06, "loss": 3.3383, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.7772, "num_token_overlap": 11.6428, "num_token_query": 31.287, "num_token_union": 65.0637, "num_word_context": 202.3105, "num_word_doc": 49.8373, "num_word_query": 23.2201, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11365.1607, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4336, "query_norm": 2.6224, "queue_k_norm": 2.7041, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.287, "sent_len_1": 66.7772, "sent_len_max_0": 127.9338, "sent_len_max_1": 208.2488, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185500 }, { "accuracy": 55.0781, "active_queue_size": 16384.0, "cl_loss": 3.3405, "doc_norm": 2.7045, "encoder_q-embeddings": 4535.5181, "encoder_q-layer.0": 3199.5381, "encoder_q-layer.1": 3353.5508, "encoder_q-layer.10": 6360.4404, "encoder_q-layer.11": 17054.0684, "encoder_q-layer.2": 3765.2036, "encoder_q-layer.3": 3864.5896, "encoder_q-layer.4": 4119.7817, "encoder_q-layer.5": 4281.2661, "encoder_q-layer.6": 4725.335, "encoder_q-layer.7": 5496.3379, "encoder_q-layer.8": 6585.7422, "encoder_q-layer.9": 6118.687, "epoch": 0.81, "inbatch_neg_score": 5.4323, "inbatch_pos_score": 6.0703, "learning_rate": 3.789473684210527e-06, "loss": 3.3405, "norm_diff": 0.0768, "norm_loss": 0.0, "num_token_doc": 66.8162, "num_token_overlap": 11.6511, "num_token_query": 31.3383, "num_token_union": 65.1418, "num_word_context": 202.4498, "num_word_doc": 49.8582, "num_word_query": 23.2802, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10503.1706, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4336, "query_norm": 2.6277, "queue_k_norm": 2.7035, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3383, "sent_len_1": 66.8162, "sent_len_max_0": 127.9887, "sent_len_max_1": 210.4137, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 185600 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3471, "doc_norm": 2.7019, "encoder_q-embeddings": 4757.2817, "encoder_q-layer.0": 3296.4844, "encoder_q-layer.1": 3537.6216, "encoder_q-layer.10": 7133.3853, "encoder_q-layer.11": 17755.0898, "encoder_q-layer.2": 3935.0339, "encoder_q-layer.3": 4165.0415, "encoder_q-layer.4": 4556.875, "encoder_q-layer.5": 4908.7402, "encoder_q-layer.6": 5368.6938, "encoder_q-layer.7": 5963.1621, "encoder_q-layer.8": 7453.3843, "encoder_q-layer.9": 6538.6274, "epoch": 0.81, "inbatch_neg_score": 5.4345, "inbatch_pos_score": 6.0469, "learning_rate": 3.763157894736842e-06, "loss": 3.3471, "norm_diff": 0.0791, "norm_loss": 0.0, "num_token_doc": 66.6375, "num_token_overlap": 11.6737, "num_token_query": 31.3395, "num_token_union": 64.9973, "num_word_context": 202.1271, "num_word_doc": 49.6891, "num_word_query": 23.2685, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11108.4119, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4336, "query_norm": 2.6228, "queue_k_norm": 2.7043, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3395, "sent_len_1": 66.6375, "sent_len_max_0": 127.9963, "sent_len_max_1": 210.3938, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 185700 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3406, "doc_norm": 2.7029, "encoder_q-embeddings": 4782.583, "encoder_q-layer.0": 3309.4104, "encoder_q-layer.1": 3567.2822, "encoder_q-layer.10": 7282.0923, "encoder_q-layer.11": 17956.2168, "encoder_q-layer.2": 3955.2009, "encoder_q-layer.3": 4057.2012, "encoder_q-layer.4": 4433.5864, "encoder_q-layer.5": 4469.3608, "encoder_q-layer.6": 5202.4873, "encoder_q-layer.7": 5705.8672, "encoder_q-layer.8": 6962.8999, "encoder_q-layer.9": 6544.8984, "epoch": 0.81, "inbatch_neg_score": 5.436, "inbatch_pos_score": 6.0547, "learning_rate": 3.736842105263158e-06, "loss": 3.3406, "norm_diff": 0.0761, "norm_loss": 0.0, "num_token_doc": 66.7934, "num_token_overlap": 11.699, "num_token_query": 31.4461, "num_token_union": 65.135, "num_word_context": 202.1313, "num_word_doc": 49.8394, "num_word_query": 23.3642, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10988.0952, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4375, "query_norm": 2.6267, "queue_k_norm": 2.7037, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4461, "sent_len_1": 66.7934, "sent_len_max_0": 127.985, "sent_len_max_1": 208.0263, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 185800 }, { "accuracy": 47.8027, "active_queue_size": 16384.0, "cl_loss": 3.3317, "doc_norm": 2.704, "encoder_q-embeddings": 4695.813, "encoder_q-layer.0": 3359.5056, "encoder_q-layer.1": 3530.5554, "encoder_q-layer.10": 8339.3379, "encoder_q-layer.11": 18786.4824, "encoder_q-layer.2": 3916.187, "encoder_q-layer.3": 3962.3269, "encoder_q-layer.4": 4329.4175, "encoder_q-layer.5": 4522.6484, "encoder_q-layer.6": 5173.6768, "encoder_q-layer.7": 5819.188, "encoder_q-layer.8": 7290.1738, "encoder_q-layer.9": 7087.6001, "epoch": 0.81, "inbatch_neg_score": 5.4394, "inbatch_pos_score": 6.043, "learning_rate": 3.710526315789474e-06, "loss": 3.3317, "norm_diff": 0.0771, "norm_loss": 0.0, "num_token_doc": 66.8217, "num_token_overlap": 11.6621, "num_token_query": 31.3359, "num_token_union": 65.1509, "num_word_context": 202.1877, "num_word_doc": 49.874, "num_word_query": 23.2865, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11279.6084, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4375, "query_norm": 2.6269, "queue_k_norm": 2.7042, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3359, "sent_len_1": 66.8217, "sent_len_max_0": 127.9737, "sent_len_max_1": 208.725, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 185900 }, { "accuracy": 52.4902, "active_queue_size": 16384.0, "cl_loss": 3.3378, "doc_norm": 2.7062, "encoder_q-embeddings": 4798.6387, "encoder_q-layer.0": 3388.5002, "encoder_q-layer.1": 3664.885, "encoder_q-layer.10": 6816.897, "encoder_q-layer.11": 17061.5215, "encoder_q-layer.2": 4122.9404, "encoder_q-layer.3": 4325.856, "encoder_q-layer.4": 4706.5308, "encoder_q-layer.5": 4785.5298, "encoder_q-layer.6": 5452.7974, "encoder_q-layer.7": 6118.4771, "encoder_q-layer.8": 7033.6206, "encoder_q-layer.9": 6348.3936, "epoch": 0.81, "inbatch_neg_score": 5.4414, "inbatch_pos_score": 6.0664, "learning_rate": 3.6842105263157892e-06, "loss": 3.3378, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.6506, "num_token_overlap": 11.6865, "num_token_query": 31.4872, "num_token_union": 65.1193, "num_word_context": 202.2798, "num_word_doc": 49.7681, "num_word_query": 23.3897, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10802.8353, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4414, "query_norm": 2.6282, "queue_k_norm": 2.7047, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4872, "sent_len_1": 66.6506, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.6863, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186000 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3234, "doc_norm": 2.704, "encoder_q-embeddings": 4541.2856, "encoder_q-layer.0": 3255.3643, "encoder_q-layer.1": 3413.8442, "encoder_q-layer.10": 6774.3784, "encoder_q-layer.11": 17113.8613, "encoder_q-layer.2": 3813.0654, "encoder_q-layer.3": 3896.0635, "encoder_q-layer.4": 4237.7837, "encoder_q-layer.5": 4364.938, "encoder_q-layer.6": 5013.7866, "encoder_q-layer.7": 5775.7119, "encoder_q-layer.8": 6931.9902, "encoder_q-layer.9": 6379.6606, "epoch": 0.81, "inbatch_neg_score": 5.4416, "inbatch_pos_score": 6.0547, "learning_rate": 3.6578947368421055e-06, "loss": 3.3234, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.7404, "num_token_overlap": 11.7326, "num_token_query": 31.4031, "num_token_union": 65.061, "num_word_context": 201.8804, "num_word_doc": 49.8108, "num_word_query": 23.3163, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10608.2873, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4414, "query_norm": 2.6253, "queue_k_norm": 2.7048, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4031, "sent_len_1": 66.7404, "sent_len_max_0": 127.9587, "sent_len_max_1": 208.125, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 186100 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.3408, "doc_norm": 2.7086, "encoder_q-embeddings": 4752.0742, "encoder_q-layer.0": 3472.5686, "encoder_q-layer.1": 3698.7173, "encoder_q-layer.10": 7016.585, "encoder_q-layer.11": 18268.8555, "encoder_q-layer.2": 4174.7402, "encoder_q-layer.3": 4393.4189, "encoder_q-layer.4": 4671.4336, "encoder_q-layer.5": 4754.208, "encoder_q-layer.6": 5224.3281, "encoder_q-layer.7": 5872.3096, "encoder_q-layer.8": 7062.791, "encoder_q-layer.9": 6529.1816, "epoch": 0.81, "inbatch_neg_score": 5.4425, "inbatch_pos_score": 6.0547, "learning_rate": 3.6315789473684213e-06, "loss": 3.3408, "norm_diff": 0.0804, "norm_loss": 0.0, "num_token_doc": 66.9494, "num_token_overlap": 11.7142, "num_token_query": 31.38, "num_token_union": 65.2243, "num_word_context": 202.2864, "num_word_doc": 49.9609, "num_word_query": 23.3026, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11253.3178, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4414, "query_norm": 2.6281, "queue_k_norm": 2.705, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.38, "sent_len_1": 66.9494, "sent_len_max_0": 127.9887, "sent_len_max_1": 208.7188, "stdk": 0.0498, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186200 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3349, "doc_norm": 2.7038, "encoder_q-embeddings": 5455.3857, "encoder_q-layer.0": 3959.2756, "encoder_q-layer.1": 4200.4834, "encoder_q-layer.10": 6861.6069, "encoder_q-layer.11": 17153.502, "encoder_q-layer.2": 4729.8765, "encoder_q-layer.3": 4722.1323, "encoder_q-layer.4": 5085.8735, "encoder_q-layer.5": 5039.3091, "encoder_q-layer.6": 5628.519, "encoder_q-layer.7": 6582.4746, "encoder_q-layer.8": 7532.313, "encoder_q-layer.9": 6455.8037, "epoch": 0.81, "inbatch_neg_score": 5.4444, "inbatch_pos_score": 6.0586, "learning_rate": 3.605263157894737e-06, "loss": 3.3349, "norm_diff": 0.0753, "norm_loss": 0.0, "num_token_doc": 66.7666, "num_token_overlap": 11.6693, "num_token_query": 31.3459, "num_token_union": 65.0807, "num_word_context": 202.2118, "num_word_doc": 49.8092, "num_word_query": 23.2852, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11336.4091, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4453, "query_norm": 2.6286, "queue_k_norm": 2.7047, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3459, "sent_len_1": 66.7666, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.6775, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 186300 }, { "accuracy": 50.4395, "active_queue_size": 16384.0, "cl_loss": 3.3388, "doc_norm": 2.7049, "encoder_q-embeddings": 4785.4541, "encoder_q-layer.0": 3647.0894, "encoder_q-layer.1": 4024.8916, "encoder_q-layer.10": 8090.627, "encoder_q-layer.11": 17556.0488, "encoder_q-layer.2": 4486.1929, "encoder_q-layer.3": 4616.0098, "encoder_q-layer.4": 5138.8931, "encoder_q-layer.5": 4939.8335, "encoder_q-layer.6": 5352.7505, "encoder_q-layer.7": 6040.9004, "encoder_q-layer.8": 7026.917, "encoder_q-layer.9": 6498.1914, "epoch": 0.81, "inbatch_neg_score": 5.4462, "inbatch_pos_score": 6.0547, "learning_rate": 3.5789473684210525e-06, "loss": 3.3388, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.8178, "num_token_overlap": 11.7091, "num_token_query": 31.4656, "num_token_union": 65.157, "num_word_context": 202.2718, "num_word_doc": 49.8248, "num_word_query": 23.3636, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11196.7588, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4453, "query_norm": 2.6268, "queue_k_norm": 2.7056, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4656, "sent_len_1": 66.8178, "sent_len_max_0": 127.9575, "sent_len_max_1": 210.8025, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 186400 }, { "accuracy": 50.4883, "active_queue_size": 16384.0, "cl_loss": 3.342, "doc_norm": 2.7083, "encoder_q-embeddings": 5072.1055, "encoder_q-layer.0": 3649.1929, "encoder_q-layer.1": 3874.26, "encoder_q-layer.10": 7702.2246, "encoder_q-layer.11": 18100.5332, "encoder_q-layer.2": 4292.856, "encoder_q-layer.3": 4427.1455, "encoder_q-layer.4": 4791.5303, "encoder_q-layer.5": 5079.4082, "encoder_q-layer.6": 5387.5649, "encoder_q-layer.7": 5979.1675, "encoder_q-layer.8": 7496.8696, "encoder_q-layer.9": 6859.5972, "epoch": 0.81, "inbatch_neg_score": 5.4479, "inbatch_pos_score": 6.0664, "learning_rate": 3.5526315789473687e-06, "loss": 3.342, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.7435, "num_token_overlap": 11.6583, "num_token_query": 31.3128, "num_token_union": 65.0589, "num_word_context": 202.1498, "num_word_doc": 49.7813, "num_word_query": 23.2408, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11228.2115, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4492, "query_norm": 2.6325, "queue_k_norm": 2.706, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3128, "sent_len_1": 66.7435, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.34, "stdk": 0.0497, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186500 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.3275, "doc_norm": 2.7067, "encoder_q-embeddings": 4755.6797, "encoder_q-layer.0": 3332.9895, "encoder_q-layer.1": 3532.5264, "encoder_q-layer.10": 7401.4609, "encoder_q-layer.11": 17569.4668, "encoder_q-layer.2": 3874.6414, "encoder_q-layer.3": 3974.678, "encoder_q-layer.4": 4375.9229, "encoder_q-layer.5": 4558.8882, "encoder_q-layer.6": 5055.1265, "encoder_q-layer.7": 5851.5605, "encoder_q-layer.8": 7447.2021, "encoder_q-layer.9": 6632.6152, "epoch": 0.81, "inbatch_neg_score": 5.4495, "inbatch_pos_score": 6.0703, "learning_rate": 3.5263157894736846e-06, "loss": 3.3275, "norm_diff": 0.0772, "norm_loss": 0.0, "num_token_doc": 66.7423, "num_token_overlap": 11.686, "num_token_query": 31.4001, "num_token_union": 65.1012, "num_word_context": 202.2739, "num_word_doc": 49.831, "num_word_query": 23.3078, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11202.2279, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4492, "query_norm": 2.6295, "queue_k_norm": 2.7065, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4001, "sent_len_1": 66.7423, "sent_len_max_0": 127.9912, "sent_len_max_1": 207.5687, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186600 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3422, "doc_norm": 2.7054, "encoder_q-embeddings": 5005.126, "encoder_q-layer.0": 3580.1729, "encoder_q-layer.1": 3997.8591, "encoder_q-layer.10": 6835.5352, "encoder_q-layer.11": 17002.832, "encoder_q-layer.2": 4414.4541, "encoder_q-layer.3": 4447.834, "encoder_q-layer.4": 4825.5483, "encoder_q-layer.5": 5007.4209, "encoder_q-layer.6": 5451.8555, "encoder_q-layer.7": 6174.6836, "encoder_q-layer.8": 6705.2427, "encoder_q-layer.9": 6213.2935, "epoch": 0.81, "inbatch_neg_score": 5.4516, "inbatch_pos_score": 6.0703, "learning_rate": 3.5000000000000004e-06, "loss": 3.3422, "norm_diff": 0.074, "norm_loss": 0.0, "num_token_doc": 66.7765, "num_token_overlap": 11.6734, "num_token_query": 31.3555, "num_token_union": 65.1074, "num_word_context": 202.2264, "num_word_doc": 49.8154, "num_word_query": 23.2847, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10945.3802, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4492, "query_norm": 2.6314, "queue_k_norm": 2.7065, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3555, "sent_len_1": 66.7765, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.5163, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186700 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3314, "doc_norm": 2.7038, "encoder_q-embeddings": 5361.522, "encoder_q-layer.0": 3859.1204, "encoder_q-layer.1": 4244.7847, "encoder_q-layer.10": 7866.2734, "encoder_q-layer.11": 18478.2598, "encoder_q-layer.2": 4962.9546, "encoder_q-layer.3": 5066.4814, "encoder_q-layer.4": 5334.272, "encoder_q-layer.5": 5585.6821, "encoder_q-layer.6": 6124.3735, "encoder_q-layer.7": 6653.3125, "encoder_q-layer.8": 7756.6235, "encoder_q-layer.9": 7147.499, "epoch": 0.81, "inbatch_neg_score": 5.4534, "inbatch_pos_score": 6.0625, "learning_rate": 3.4736842105263158e-06, "loss": 3.3314, "norm_diff": 0.0736, "norm_loss": 0.0, "num_token_doc": 66.8382, "num_token_overlap": 11.6436, "num_token_query": 31.3425, "num_token_union": 65.1507, "num_word_context": 202.3818, "num_word_doc": 49.8438, "num_word_query": 23.262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11891.569, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4531, "query_norm": 2.6302, "queue_k_norm": 2.707, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3425, "sent_len_1": 66.8382, "sent_len_max_0": 127.9688, "sent_len_max_1": 211.4162, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 186800 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3398, "doc_norm": 2.7058, "encoder_q-embeddings": 5399.0181, "encoder_q-layer.0": 3893.0969, "encoder_q-layer.1": 4312.5522, "encoder_q-layer.10": 7436.4692, "encoder_q-layer.11": 17939.2246, "encoder_q-layer.2": 4916.6064, "encoder_q-layer.3": 5254.0298, "encoder_q-layer.4": 5783.2275, "encoder_q-layer.5": 5848.2373, "encoder_q-layer.6": 6299.3799, "encoder_q-layer.7": 6333.2017, "encoder_q-layer.8": 7184.5415, "encoder_q-layer.9": 6579.6948, "epoch": 0.81, "inbatch_neg_score": 5.4542, "inbatch_pos_score": 6.0664, "learning_rate": 3.4473684210526316e-06, "loss": 3.3398, "norm_diff": 0.0783, "norm_loss": 0.0, "num_token_doc": 66.6213, "num_token_overlap": 11.6578, "num_token_query": 31.3527, "num_token_union": 65.027, "num_word_context": 201.8943, "num_word_doc": 49.6949, "num_word_query": 23.2835, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11849.1539, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.4531, "query_norm": 2.6275, "queue_k_norm": 2.7066, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3527, "sent_len_1": 66.6213, "sent_len_max_0": 127.9862, "sent_len_max_1": 207.1763, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 186900 }, { "accuracy": 53.3203, "active_queue_size": 16384.0, "cl_loss": 3.3418, "doc_norm": 2.7091, "encoder_q-embeddings": 4950.5737, "encoder_q-layer.0": 3581.3311, "encoder_q-layer.1": 3839.5391, "encoder_q-layer.10": 6663.9414, "encoder_q-layer.11": 16974.1895, "encoder_q-layer.2": 4175.104, "encoder_q-layer.3": 4232.2637, "encoder_q-layer.4": 4496.771, "encoder_q-layer.5": 4498.7202, "encoder_q-layer.6": 5016.1377, "encoder_q-layer.7": 5784.7866, "encoder_q-layer.8": 6665.582, "encoder_q-layer.9": 6473.7114, "epoch": 0.81, "inbatch_neg_score": 5.455, "inbatch_pos_score": 6.0859, "learning_rate": 3.421052631578948e-06, "loss": 3.3418, "norm_diff": 0.0782, "norm_loss": 0.0, "num_token_doc": 67.0003, "num_token_overlap": 11.6681, "num_token_query": 31.3011, "num_token_union": 65.2171, "num_word_context": 202.4908, "num_word_doc": 49.9929, "num_word_query": 23.2405, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10617.0545, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.457, "query_norm": 2.6309, "queue_k_norm": 2.7072, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3011, "sent_len_1": 67.0003, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.5325, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187000 }, { "accuracy": 53.5156, "active_queue_size": 16384.0, "cl_loss": 3.3281, "doc_norm": 2.709, "encoder_q-embeddings": 4759.1543, "encoder_q-layer.0": 3257.49, "encoder_q-layer.1": 3495.7297, "encoder_q-layer.10": 7793.6763, "encoder_q-layer.11": 18860.125, "encoder_q-layer.2": 3907.5215, "encoder_q-layer.3": 4058.865, "encoder_q-layer.4": 4384.6235, "encoder_q-layer.5": 4502.8472, "encoder_q-layer.6": 5026.3291, "encoder_q-layer.7": 6186.7812, "encoder_q-layer.8": 7607.5654, "encoder_q-layer.9": 6943.8311, "epoch": 0.81, "inbatch_neg_score": 5.4571, "inbatch_pos_score": 6.0859, "learning_rate": 3.3947368421052632e-06, "loss": 3.3281, "norm_diff": 0.0767, "norm_loss": 0.0, "num_token_doc": 66.7317, "num_token_overlap": 11.6923, "num_token_query": 31.4155, "num_token_union": 65.0928, "num_word_context": 202.1371, "num_word_doc": 49.7672, "num_word_query": 23.3319, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11512.2898, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.457, "query_norm": 2.6323, "queue_k_norm": 2.7076, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4155, "sent_len_1": 66.7317, "sent_len_max_0": 127.9813, "sent_len_max_1": 211.105, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187100 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.346, "doc_norm": 2.7093, "encoder_q-embeddings": 4894.6636, "encoder_q-layer.0": 3535.6599, "encoder_q-layer.1": 3715.6318, "encoder_q-layer.10": 7359.1162, "encoder_q-layer.11": 17789.5664, "encoder_q-layer.2": 4186.2178, "encoder_q-layer.3": 4452.1406, "encoder_q-layer.4": 4709.8076, "encoder_q-layer.5": 4792.939, "encoder_q-layer.6": 5199.0615, "encoder_q-layer.7": 5682.1104, "encoder_q-layer.8": 7162.9336, "encoder_q-layer.9": 6499.752, "epoch": 0.81, "inbatch_neg_score": 5.4583, "inbatch_pos_score": 6.0703, "learning_rate": 3.368421052631579e-06, "loss": 3.346, "norm_diff": 0.0847, "norm_loss": 0.0, "num_token_doc": 66.8669, "num_token_overlap": 11.6779, "num_token_query": 31.3211, "num_token_union": 65.0969, "num_word_context": 202.2837, "num_word_doc": 49.8639, "num_word_query": 23.2581, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11152.7507, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.457, "query_norm": 2.6246, "queue_k_norm": 2.7079, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3211, "sent_len_1": 66.8669, "sent_len_max_0": 127.9775, "sent_len_max_1": 210.595, "stdk": 0.0496, "stdq": 0.0427, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187200 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.347, "doc_norm": 2.7081, "encoder_q-embeddings": 9155.5654, "encoder_q-layer.0": 6399.5918, "encoder_q-layer.1": 6826.9258, "encoder_q-layer.10": 13470.8857, "encoder_q-layer.11": 34560.3281, "encoder_q-layer.2": 7794.041, "encoder_q-layer.3": 7675.4775, "encoder_q-layer.4": 8379.4121, "encoder_q-layer.5": 8815.002, "encoder_q-layer.6": 9901.3047, "encoder_q-layer.7": 11289.8467, "encoder_q-layer.8": 13276.0596, "encoder_q-layer.9": 12582.8271, "epoch": 0.81, "inbatch_neg_score": 5.4608, "inbatch_pos_score": 6.082, "learning_rate": 3.342105263157895e-06, "loss": 3.347, "norm_diff": 0.0778, "norm_loss": 0.0, "num_token_doc": 66.578, "num_token_overlap": 11.6408, "num_token_query": 31.2434, "num_token_union": 64.9329, "num_word_context": 202.0312, "num_word_doc": 49.6715, "num_word_query": 23.1974, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21361.1471, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4609, "query_norm": 2.6302, "queue_k_norm": 2.7073, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2434, "sent_len_1": 66.578, "sent_len_max_0": 127.9663, "sent_len_max_1": 207.9338, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 187300 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3233, "doc_norm": 2.7088, "encoder_q-embeddings": 10105.6279, "encoder_q-layer.0": 7272.0811, "encoder_q-layer.1": 8114.3188, "encoder_q-layer.10": 14853.21, "encoder_q-layer.11": 35115.082, "encoder_q-layer.2": 9350.7578, "encoder_q-layer.3": 9535.8809, "encoder_q-layer.4": 10477.0645, "encoder_q-layer.5": 10972.8496, "encoder_q-layer.6": 11730.0518, "encoder_q-layer.7": 12926.2529, "encoder_q-layer.8": 14878.3887, "encoder_q-layer.9": 13205.083, "epoch": 0.81, "inbatch_neg_score": 5.4602, "inbatch_pos_score": 6.0703, "learning_rate": 3.315789473684211e-06, "loss": 3.3233, "norm_diff": 0.0781, "norm_loss": 0.0, "num_token_doc": 66.8727, "num_token_overlap": 11.7272, "num_token_query": 31.4815, "num_token_union": 65.2169, "num_word_context": 202.5226, "num_word_doc": 49.93, "num_word_query": 23.3977, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22788.0288, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4609, "query_norm": 2.6307, "queue_k_norm": 2.7077, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4815, "sent_len_1": 66.8727, "sent_len_max_0": 127.9975, "sent_len_max_1": 208.4112, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 187400 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.331, "doc_norm": 2.7071, "encoder_q-embeddings": 9168.9707, "encoder_q-layer.0": 6656.6626, "encoder_q-layer.1": 6965.583, "encoder_q-layer.10": 16256.9883, "encoder_q-layer.11": 35753.5742, "encoder_q-layer.2": 7781.9189, "encoder_q-layer.3": 8327.5566, "encoder_q-layer.4": 8746.0605, "encoder_q-layer.5": 9167.5146, "encoder_q-layer.6": 9963.0234, "encoder_q-layer.7": 11362.1777, "encoder_q-layer.8": 13592.2227, "encoder_q-layer.9": 12574.8311, "epoch": 0.81, "inbatch_neg_score": 5.4623, "inbatch_pos_score": 6.082, "learning_rate": 3.2894736842105265e-06, "loss": 3.331, "norm_diff": 0.073, "norm_loss": 0.0, "num_token_doc": 66.6194, "num_token_overlap": 11.6806, "num_token_query": 31.4421, "num_token_union": 65.0471, "num_word_context": 202.2777, "num_word_doc": 49.7207, "num_word_query": 23.3538, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22173.4218, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4609, "query_norm": 2.6342, "queue_k_norm": 2.7091, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4421, "sent_len_1": 66.6194, "sent_len_max_0": 128.0, "sent_len_max_1": 206.145, "stdk": 0.0494, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187500 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3402, "doc_norm": 2.71, "encoder_q-embeddings": 9924.7285, "encoder_q-layer.0": 6757.1938, "encoder_q-layer.1": 7412.6724, "encoder_q-layer.10": 14088.2725, "encoder_q-layer.11": 36636.1445, "encoder_q-layer.2": 8196.8652, "encoder_q-layer.3": 8704.4521, "encoder_q-layer.4": 9359.1143, "encoder_q-layer.5": 9788.6973, "encoder_q-layer.6": 10874.7812, "encoder_q-layer.7": 12322.3926, "encoder_q-layer.8": 14625.9893, "encoder_q-layer.9": 13575.6064, "epoch": 0.81, "inbatch_neg_score": 5.4648, "inbatch_pos_score": 6.082, "learning_rate": 3.2631578947368423e-06, "loss": 3.3402, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7089, "num_token_overlap": 11.6859, "num_token_query": 31.3708, "num_token_union": 65.0597, "num_word_context": 202.342, "num_word_doc": 49.7799, "num_word_query": 23.3019, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22949.3389, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4648, "query_norm": 2.6303, "queue_k_norm": 2.7091, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3708, "sent_len_1": 66.7089, "sent_len_max_0": 127.995, "sent_len_max_1": 207.7363, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187600 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.34, "doc_norm": 2.7118, "encoder_q-embeddings": 10192.915, "encoder_q-layer.0": 7341.1562, "encoder_q-layer.1": 7843.1763, "encoder_q-layer.10": 14007.8828, "encoder_q-layer.11": 35677.4648, "encoder_q-layer.2": 9066.8213, "encoder_q-layer.3": 9324.8545, "encoder_q-layer.4": 9966.7539, "encoder_q-layer.5": 10283.3828, "encoder_q-layer.6": 11354.8223, "encoder_q-layer.7": 12626.2979, "encoder_q-layer.8": 14339.5176, "encoder_q-layer.9": 13297.1934, "epoch": 0.81, "inbatch_neg_score": 5.4658, "inbatch_pos_score": 6.0859, "learning_rate": 3.236842105263158e-06, "loss": 3.34, "norm_diff": 0.0805, "norm_loss": 0.0, "num_token_doc": 66.8917, "num_token_overlap": 11.6707, "num_token_query": 31.3622, "num_token_union": 65.2027, "num_word_context": 202.3193, "num_word_doc": 49.8928, "num_word_query": 23.2926, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23011.8002, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4648, "query_norm": 2.6313, "queue_k_norm": 2.7097, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3622, "sent_len_1": 66.8917, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.2125, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187700 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.3262, "doc_norm": 2.7104, "encoder_q-embeddings": 9409.0244, "encoder_q-layer.0": 6531.7837, "encoder_q-layer.1": 7036.3037, "encoder_q-layer.10": 13842.125, "encoder_q-layer.11": 34616.6523, "encoder_q-layer.2": 8020.6504, "encoder_q-layer.3": 8095.1719, "encoder_q-layer.4": 8797.4443, "encoder_q-layer.5": 9231.0664, "encoder_q-layer.6": 10602.3135, "encoder_q-layer.7": 12663.6445, "encoder_q-layer.8": 13985.1709, "encoder_q-layer.9": 12801.3076, "epoch": 0.81, "inbatch_neg_score": 5.4666, "inbatch_pos_score": 6.0898, "learning_rate": 3.2105263157894735e-06, "loss": 3.3262, "norm_diff": 0.0758, "norm_loss": 0.0, "num_token_doc": 66.5987, "num_token_overlap": 11.6634, "num_token_query": 31.372, "num_token_union": 65.0208, "num_word_context": 202.0702, "num_word_doc": 49.7166, "num_word_query": 23.3164, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21828.693, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4688, "query_norm": 2.6346, "queue_k_norm": 2.7089, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.372, "sent_len_1": 66.5987, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.9925, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 187800 }, { "accuracy": 52.8809, "active_queue_size": 16384.0, "cl_loss": 3.3447, "doc_norm": 2.711, "encoder_q-embeddings": 9414.3887, "encoder_q-layer.0": 6734.9468, "encoder_q-layer.1": 7346.8052, "encoder_q-layer.10": 15342.7246, "encoder_q-layer.11": 35234.8984, "encoder_q-layer.2": 8560.8779, "encoder_q-layer.3": 8859.4844, "encoder_q-layer.4": 9192.7725, "encoder_q-layer.5": 9717.2188, "encoder_q-layer.6": 10390.2588, "encoder_q-layer.7": 11765.4062, "encoder_q-layer.8": 13736.5332, "encoder_q-layer.9": 12789.8174, "epoch": 0.82, "inbatch_neg_score": 5.4671, "inbatch_pos_score": 6.0977, "learning_rate": 3.1842105263157893e-06, "loss": 3.3447, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.7803, "num_token_overlap": 11.6594, "num_token_query": 31.3481, "num_token_union": 65.1559, "num_word_context": 202.3741, "num_word_doc": 49.8213, "num_word_query": 23.2772, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22071.3686, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4688, "query_norm": 2.6335, "queue_k_norm": 2.7105, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3481, "sent_len_1": 66.7803, "sent_len_max_0": 127.9938, "sent_len_max_1": 207.9875, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 187900 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.3237, "doc_norm": 2.7131, "encoder_q-embeddings": 9318.1406, "encoder_q-layer.0": 6430.3955, "encoder_q-layer.1": 6906.6489, "encoder_q-layer.10": 16064.5693, "encoder_q-layer.11": 37131.1562, "encoder_q-layer.2": 7548.0576, "encoder_q-layer.3": 7755.2202, "encoder_q-layer.4": 8342.2188, "encoder_q-layer.5": 8448.0244, "encoder_q-layer.6": 9862.7549, "encoder_q-layer.7": 10897.959, "encoder_q-layer.8": 13375.6064, "encoder_q-layer.9": 13091.9971, "epoch": 0.82, "inbatch_neg_score": 5.4704, "inbatch_pos_score": 6.082, "learning_rate": 3.1578947368421056e-06, "loss": 3.3237, "norm_diff": 0.081, "norm_loss": 0.0, "num_token_doc": 66.8573, "num_token_overlap": 11.7058, "num_token_query": 31.432, "num_token_union": 65.1929, "num_word_context": 202.5443, "num_word_doc": 49.8643, "num_word_query": 23.3559, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22357.4284, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4688, "query_norm": 2.6321, "queue_k_norm": 2.7101, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.432, "sent_len_1": 66.8573, "sent_len_max_0": 127.9587, "sent_len_max_1": 207.0863, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188000 }, { "accuracy": 50.6836, "active_queue_size": 16384.0, "cl_loss": 3.3438, "doc_norm": 2.7104, "encoder_q-embeddings": 9928.916, "encoder_q-layer.0": 7000.1221, "encoder_q-layer.1": 7368.7212, "encoder_q-layer.10": 15421.1484, "encoder_q-layer.11": 36728.2422, "encoder_q-layer.2": 8218.835, "encoder_q-layer.3": 8428.333, "encoder_q-layer.4": 8905.002, "encoder_q-layer.5": 9599.376, "encoder_q-layer.6": 11116.9434, "encoder_q-layer.7": 12190.9658, "encoder_q-layer.8": 14797.3604, "encoder_q-layer.9": 13640.3809, "epoch": 0.82, "inbatch_neg_score": 5.4719, "inbatch_pos_score": 6.0664, "learning_rate": 3.1315789473684214e-06, "loss": 3.3438, "norm_diff": 0.0825, "norm_loss": 0.0, "num_token_doc": 66.8696, "num_token_overlap": 11.6774, "num_token_query": 31.374, "num_token_union": 65.1795, "num_word_context": 202.361, "num_word_doc": 49.8706, "num_word_query": 23.3046, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22909.2282, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4727, "query_norm": 2.628, "queue_k_norm": 2.7106, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.374, "sent_len_1": 66.8696, "sent_len_max_0": 127.9712, "sent_len_max_1": 208.2312, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188100 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3369, "doc_norm": 2.7118, "encoder_q-embeddings": 10089.2656, "encoder_q-layer.0": 7106.188, "encoder_q-layer.1": 7808.2749, "encoder_q-layer.10": 14606.6406, "encoder_q-layer.11": 35106.3203, "encoder_q-layer.2": 9062.1162, "encoder_q-layer.3": 9616.6543, "encoder_q-layer.4": 9952.7227, "encoder_q-layer.5": 10238.7451, "encoder_q-layer.6": 11274.4775, "encoder_q-layer.7": 12007.5996, "encoder_q-layer.8": 14412.252, "encoder_q-layer.9": 13542.5049, "epoch": 0.82, "inbatch_neg_score": 5.4701, "inbatch_pos_score": 6.0938, "learning_rate": 3.1052631578947372e-06, "loss": 3.3369, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.9742, "num_token_overlap": 11.6791, "num_token_query": 31.4363, "num_token_union": 65.2652, "num_word_context": 202.7105, "num_word_doc": 49.9596, "num_word_query": 23.356, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22647.0775, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4688, "query_norm": 2.6334, "queue_k_norm": 2.7126, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4363, "sent_len_1": 66.9742, "sent_len_max_0": 127.9737, "sent_len_max_1": 207.165, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 188200 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3267, "doc_norm": 2.7124, "encoder_q-embeddings": 10611.3564, "encoder_q-layer.0": 7675.8975, "encoder_q-layer.1": 8372.5664, "encoder_q-layer.10": 14572.9473, "encoder_q-layer.11": 34969.4492, "encoder_q-layer.2": 9627.1885, "encoder_q-layer.3": 9966.7441, "encoder_q-layer.4": 10527.917, "encoder_q-layer.5": 10718.2744, "encoder_q-layer.6": 11636.6006, "encoder_q-layer.7": 12495.3701, "encoder_q-layer.8": 14136.6289, "encoder_q-layer.9": 12613.79, "epoch": 0.82, "inbatch_neg_score": 5.4744, "inbatch_pos_score": 6.0938, "learning_rate": 3.0789473684210526e-06, "loss": 3.3267, "norm_diff": 0.0775, "norm_loss": 0.0, "num_token_doc": 66.9286, "num_token_overlap": 11.703, "num_token_query": 31.3668, "num_token_union": 65.1843, "num_word_context": 202.3088, "num_word_doc": 49.9379, "num_word_query": 23.3, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22918.7947, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.6349, "queue_k_norm": 2.711, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3668, "sent_len_1": 66.9286, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.9275, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188300 }, { "accuracy": 52.6855, "active_queue_size": 16384.0, "cl_loss": 3.3341, "doc_norm": 2.7108, "encoder_q-embeddings": 9518.6094, "encoder_q-layer.0": 6504.1963, "encoder_q-layer.1": 6937.7554, "encoder_q-layer.10": 13801.8105, "encoder_q-layer.11": 34449.8047, "encoder_q-layer.2": 7725.0713, "encoder_q-layer.3": 7828.3345, "encoder_q-layer.4": 8464.9639, "encoder_q-layer.5": 8752.1787, "encoder_q-layer.6": 9882.8232, "encoder_q-layer.7": 10815.3955, "encoder_q-layer.8": 13406.1445, "encoder_q-layer.9": 12480.625, "epoch": 0.82, "inbatch_neg_score": 5.4737, "inbatch_pos_score": 6.0859, "learning_rate": 3.0526315789473684e-06, "loss": 3.3341, "norm_diff": 0.081, "norm_loss": 0.0, "num_token_doc": 66.7297, "num_token_overlap": 11.6466, "num_token_query": 31.3513, "num_token_union": 65.0972, "num_word_context": 202.2152, "num_word_doc": 49.7918, "num_word_query": 23.2864, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21120.3451, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4727, "query_norm": 2.6297, "queue_k_norm": 2.7119, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3513, "sent_len_1": 66.7297, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.5838, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188400 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3297, "doc_norm": 2.7142, "encoder_q-embeddings": 9075.7959, "encoder_q-layer.0": 6403.9883, "encoder_q-layer.1": 6890.8887, "encoder_q-layer.10": 13726.3506, "encoder_q-layer.11": 34751.0664, "encoder_q-layer.2": 7771.3042, "encoder_q-layer.3": 7980.2134, "encoder_q-layer.4": 8835.4922, "encoder_q-layer.5": 8676.5615, "encoder_q-layer.6": 9834.8125, "encoder_q-layer.7": 11099.9375, "encoder_q-layer.8": 13140.2148, "encoder_q-layer.9": 12155.1309, "epoch": 0.82, "inbatch_neg_score": 5.4755, "inbatch_pos_score": 6.0938, "learning_rate": 3.0263157894736843e-06, "loss": 3.3297, "norm_diff": 0.082, "norm_loss": 0.0, "num_token_doc": 66.8486, "num_token_overlap": 11.7078, "num_token_query": 31.4234, "num_token_union": 65.1938, "num_word_context": 202.5718, "num_word_doc": 49.9005, "num_word_query": 23.3561, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21526.4683, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.6322, "queue_k_norm": 2.7122, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4234, "sent_len_1": 66.8486, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.2837, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188500 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3431, "doc_norm": 2.712, "encoder_q-embeddings": 8911.4062, "encoder_q-layer.0": 6293.3608, "encoder_q-layer.1": 6775.8818, "encoder_q-layer.10": 13792.3896, "encoder_q-layer.11": 34203.9648, "encoder_q-layer.2": 7604.3408, "encoder_q-layer.3": 7838.2573, "encoder_q-layer.4": 8620.5557, "encoder_q-layer.5": 8891.8027, "encoder_q-layer.6": 10519.5107, "encoder_q-layer.7": 11613.7705, "encoder_q-layer.8": 13924.9512, "encoder_q-layer.9": 12336.2881, "epoch": 0.82, "inbatch_neg_score": 5.476, "inbatch_pos_score": 6.0977, "learning_rate": 3e-06, "loss": 3.3431, "norm_diff": 0.075, "norm_loss": 0.0, "num_token_doc": 66.9099, "num_token_overlap": 11.6815, "num_token_query": 31.3855, "num_token_union": 65.2428, "num_word_context": 202.765, "num_word_doc": 49.9468, "num_word_query": 23.3074, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21204.5818, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.637, "queue_k_norm": 2.7113, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3855, "sent_len_1": 66.9099, "sent_len_max_0": 128.0, "sent_len_max_1": 207.72, "stdk": 0.0496, "stdq": 0.0435, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 188600 }, { "accuracy": 50.7812, "active_queue_size": 16384.0, "cl_loss": 3.337, "doc_norm": 2.7124, "encoder_q-embeddings": 9355.1035, "encoder_q-layer.0": 6456.917, "encoder_q-layer.1": 7074.4595, "encoder_q-layer.10": 14940.0801, "encoder_q-layer.11": 36425.6211, "encoder_q-layer.2": 7748.0171, "encoder_q-layer.3": 8017.6343, "encoder_q-layer.4": 8423.3818, "encoder_q-layer.5": 8788.3096, "encoder_q-layer.6": 9736.1768, "encoder_q-layer.7": 11119.3125, "encoder_q-layer.8": 14307.6475, "encoder_q-layer.9": 13076.3594, "epoch": 0.82, "inbatch_neg_score": 5.4769, "inbatch_pos_score": 6.0898, "learning_rate": 2.973684210526316e-06, "loss": 3.337, "norm_diff": 0.0804, "norm_loss": 0.0, "num_token_doc": 66.8974, "num_token_overlap": 11.6748, "num_token_query": 31.3254, "num_token_union": 65.1562, "num_word_context": 202.3407, "num_word_doc": 49.9259, "num_word_query": 23.2619, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22354.0365, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.6319, "queue_k_norm": 2.7132, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3254, "sent_len_1": 66.8974, "sent_len_max_0": 127.9688, "sent_len_max_1": 209.31, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 188700 }, { "accuracy": 52.9785, "active_queue_size": 16384.0, "cl_loss": 3.3334, "doc_norm": 2.7125, "encoder_q-embeddings": 9441.3545, "encoder_q-layer.0": 6751.6587, "encoder_q-layer.1": 7092.0234, "encoder_q-layer.10": 14769.2139, "encoder_q-layer.11": 35576.2266, "encoder_q-layer.2": 7986.9609, "encoder_q-layer.3": 8245.5312, "encoder_q-layer.4": 9192.1123, "encoder_q-layer.5": 9377.5615, "encoder_q-layer.6": 10516.1914, "encoder_q-layer.7": 11496.5693, "encoder_q-layer.8": 13696.9512, "encoder_q-layer.9": 12734.8545, "epoch": 0.82, "inbatch_neg_score": 5.4784, "inbatch_pos_score": 6.1016, "learning_rate": 2.9473684210526317e-06, "loss": 3.3334, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.7262, "num_token_overlap": 11.6932, "num_token_query": 31.3987, "num_token_union": 65.1291, "num_word_context": 202.1294, "num_word_doc": 49.8161, "num_word_query": 23.3139, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22447.2846, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.6332, "queue_k_norm": 2.7119, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3987, "sent_len_1": 66.7262, "sent_len_max_0": 128.0, "sent_len_max_1": 208.085, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188800 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3227, "doc_norm": 2.7136, "encoder_q-embeddings": 9660.2393, "encoder_q-layer.0": 7049.0186, "encoder_q-layer.1": 7638.6001, "encoder_q-layer.10": 13922.0977, "encoder_q-layer.11": 34554.9531, "encoder_q-layer.2": 8652.5039, "encoder_q-layer.3": 9080.5938, "encoder_q-layer.4": 9862.7109, "encoder_q-layer.5": 10103.668, "encoder_q-layer.6": 11396.9277, "encoder_q-layer.7": 11873.2363, "encoder_q-layer.8": 13974.8613, "encoder_q-layer.9": 12586.1211, "epoch": 0.82, "inbatch_neg_score": 5.4784, "inbatch_pos_score": 6.0977, "learning_rate": 2.9210526315789475e-06, "loss": 3.3227, "norm_diff": 0.0778, "norm_loss": 0.0, "num_token_doc": 66.8599, "num_token_overlap": 11.6808, "num_token_query": 31.3913, "num_token_union": 65.203, "num_word_context": 202.3708, "num_word_doc": 49.891, "num_word_query": 23.3023, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22215.2959, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4766, "query_norm": 2.6357, "queue_k_norm": 2.7129, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3913, "sent_len_1": 66.8599, "sent_len_max_0": 127.97, "sent_len_max_1": 208.3013, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 188900 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3259, "doc_norm": 2.7133, "encoder_q-embeddings": 9470.3457, "encoder_q-layer.0": 6549.873, "encoder_q-layer.1": 7077.5054, "encoder_q-layer.10": 14595.5352, "encoder_q-layer.11": 36573.7305, "encoder_q-layer.2": 7940.8066, "encoder_q-layer.3": 8144.1455, "encoder_q-layer.4": 8812.498, "encoder_q-layer.5": 9024.3145, "encoder_q-layer.6": 9906.4248, "encoder_q-layer.7": 11044.6592, "encoder_q-layer.8": 14309.2324, "encoder_q-layer.9": 13020.6201, "epoch": 0.82, "inbatch_neg_score": 5.479, "inbatch_pos_score": 6.1016, "learning_rate": 2.8947368421052634e-06, "loss": 3.3259, "norm_diff": 0.0793, "norm_loss": 0.0, "num_token_doc": 66.8422, "num_token_overlap": 11.6814, "num_token_query": 31.3763, "num_token_union": 65.1621, "num_word_context": 202.3418, "num_word_doc": 49.8898, "num_word_query": 23.2958, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22275.11, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4805, "query_norm": 2.634, "queue_k_norm": 2.7132, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3763, "sent_len_1": 66.8422, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.2463, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189000 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3414, "doc_norm": 2.7145, "encoder_q-embeddings": 9836.6436, "encoder_q-layer.0": 6847.4243, "encoder_q-layer.1": 7369.3667, "encoder_q-layer.10": 13576.9648, "encoder_q-layer.11": 34710.3828, "encoder_q-layer.2": 8377.9805, "encoder_q-layer.3": 8689.1816, "encoder_q-layer.4": 9249.6484, "encoder_q-layer.5": 9675.917, "encoder_q-layer.6": 10671.7744, "encoder_q-layer.7": 11318.4316, "encoder_q-layer.8": 13859.7344, "encoder_q-layer.9": 12557.2129, "epoch": 0.82, "inbatch_neg_score": 5.4795, "inbatch_pos_score": 6.0977, "learning_rate": 2.8684210526315787e-06, "loss": 3.3414, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.8364, "num_token_overlap": 11.6748, "num_token_query": 31.3853, "num_token_union": 65.1708, "num_word_context": 202.2727, "num_word_doc": 49.8584, "num_word_query": 23.3245, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21643.2895, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4805, "query_norm": 2.6319, "queue_k_norm": 2.7132, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3853, "sent_len_1": 66.8364, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.2738, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189100 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3346, "doc_norm": 2.7117, "encoder_q-embeddings": 9151.1465, "encoder_q-layer.0": 6361.7563, "encoder_q-layer.1": 6896.5586, "encoder_q-layer.10": 13448.6709, "encoder_q-layer.11": 34174.7031, "encoder_q-layer.2": 7621.6523, "encoder_q-layer.3": 7641.9229, "encoder_q-layer.4": 8145.1104, "encoder_q-layer.5": 8406.3135, "encoder_q-layer.6": 9461.791, "encoder_q-layer.7": 11001.2822, "encoder_q-layer.8": 12828.0127, "encoder_q-layer.9": 12364.6621, "epoch": 0.82, "inbatch_neg_score": 5.4813, "inbatch_pos_score": 6.0938, "learning_rate": 2.842105263157895e-06, "loss": 3.3346, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.8894, "num_token_overlap": 11.6664, "num_token_query": 31.3298, "num_token_union": 65.1552, "num_word_context": 202.2572, "num_word_doc": 49.8953, "num_word_query": 23.2497, "postclip_grad_norm": 1.0, "preclip_grad_norm": 20984.9673, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4805, "query_norm": 2.6318, "queue_k_norm": 2.7141, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3298, "sent_len_1": 66.8894, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.99, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 189200 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.7165, "encoder_q-embeddings": 18676.6074, "encoder_q-layer.0": 13425.5264, "encoder_q-layer.1": 14156.6592, "encoder_q-layer.10": 28959.8027, "encoder_q-layer.11": 70579.7734, "encoder_q-layer.2": 16118.0039, "encoder_q-layer.3": 16535.0234, "encoder_q-layer.4": 17543.416, "encoder_q-layer.5": 18108.5234, "encoder_q-layer.6": 20620.9766, "encoder_q-layer.7": 22969.8594, "encoder_q-layer.8": 27039.3535, "encoder_q-layer.9": 25796.5332, "epoch": 0.82, "inbatch_neg_score": 5.4831, "inbatch_pos_score": 6.1016, "learning_rate": 2.8157894736842104e-06, "loss": 3.3331, "norm_diff": 0.0842, "norm_loss": 0.0, "num_token_doc": 66.7522, "num_token_overlap": 11.6754, "num_token_query": 31.3706, "num_token_union": 65.0925, "num_word_context": 202.2758, "num_word_doc": 49.8015, "num_word_query": 23.3138, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43521.9691, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.4844, "query_norm": 2.6324, "queue_k_norm": 2.7135, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3706, "sent_len_1": 66.7522, "sent_len_max_0": 127.9875, "sent_len_max_1": 210.1788, "stdk": 0.0497, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189300 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3289, "doc_norm": 2.7115, "encoder_q-embeddings": 18751.5371, "encoder_q-layer.0": 13611.4082, "encoder_q-layer.1": 14742.833, "encoder_q-layer.10": 27058.707, "encoder_q-layer.11": 67193.9531, "encoder_q-layer.2": 16807.2715, "encoder_q-layer.3": 17026.8242, "encoder_q-layer.4": 18702.9473, "encoder_q-layer.5": 18976.3262, "encoder_q-layer.6": 20961.1172, "encoder_q-layer.7": 23270.3789, "encoder_q-layer.8": 27714.2559, "encoder_q-layer.9": 24701.4004, "epoch": 0.82, "inbatch_neg_score": 5.4837, "inbatch_pos_score": 6.1055, "learning_rate": 2.7894736842105266e-06, "loss": 3.3289, "norm_diff": 0.0744, "norm_loss": 0.0, "num_token_doc": 66.7445, "num_token_overlap": 11.6905, "num_token_query": 31.3896, "num_token_union": 65.0881, "num_word_context": 202.3119, "num_word_doc": 49.8084, "num_word_query": 23.2964, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42842.0554, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.4844, "query_norm": 2.6371, "queue_k_norm": 2.7136, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3896, "sent_len_1": 66.7445, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.07, "stdk": 0.0494, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189400 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3288, "doc_norm": 2.7147, "encoder_q-embeddings": 17586.9805, "encoder_q-layer.0": 12643.7871, "encoder_q-layer.1": 13759.2402, "encoder_q-layer.10": 28965.7344, "encoder_q-layer.11": 70878.1016, "encoder_q-layer.2": 15428.7207, "encoder_q-layer.3": 15613.582, "encoder_q-layer.4": 16784.6367, "encoder_q-layer.5": 16808.7305, "encoder_q-layer.6": 19137.9375, "encoder_q-layer.7": 22096.0, "encoder_q-layer.8": 27219.0938, "encoder_q-layer.9": 26315.709, "epoch": 0.82, "inbatch_neg_score": 5.4828, "inbatch_pos_score": 6.0898, "learning_rate": 2.763157894736842e-06, "loss": 3.3288, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.6764, "num_token_overlap": 11.6649, "num_token_query": 31.331, "num_token_union": 65.0477, "num_word_context": 202.0807, "num_word_doc": 49.7617, "num_word_query": 23.2721, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43266.5385, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.4844, "query_norm": 2.6338, "queue_k_norm": 2.714, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.331, "sent_len_1": 66.6764, "sent_len_max_0": 127.9813, "sent_len_max_1": 206.8963, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189500 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3508, "doc_norm": 2.7142, "encoder_q-embeddings": 9596.1152, "encoder_q-layer.0": 6940.7515, "encoder_q-layer.1": 7626.8623, "encoder_q-layer.10": 14244.377, "encoder_q-layer.11": 36517.75, "encoder_q-layer.2": 8265.9023, "encoder_q-layer.3": 8656.6289, "encoder_q-layer.4": 9354.4277, "encoder_q-layer.5": 9411.377, "encoder_q-layer.6": 10768.1191, "encoder_q-layer.7": 12299.9023, "encoder_q-layer.8": 14584.4453, "encoder_q-layer.9": 13818.8047, "epoch": 0.82, "inbatch_neg_score": 5.4842, "inbatch_pos_score": 6.0938, "learning_rate": 2.7368421052631583e-06, "loss": 3.3508, "norm_diff": 0.0844, "norm_loss": 0.0, "num_token_doc": 66.5217, "num_token_overlap": 11.6764, "num_token_query": 31.4062, "num_token_union": 64.9674, "num_word_context": 202.0896, "num_word_doc": 49.6686, "num_word_query": 23.3261, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22987.2495, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4844, "query_norm": 2.6298, "queue_k_norm": 2.7129, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4062, "sent_len_1": 66.5217, "sent_len_max_0": 127.975, "sent_len_max_1": 206.0675, "stdk": 0.0496, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 189600 }, { "accuracy": 52.4414, "active_queue_size": 16384.0, "cl_loss": 3.3387, "doc_norm": 2.7125, "encoder_q-embeddings": 10287.9287, "encoder_q-layer.0": 7336.6348, "encoder_q-layer.1": 8120.6655, "encoder_q-layer.10": 13781.1143, "encoder_q-layer.11": 34521.4531, "encoder_q-layer.2": 9276.9053, "encoder_q-layer.3": 9452.626, "encoder_q-layer.4": 10160.46, "encoder_q-layer.5": 9965.9336, "encoder_q-layer.6": 11212.917, "encoder_q-layer.7": 12261.0166, "encoder_q-layer.8": 14612.3447, "encoder_q-layer.9": 12690.7207, "epoch": 0.82, "inbatch_neg_score": 5.4849, "inbatch_pos_score": 6.1094, "learning_rate": 2.7105263157894737e-06, "loss": 3.3387, "norm_diff": 0.0786, "norm_loss": 0.0, "num_token_doc": 66.6386, "num_token_overlap": 11.6781, "num_token_query": 31.3878, "num_token_union": 65.0646, "num_word_context": 202.221, "num_word_doc": 49.7084, "num_word_query": 23.3206, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22280.7731, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4844, "query_norm": 2.6339, "queue_k_norm": 2.7134, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3878, "sent_len_1": 66.6386, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.12, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 189700 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3411, "doc_norm": 2.713, "encoder_q-embeddings": 9534.6406, "encoder_q-layer.0": 6715.5732, "encoder_q-layer.1": 7246.6699, "encoder_q-layer.10": 15785.7842, "encoder_q-layer.11": 37611.8594, "encoder_q-layer.2": 8169.7021, "encoder_q-layer.3": 8453.501, "encoder_q-layer.4": 8930.3711, "encoder_q-layer.5": 8795.3955, "encoder_q-layer.6": 9953.3027, "encoder_q-layer.7": 11764.1885, "encoder_q-layer.8": 13811.6729, "encoder_q-layer.9": 13336.2949, "epoch": 0.82, "inbatch_neg_score": 5.4873, "inbatch_pos_score": 6.1016, "learning_rate": 2.68421052631579e-06, "loss": 3.3411, "norm_diff": 0.0782, "norm_loss": 0.0, "num_token_doc": 66.733, "num_token_overlap": 11.6777, "num_token_query": 31.3531, "num_token_union": 65.0911, "num_word_context": 202.1961, "num_word_doc": 49.7846, "num_word_query": 23.2885, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22871.9896, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4883, "query_norm": 2.6348, "queue_k_norm": 2.7138, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3531, "sent_len_1": 66.733, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.0575, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 189800 }, { "accuracy": 52.9297, "active_queue_size": 16384.0, "cl_loss": 3.3474, "doc_norm": 2.7156, "encoder_q-embeddings": 8997.2881, "encoder_q-layer.0": 6657.9478, "encoder_q-layer.1": 6981.7603, "encoder_q-layer.10": 13651.7002, "encoder_q-layer.11": 34260.3281, "encoder_q-layer.2": 7793.291, "encoder_q-layer.3": 8115.5439, "encoder_q-layer.4": 8833.7461, "encoder_q-layer.5": 9031.3115, "encoder_q-layer.6": 10659.582, "encoder_q-layer.7": 12285.625, "encoder_q-layer.8": 14001.5781, "encoder_q-layer.9": 12674.5508, "epoch": 0.82, "inbatch_neg_score": 5.4883, "inbatch_pos_score": 6.0977, "learning_rate": 2.6578947368421053e-06, "loss": 3.3474, "norm_diff": 0.0821, "norm_loss": 0.0, "num_token_doc": 66.6753, "num_token_overlap": 11.6243, "num_token_query": 31.3127, "num_token_union": 65.0515, "num_word_context": 202.1996, "num_word_doc": 49.7367, "num_word_query": 23.2518, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21497.0751, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4883, "query_norm": 2.6335, "queue_k_norm": 2.7131, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3127, "sent_len_1": 66.6753, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.765, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 189900 }, { "accuracy": 54.7363, "active_queue_size": 16384.0, "cl_loss": 3.3302, "doc_norm": 2.7153, "encoder_q-embeddings": 10255.8652, "encoder_q-layer.0": 7021.9463, "encoder_q-layer.1": 7668.9595, "encoder_q-layer.10": 14246.3799, "encoder_q-layer.11": 34652.6406, "encoder_q-layer.2": 8692.7871, "encoder_q-layer.3": 9007.4443, "encoder_q-layer.4": 9887.9736, "encoder_q-layer.5": 10083.1094, "encoder_q-layer.6": 11231.042, "encoder_q-layer.7": 12073.6631, "encoder_q-layer.8": 14056.6748, "encoder_q-layer.9": 12560.7139, "epoch": 0.82, "inbatch_neg_score": 5.4881, "inbatch_pos_score": 6.125, "learning_rate": 2.631578947368421e-06, "loss": 3.3302, "norm_diff": 0.0776, "norm_loss": 0.0, "num_token_doc": 66.7055, "num_token_overlap": 11.6623, "num_token_query": 31.4134, "num_token_union": 65.113, "num_word_context": 202.1604, "num_word_doc": 49.7555, "num_word_query": 23.3316, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21998.7156, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4883, "query_norm": 2.6377, "queue_k_norm": 2.7146, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4134, "sent_len_1": 66.7055, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.31, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 190000 }, { "dev_runtime": 28.2261, "dev_samples_per_second": 1.134, "dev_steps_per_second": 0.035, "epoch": 0.82, "step": 190000, "test_accuracy": 94.091796875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3210689425468445, "test_doc_norm": 2.6982016563415527, "test_inbatch_neg_score": 5.7754716873168945, "test_inbatch_pos_score": 6.80811882019043, "test_loss": 0.3210689425468445, "test_loss_align": 0.9824433326721191, "test_loss_unif": -53.98189926147461, "test_loss_unif_q@queue": -53.98189926147461, "test_norm_diff": 0.005644269287586212, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.480362415313721, "test_query_norm": 2.7036218643188477, "test_queue_k_norm": 2.7148678302764893, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04417052119970322, "test_stdq": 0.044199176132678986, "test_stdqueue_k": 0.04961264133453369, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.2261, "dev_samples_per_second": 1.134, "dev_steps_per_second": 0.035, "epoch": 0.82, "eval_beir-arguana_ndcg@10": 0.41181, "eval_beir-arguana_recall@10": 0.69772, "eval_beir-arguana_recall@100": 0.9495, "eval_beir-arguana_recall@20": 0.81792, "eval_beir-avg_ndcg@10": 0.4042180833333333, "eval_beir-avg_recall@10": 0.47702358333333333, "eval_beir-avg_recall@100": 0.65483125, "eval_beir-avg_recall@20": 0.5391926666666667, "eval_beir-cqadupstack_ndcg@10": 0.29196083333333334, "eval_beir-cqadupstack_recall@10": 0.39461583333333333, "eval_beir-cqadupstack_recall@100": 0.6326525, "eval_beir-cqadupstack_recall@20": 0.46394666666666673, "eval_beir-fiqa_ndcg@10": 0.27276, "eval_beir-fiqa_recall@10": 0.34428, "eval_beir-fiqa_recall@100": 0.61686, "eval_beir-fiqa_recall@20": 0.41352, "eval_beir-nfcorpus_ndcg@10": 0.32639, "eval_beir-nfcorpus_recall@10": 0.16016, "eval_beir-nfcorpus_recall@100": 0.31603, "eval_beir-nfcorpus_recall@20": 0.20419, "eval_beir-nq_ndcg@10": 0.32573, "eval_beir-nq_recall@10": 0.52141, "eval_beir-nq_recall@100": 0.8442, "eval_beir-nq_recall@20": 0.64371, "eval_beir-quora_ndcg@10": 0.79453, "eval_beir-quora_recall@10": 0.89856, "eval_beir-quora_recall@100": 0.98189, "eval_beir-quora_recall@20": 0.93713, "eval_beir-scidocs_ndcg@10": 0.16731, "eval_beir-scidocs_recall@10": 0.17493, "eval_beir-scidocs_recall@100": 0.3908, "eval_beir-scidocs_recall@20": 0.23492, "eval_beir-scifact_ndcg@10": 0.68411, "eval_beir-scifact_recall@10": 0.80067, "eval_beir-scifact_recall@100": 0.92989, "eval_beir-scifact_recall@20": 0.85078, "eval_beir-trec-covid_ndcg@10": 0.58894, "eval_beir-trec-covid_recall@10": 0.644, "eval_beir-trec-covid_recall@100": 0.4572, "eval_beir-trec-covid_recall@20": 0.608, "eval_beir-webis-touche2020_ndcg@10": 0.17864, "eval_beir-webis-touche2020_recall@10": 0.13389, "eval_beir-webis-touche2020_recall@100": 0.42929, "eval_beir-webis-touche2020_recall@20": 0.21781, "eval_senteval-avg_sts": 0.7329381103218209, "eval_senteval-sickr_spearman": 0.7258123351306504, "eval_senteval-stsb_spearman": 0.7400638855129914, "step": 190000, "test_accuracy": 94.091796875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3210689425468445, "test_doc_norm": 2.6982016563415527, "test_inbatch_neg_score": 5.7754716873168945, "test_inbatch_pos_score": 6.80811882019043, "test_loss": 0.3210689425468445, "test_loss_align": 0.9824433326721191, "test_loss_unif": -53.98189926147461, "test_loss_unif_q@queue": -53.98189926147461, "test_norm_diff": 0.005644269287586212, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.480362415313721, "test_query_norm": 2.7036218643188477, "test_queue_k_norm": 2.7148678302764893, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04417052119970322, "test_stdq": 0.044199176132678986, "test_stdqueue_k": 0.04961264133453369, "test_stdqueue_q": 0.0 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.341, "doc_norm": 2.7159, "encoder_q-embeddings": 9684.5977, "encoder_q-layer.0": 6661.5405, "encoder_q-layer.1": 6979.9551, "encoder_q-layer.10": 14059.6279, "encoder_q-layer.11": 34330.1641, "encoder_q-layer.2": 7841.5815, "encoder_q-layer.3": 7820.4023, "encoder_q-layer.4": 8475.375, "encoder_q-layer.5": 9121.9492, "encoder_q-layer.6": 10001.1289, "encoder_q-layer.7": 11291.5244, "encoder_q-layer.8": 14059.0703, "encoder_q-layer.9": 12657.291, "epoch": 0.82, "inbatch_neg_score": 5.4895, "inbatch_pos_score": 6.1094, "learning_rate": 2.605263157894737e-06, "loss": 3.341, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.7109, "num_token_overlap": 11.6869, "num_token_query": 31.3513, "num_token_union": 65.072, "num_word_context": 202.3012, "num_word_doc": 49.8107, "num_word_query": 23.2824, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21529.2966, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4883, "query_norm": 2.6357, "queue_k_norm": 2.7139, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3513, "sent_len_1": 66.7109, "sent_len_max_0": 127.9825, "sent_len_max_1": 207.4812, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 190100 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3105, "doc_norm": 2.7161, "encoder_q-embeddings": 9439.8037, "encoder_q-layer.0": 6923.166, "encoder_q-layer.1": 7486.2256, "encoder_q-layer.10": 13306.335, "encoder_q-layer.11": 34285.6406, "encoder_q-layer.2": 8382.6035, "encoder_q-layer.3": 8397.6201, "encoder_q-layer.4": 8721.9238, "encoder_q-layer.5": 9147.6855, "encoder_q-layer.6": 10306.2148, "encoder_q-layer.7": 11466.7832, "encoder_q-layer.8": 13575.1221, "encoder_q-layer.9": 13136.2021, "epoch": 0.83, "inbatch_neg_score": 5.4901, "inbatch_pos_score": 6.1094, "learning_rate": 2.5789473684210527e-06, "loss": 3.3105, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.8535, "num_token_overlap": 11.7245, "num_token_query": 31.4316, "num_token_union": 65.1639, "num_word_context": 202.5472, "num_word_doc": 49.8956, "num_word_query": 23.3505, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21505.6208, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4883, "query_norm": 2.6375, "queue_k_norm": 2.7146, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4316, "sent_len_1": 66.8535, "sent_len_max_0": 127.9887, "sent_len_max_1": 207.5938, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 190200 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3308, "doc_norm": 2.7145, "encoder_q-embeddings": 9677.3965, "encoder_q-layer.0": 6610.8706, "encoder_q-layer.1": 7268.9995, "encoder_q-layer.10": 15020.835, "encoder_q-layer.11": 35387.0977, "encoder_q-layer.2": 7889.6938, "encoder_q-layer.3": 8152.6699, "encoder_q-layer.4": 8536.7324, "encoder_q-layer.5": 9339.6279, "encoder_q-layer.6": 10105.7861, "encoder_q-layer.7": 10775.5303, "encoder_q-layer.8": 13645.707, "encoder_q-layer.9": 12714.9131, "epoch": 0.83, "inbatch_neg_score": 5.4909, "inbatch_pos_score": 6.1016, "learning_rate": 2.5526315789473686e-06, "loss": 3.3308, "norm_diff": 0.0815, "norm_loss": 0.0, "num_token_doc": 66.8182, "num_token_overlap": 11.6903, "num_token_query": 31.4611, "num_token_union": 65.1702, "num_word_context": 202.6474, "num_word_doc": 49.8715, "num_word_query": 23.3723, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21752.5827, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.633, "queue_k_norm": 2.7144, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4611, "sent_len_1": 66.8182, "sent_len_max_0": 127.9775, "sent_len_max_1": 208.4762, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 190300 }, { "accuracy": 51.0254, "active_queue_size": 16384.0, "cl_loss": 3.3581, "doc_norm": 2.7137, "encoder_q-embeddings": 10310.834, "encoder_q-layer.0": 7239.229, "encoder_q-layer.1": 7831.4863, "encoder_q-layer.10": 16448.1035, "encoder_q-layer.11": 37801.1641, "encoder_q-layer.2": 8783.4736, "encoder_q-layer.3": 9485.7295, "encoder_q-layer.4": 10038.2295, "encoder_q-layer.5": 11571.9434, "encoder_q-layer.6": 12972.1025, "encoder_q-layer.7": 13240.0703, "encoder_q-layer.8": 15614.8184, "encoder_q-layer.9": 13784.7559, "epoch": 0.83, "inbatch_neg_score": 5.4912, "inbatch_pos_score": 6.1016, "learning_rate": 2.5263157894736844e-06, "loss": 3.3581, "norm_diff": 0.0792, "norm_loss": 0.0, "num_token_doc": 66.5535, "num_token_overlap": 11.6196, "num_token_query": 31.3087, "num_token_union": 64.993, "num_word_context": 201.8434, "num_word_doc": 49.664, "num_word_query": 23.2392, "postclip_grad_norm": 1.0, "preclip_grad_norm": 24091.2321, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6345, "queue_k_norm": 2.714, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3087, "sent_len_1": 66.5535, "sent_len_max_0": 127.9638, "sent_len_max_1": 208.5975, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 190400 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.3352, "doc_norm": 2.7153, "encoder_q-embeddings": 9487.458, "encoder_q-layer.0": 6705.2915, "encoder_q-layer.1": 7152.8174, "encoder_q-layer.10": 14570.168, "encoder_q-layer.11": 35518.8438, "encoder_q-layer.2": 8026.7153, "encoder_q-layer.3": 8271.3369, "encoder_q-layer.4": 8983.4229, "encoder_q-layer.5": 8996.9902, "encoder_q-layer.6": 10248.7676, "encoder_q-layer.7": 11854.7754, "encoder_q-layer.8": 13999.3086, "encoder_q-layer.9": 13388.3379, "epoch": 0.83, "inbatch_neg_score": 5.4926, "inbatch_pos_score": 6.1094, "learning_rate": 2.5e-06, "loss": 3.3352, "norm_diff": 0.0839, "norm_loss": 0.0, "num_token_doc": 66.7995, "num_token_overlap": 11.6718, "num_token_query": 31.3915, "num_token_union": 65.1449, "num_word_context": 202.335, "num_word_doc": 49.8316, "num_word_query": 23.3082, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22273.8824, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6313, "queue_k_norm": 2.7151, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3915, "sent_len_1": 66.7995, "sent_len_max_0": 127.98, "sent_len_max_1": 209.6587, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 190500 }, { "accuracy": 51.5137, "active_queue_size": 16384.0, "cl_loss": 3.3323, "doc_norm": 2.7124, "encoder_q-embeddings": 9321.8564, "encoder_q-layer.0": 6813.3867, "encoder_q-layer.1": 7234.9995, "encoder_q-layer.10": 15055.2539, "encoder_q-layer.11": 37205.9648, "encoder_q-layer.2": 8125.2607, "encoder_q-layer.3": 8307.3066, "encoder_q-layer.4": 8968.0029, "encoder_q-layer.5": 9610.4834, "encoder_q-layer.6": 10477.7705, "encoder_q-layer.7": 11826.6055, "encoder_q-layer.8": 13733.9141, "encoder_q-layer.9": 12968.334, "epoch": 0.83, "inbatch_neg_score": 5.4918, "inbatch_pos_score": 6.1016, "learning_rate": 2.473684210526316e-06, "loss": 3.3323, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.9076, "num_token_overlap": 11.6495, "num_token_query": 31.3167, "num_token_union": 65.1861, "num_word_context": 202.3289, "num_word_doc": 49.9047, "num_word_query": 23.2318, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23036.602, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.634, "queue_k_norm": 2.7149, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3167, "sent_len_1": 66.9076, "sent_len_max_0": 127.9838, "sent_len_max_1": 211.1637, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 190600 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3345, "doc_norm": 2.7147, "encoder_q-embeddings": 9106.8203, "encoder_q-layer.0": 6604.6797, "encoder_q-layer.1": 7037.145, "encoder_q-layer.10": 16116.1816, "encoder_q-layer.11": 36883.1289, "encoder_q-layer.2": 7829.2227, "encoder_q-layer.3": 8000.5635, "encoder_q-layer.4": 8522.4473, "encoder_q-layer.5": 8956.166, "encoder_q-layer.6": 10199.6045, "encoder_q-layer.7": 11329.9824, "encoder_q-layer.8": 13232.293, "encoder_q-layer.9": 12713.5879, "epoch": 0.83, "inbatch_neg_score": 5.4934, "inbatch_pos_score": 6.1016, "learning_rate": 2.4473684210526314e-06, "loss": 3.3345, "norm_diff": 0.0816, "norm_loss": 0.0, "num_token_doc": 66.6909, "num_token_overlap": 11.6514, "num_token_query": 31.3669, "num_token_union": 65.1037, "num_word_context": 202.1202, "num_word_doc": 49.7607, "num_word_query": 23.2996, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22593.8674, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6332, "queue_k_norm": 2.7157, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3669, "sent_len_1": 66.6909, "sent_len_max_0": 127.9963, "sent_len_max_1": 209.2438, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 190700 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3277, "doc_norm": 2.7157, "encoder_q-embeddings": 9552.3408, "encoder_q-layer.0": 7104.311, "encoder_q-layer.1": 7750.6655, "encoder_q-layer.10": 12999.4023, "encoder_q-layer.11": 34036.8359, "encoder_q-layer.2": 8825.6553, "encoder_q-layer.3": 8938.9111, "encoder_q-layer.4": 9460.5869, "encoder_q-layer.5": 9644.8242, "encoder_q-layer.6": 11040.8018, "encoder_q-layer.7": 12802.082, "encoder_q-layer.8": 14143.7852, "encoder_q-layer.9": 12744.0938, "epoch": 0.83, "inbatch_neg_score": 5.4951, "inbatch_pos_score": 6.1172, "learning_rate": 2.4210526315789477e-06, "loss": 3.3277, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.8219, "num_token_overlap": 11.7172, "num_token_query": 31.4032, "num_token_union": 65.1468, "num_word_context": 202.3989, "num_word_doc": 49.8355, "num_word_query": 23.3258, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21664.8303, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6372, "queue_k_norm": 2.715, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4032, "sent_len_1": 66.8219, "sent_len_max_0": 127.9712, "sent_len_max_1": 209.095, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 190800 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3226, "doc_norm": 2.7166, "encoder_q-embeddings": 9436.6465, "encoder_q-layer.0": 6871.8335, "encoder_q-layer.1": 7760.7017, "encoder_q-layer.10": 13755.5244, "encoder_q-layer.11": 35379.3984, "encoder_q-layer.2": 8558.2559, "encoder_q-layer.3": 8668.3701, "encoder_q-layer.4": 9168.9102, "encoder_q-layer.5": 9658.8418, "encoder_q-layer.6": 10348.5713, "encoder_q-layer.7": 11868.1865, "encoder_q-layer.8": 14684.0977, "encoder_q-layer.9": 13526.291, "epoch": 0.83, "inbatch_neg_score": 5.4938, "inbatch_pos_score": 6.1094, "learning_rate": 2.394736842105263e-06, "loss": 3.3226, "norm_diff": 0.0789, "norm_loss": 0.0, "num_token_doc": 66.8765, "num_token_overlap": 11.6699, "num_token_query": 31.356, "num_token_union": 65.1886, "num_word_context": 202.4891, "num_word_doc": 49.9093, "num_word_query": 23.2954, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22519.6891, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4961, "query_norm": 2.6377, "queue_k_norm": 2.7165, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.356, "sent_len_1": 66.8765, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.9988, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 190900 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3425, "doc_norm": 2.7184, "encoder_q-embeddings": 9852.4004, "encoder_q-layer.0": 7178.646, "encoder_q-layer.1": 7784.6274, "encoder_q-layer.10": 13620.2061, "encoder_q-layer.11": 34584.8281, "encoder_q-layer.2": 8838.6455, "encoder_q-layer.3": 9212.043, "encoder_q-layer.4": 10245.3066, "encoder_q-layer.5": 10535.1758, "encoder_q-layer.6": 11703.0068, "encoder_q-layer.7": 13018.7617, "encoder_q-layer.8": 14756.9795, "encoder_q-layer.9": 13112.8203, "epoch": 0.83, "inbatch_neg_score": 5.4931, "inbatch_pos_score": 6.1016, "learning_rate": 2.3684210526315793e-06, "loss": 3.3425, "norm_diff": 0.082, "norm_loss": 0.0, "num_token_doc": 66.8252, "num_token_overlap": 11.6527, "num_token_query": 31.3272, "num_token_union": 65.1669, "num_word_context": 202.3086, "num_word_doc": 49.8558, "num_word_query": 23.2629, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22482.1611, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6365, "queue_k_norm": 2.7156, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3272, "sent_len_1": 66.8252, "sent_len_max_0": 127.9838, "sent_len_max_1": 209.0062, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 191000 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3332, "doc_norm": 2.7138, "encoder_q-embeddings": 8962.8965, "encoder_q-layer.0": 6423.2134, "encoder_q-layer.1": 6835.9238, "encoder_q-layer.10": 14546.082, "encoder_q-layer.11": 34930.1016, "encoder_q-layer.2": 7768.9731, "encoder_q-layer.3": 7932.3442, "encoder_q-layer.4": 8591.9893, "encoder_q-layer.5": 9025.0518, "encoder_q-layer.6": 10279.6738, "encoder_q-layer.7": 11753.1729, "encoder_q-layer.8": 13345.084, "encoder_q-layer.9": 12509.2881, "epoch": 0.83, "inbatch_neg_score": 5.4963, "inbatch_pos_score": 6.1094, "learning_rate": 2.3421052631578947e-06, "loss": 3.3332, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.7065, "num_token_overlap": 11.6629, "num_token_query": 31.4261, "num_token_union": 65.1347, "num_word_context": 202.2149, "num_word_doc": 49.7875, "num_word_query": 23.3455, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21774.8469, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4961, "query_norm": 2.6336, "queue_k_norm": 2.7161, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4261, "sent_len_1": 66.7065, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.7012, "stdk": 0.0493, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 191100 }, { "accuracy": 52.3438, "active_queue_size": 16384.0, "cl_loss": 3.3357, "doc_norm": 2.718, "encoder_q-embeddings": 11184.0977, "encoder_q-layer.0": 8334.4785, "encoder_q-layer.1": 9327.3779, "encoder_q-layer.10": 13987.1816, "encoder_q-layer.11": 34913.6016, "encoder_q-layer.2": 10803.998, "encoder_q-layer.3": 11371.3477, "encoder_q-layer.4": 12010.3213, "encoder_q-layer.5": 12200.8672, "encoder_q-layer.6": 12199.7754, "encoder_q-layer.7": 12847.9727, "encoder_q-layer.8": 14359.4336, "encoder_q-layer.9": 13016.916, "epoch": 0.83, "inbatch_neg_score": 5.4943, "inbatch_pos_score": 6.1172, "learning_rate": 2.315789473684211e-06, "loss": 3.3357, "norm_diff": 0.0834, "norm_loss": 0.0, "num_token_doc": 66.8522, "num_token_overlap": 11.7213, "num_token_query": 31.5191, "num_token_union": 65.2459, "num_word_context": 202.4722, "num_word_doc": 49.9094, "num_word_query": 23.4127, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23383.146, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4922, "query_norm": 2.6346, "queue_k_norm": 2.7154, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.5191, "sent_len_1": 66.8522, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.1325, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 191200 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3281, "doc_norm": 2.717, "encoder_q-embeddings": 10013.4648, "encoder_q-layer.0": 7170.0591, "encoder_q-layer.1": 8031.7998, "encoder_q-layer.10": 14402.0156, "encoder_q-layer.11": 36841.0664, "encoder_q-layer.2": 9036.8018, "encoder_q-layer.3": 9277.207, "encoder_q-layer.4": 9777.0781, "encoder_q-layer.5": 9699.5674, "encoder_q-layer.6": 10391.6777, "encoder_q-layer.7": 11579.3369, "encoder_q-layer.8": 13952.2773, "encoder_q-layer.9": 12913.8779, "epoch": 0.83, "inbatch_neg_score": 5.4986, "inbatch_pos_score": 6.1055, "learning_rate": 2.2894736842105263e-06, "loss": 3.3281, "norm_diff": 0.0821, "norm_loss": 0.0, "num_token_doc": 66.9552, "num_token_overlap": 11.7136, "num_token_query": 31.4736, "num_token_union": 65.2528, "num_word_context": 202.602, "num_word_doc": 49.9743, "num_word_query": 23.3827, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23001.8191, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5, "query_norm": 2.635, "queue_k_norm": 2.7162, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4736, "sent_len_1": 66.9552, "sent_len_max_0": 128.0, "sent_len_max_1": 210.5613, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 191300 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3276, "doc_norm": 2.717, "encoder_q-embeddings": 9681.9648, "encoder_q-layer.0": 7121.5815, "encoder_q-layer.1": 7772.2378, "encoder_q-layer.10": 13291.582, "encoder_q-layer.11": 35142.2734, "encoder_q-layer.2": 8776.5596, "encoder_q-layer.3": 9176.0898, "encoder_q-layer.4": 10006.2461, "encoder_q-layer.5": 10719.8662, "encoder_q-layer.6": 11322.5215, "encoder_q-layer.7": 12364.042, "encoder_q-layer.8": 14011.1426, "encoder_q-layer.9": 12538.4121, "epoch": 0.83, "inbatch_neg_score": 5.4977, "inbatch_pos_score": 6.1094, "learning_rate": 2.2631578947368426e-06, "loss": 3.3276, "norm_diff": 0.0836, "norm_loss": 0.0, "num_token_doc": 66.7256, "num_token_overlap": 11.6647, "num_token_query": 31.379, "num_token_union": 65.1116, "num_word_context": 202.3394, "num_word_doc": 49.7699, "num_word_query": 23.2924, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22346.8075, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.4961, "query_norm": 2.6334, "queue_k_norm": 2.7167, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.379, "sent_len_1": 66.7256, "sent_len_max_0": 127.9813, "sent_len_max_1": 211.9075, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 191400 }, { "accuracy": 52.832, "active_queue_size": 16384.0, "cl_loss": 3.3302, "doc_norm": 2.7152, "encoder_q-embeddings": 9333.2939, "encoder_q-layer.0": 6660.5249, "encoder_q-layer.1": 7314.7056, "encoder_q-layer.10": 13209.668, "encoder_q-layer.11": 34685.5391, "encoder_q-layer.2": 8136.7734, "encoder_q-layer.3": 8345.6914, "encoder_q-layer.4": 8788.3076, "encoder_q-layer.5": 8746.2041, "encoder_q-layer.6": 10067.2227, "encoder_q-layer.7": 11002.9473, "encoder_q-layer.8": 13013.0225, "encoder_q-layer.9": 12183.7715, "epoch": 0.83, "inbatch_neg_score": 5.4972, "inbatch_pos_score": 6.1172, "learning_rate": 2.236842105263158e-06, "loss": 3.3302, "norm_diff": 0.079, "norm_loss": 0.0, "num_token_doc": 66.7956, "num_token_overlap": 11.6889, "num_token_query": 31.4357, "num_token_union": 65.1659, "num_word_context": 202.6809, "num_word_doc": 49.8627, "num_word_query": 23.3461, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21619.382, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5, "query_norm": 2.6362, "queue_k_norm": 2.7165, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4357, "sent_len_1": 66.7956, "sent_len_max_0": 127.9975, "sent_len_max_1": 207.1538, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 191500 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3534, "doc_norm": 2.717, "encoder_q-embeddings": 18731.3027, "encoder_q-layer.0": 13014.8574, "encoder_q-layer.1": 13654.8818, "encoder_q-layer.10": 28395.7207, "encoder_q-layer.11": 68395.3672, "encoder_q-layer.2": 15390.3906, "encoder_q-layer.3": 15937.0947, "encoder_q-layer.4": 17153.3438, "encoder_q-layer.5": 17141.3027, "encoder_q-layer.6": 19297.9453, "encoder_q-layer.7": 22171.8574, "encoder_q-layer.8": 26816.0762, "encoder_q-layer.9": 24971.8398, "epoch": 0.83, "inbatch_neg_score": 5.5, "inbatch_pos_score": 6.1172, "learning_rate": 2.2105263157894738e-06, "loss": 3.3534, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.6073, "num_token_overlap": 11.6412, "num_token_query": 31.3056, "num_token_union": 65.0017, "num_word_context": 202.1074, "num_word_doc": 49.6871, "num_word_query": 23.2425, "postclip_grad_norm": 1.0, "preclip_grad_norm": 42415.8342, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.5, "query_norm": 2.6372, "queue_k_norm": 2.7167, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3056, "sent_len_1": 66.6073, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.5913, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 191600 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3319, "doc_norm": 2.7156, "encoder_q-embeddings": 18789.7871, "encoder_q-layer.0": 13456.7432, "encoder_q-layer.1": 14097.2617, "encoder_q-layer.10": 29756.0195, "encoder_q-layer.11": 70772.6328, "encoder_q-layer.2": 15899.3838, "encoder_q-layer.3": 15968.0166, "encoder_q-layer.4": 16470.4219, "encoder_q-layer.5": 17243.8691, "encoder_q-layer.6": 18424.8867, "encoder_q-layer.7": 20867.4902, "encoder_q-layer.8": 25767.7988, "encoder_q-layer.9": 24913.1895, "epoch": 0.83, "inbatch_neg_score": 5.5007, "inbatch_pos_score": 6.1172, "learning_rate": 2.1842105263157896e-06, "loss": 3.3319, "norm_diff": 0.0803, "norm_loss": 0.0, "num_token_doc": 66.8107, "num_token_overlap": 11.697, "num_token_query": 31.4515, "num_token_union": 65.17, "num_word_context": 202.0288, "num_word_doc": 49.827, "num_word_query": 23.3661, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43447.4472, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.5, "query_norm": 2.6353, "queue_k_norm": 2.717, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4515, "sent_len_1": 66.8107, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.4387, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 191700 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3471, "doc_norm": 2.7188, "encoder_q-embeddings": 19019.875, "encoder_q-layer.0": 13664.0693, "encoder_q-layer.1": 14356.0244, "encoder_q-layer.10": 25597.8438, "encoder_q-layer.11": 68988.3281, "encoder_q-layer.2": 15647.4287, "encoder_q-layer.3": 16395.0664, "encoder_q-layer.4": 17645.6484, "encoder_q-layer.5": 18113.6641, "encoder_q-layer.6": 19982.5762, "encoder_q-layer.7": 22507.2227, "encoder_q-layer.8": 26142.9023, "encoder_q-layer.9": 24700.2324, "epoch": 0.83, "inbatch_neg_score": 5.5008, "inbatch_pos_score": 6.1172, "learning_rate": 2.1578947368421054e-06, "loss": 3.3471, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.7728, "num_token_overlap": 11.6484, "num_token_query": 31.3284, "num_token_union": 65.1159, "num_word_context": 202.2113, "num_word_doc": 49.825, "num_word_query": 23.2547, "postclip_grad_norm": 1.0, "preclip_grad_norm": 43159.3462, "preclip_grad_norm_avg": 0.0004, "q@queue_neg_score": 5.5, "query_norm": 2.6362, "queue_k_norm": 2.7169, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3284, "sent_len_1": 66.7728, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.3325, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 191800 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3227, "doc_norm": 2.7186, "encoder_q-embeddings": 9319.4424, "encoder_q-layer.0": 6782.8916, "encoder_q-layer.1": 7211.2905, "encoder_q-layer.10": 13574.0967, "encoder_q-layer.11": 35576.6758, "encoder_q-layer.2": 7912.5078, "encoder_q-layer.3": 7919.0444, "encoder_q-layer.4": 8464.3096, "encoder_q-layer.5": 8660.1406, "encoder_q-layer.6": 10064.7881, "encoder_q-layer.7": 11303.8867, "encoder_q-layer.8": 13259.9277, "encoder_q-layer.9": 12462.1904, "epoch": 0.83, "inbatch_neg_score": 5.5007, "inbatch_pos_score": 6.1133, "learning_rate": 2.1315789473684212e-06, "loss": 3.3227, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.6316, "num_token_overlap": 11.6454, "num_token_query": 31.318, "num_token_union": 65.0507, "num_word_context": 202.3408, "num_word_doc": 49.7725, "num_word_query": 23.2727, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22092.172, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5, "query_norm": 2.6369, "queue_k_norm": 2.7174, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.318, "sent_len_1": 66.6316, "sent_len_max_0": 127.9488, "sent_len_max_1": 206.3512, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 191900 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3282, "doc_norm": 2.7175, "encoder_q-embeddings": 9768.3145, "encoder_q-layer.0": 6695.2686, "encoder_q-layer.1": 7313.5596, "encoder_q-layer.10": 13316.0869, "encoder_q-layer.11": 35361.8398, "encoder_q-layer.2": 7865.2744, "encoder_q-layer.3": 8324.2217, "encoder_q-layer.4": 8777.1475, "encoder_q-layer.5": 9217.2539, "encoder_q-layer.6": 10238.4121, "encoder_q-layer.7": 11857.4414, "encoder_q-layer.8": 14525.7227, "encoder_q-layer.9": 13052.0469, "epoch": 0.83, "inbatch_neg_score": 5.5031, "inbatch_pos_score": 6.1211, "learning_rate": 2.105263157894737e-06, "loss": 3.3282, "norm_diff": 0.0801, "norm_loss": 0.0, "num_token_doc": 66.7467, "num_token_overlap": 11.6563, "num_token_query": 31.3598, "num_token_union": 65.1237, "num_word_context": 202.254, "num_word_doc": 49.847, "num_word_query": 23.3092, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22206.4686, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5039, "query_norm": 2.6375, "queue_k_norm": 2.7173, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3598, "sent_len_1": 66.7467, "sent_len_max_0": 127.9887, "sent_len_max_1": 205.6562, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 192000 }, { "accuracy": 52.1484, "active_queue_size": 16384.0, "cl_loss": 3.3454, "doc_norm": 2.7183, "encoder_q-embeddings": 9365.042, "encoder_q-layer.0": 6725.8638, "encoder_q-layer.1": 7203.7061, "encoder_q-layer.10": 14671.1416, "encoder_q-layer.11": 37550.6953, "encoder_q-layer.2": 8142.0088, "encoder_q-layer.3": 8315.8135, "encoder_q-layer.4": 9183.9102, "encoder_q-layer.5": 9127.8281, "encoder_q-layer.6": 10426.7715, "encoder_q-layer.7": 12224.1299, "encoder_q-layer.8": 14555.7236, "encoder_q-layer.9": 13610.2402, "epoch": 0.83, "inbatch_neg_score": 5.5034, "inbatch_pos_score": 6.1172, "learning_rate": 2.078947368421053e-06, "loss": 3.3454, "norm_diff": 0.0802, "norm_loss": 0.0, "num_token_doc": 66.7175, "num_token_overlap": 11.6911, "num_token_query": 31.4231, "num_token_union": 65.1125, "num_word_context": 202.3983, "num_word_doc": 49.8296, "num_word_query": 23.355, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23106.2793, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5039, "query_norm": 2.6381, "queue_k_norm": 2.7175, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4231, "sent_len_1": 66.7175, "sent_len_max_0": 127.9775, "sent_len_max_1": 206.2188, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192100 }, { "accuracy": 51.3184, "active_queue_size": 16384.0, "cl_loss": 3.3474, "doc_norm": 2.7184, "encoder_q-embeddings": 10200.0273, "encoder_q-layer.0": 7164.9014, "encoder_q-layer.1": 8121.1436, "encoder_q-layer.10": 14153.4785, "encoder_q-layer.11": 37175.2539, "encoder_q-layer.2": 9439.1016, "encoder_q-layer.3": 9401.5049, "encoder_q-layer.4": 10199.7168, "encoder_q-layer.5": 10323.0889, "encoder_q-layer.6": 11626.7783, "encoder_q-layer.7": 13655.5625, "encoder_q-layer.8": 14921.7246, "encoder_q-layer.9": 13781.7646, "epoch": 0.83, "inbatch_neg_score": 5.5026, "inbatch_pos_score": 6.1133, "learning_rate": 2.0526315789473687e-06, "loss": 3.3474, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.857, "num_token_overlap": 11.6536, "num_token_query": 31.35, "num_token_union": 65.1534, "num_word_context": 202.5249, "num_word_doc": 49.8512, "num_word_query": 23.2736, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23707.661, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5039, "query_norm": 2.6338, "queue_k_norm": 2.7177, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.35, "sent_len_1": 66.857, "sent_len_max_0": 127.9762, "sent_len_max_1": 208.5425, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192200 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3284, "doc_norm": 2.7172, "encoder_q-embeddings": 9209.415, "encoder_q-layer.0": 6332.8638, "encoder_q-layer.1": 6914.314, "encoder_q-layer.10": 13370.2725, "encoder_q-layer.11": 34889.2031, "encoder_q-layer.2": 7947.1318, "encoder_q-layer.3": 8047.8506, "encoder_q-layer.4": 8712.2529, "encoder_q-layer.5": 9168.998, "encoder_q-layer.6": 10244.8301, "encoder_q-layer.7": 11649.1328, "encoder_q-layer.8": 13381.7432, "encoder_q-layer.9": 12769.416, "epoch": 0.83, "inbatch_neg_score": 5.5044, "inbatch_pos_score": 6.125, "learning_rate": 2.026315789473684e-06, "loss": 3.3284, "norm_diff": 0.08, "norm_loss": 0.0, "num_token_doc": 66.8463, "num_token_overlap": 11.6492, "num_token_query": 31.2917, "num_token_union": 65.1538, "num_word_context": 202.259, "num_word_doc": 49.8914, "num_word_query": 23.2467, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21548.8671, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5039, "query_norm": 2.6373, "queue_k_norm": 2.7181, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.2917, "sent_len_1": 66.8463, "sent_len_max_0": 127.9725, "sent_len_max_1": 207.1738, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192300 }, { "accuracy": 48.584, "active_queue_size": 16384.0, "cl_loss": 3.3312, "doc_norm": 2.7193, "encoder_q-embeddings": 10050.4766, "encoder_q-layer.0": 7116.7456, "encoder_q-layer.1": 7514.7344, "encoder_q-layer.10": 16622.0391, "encoder_q-layer.11": 37160.4492, "encoder_q-layer.2": 8347.6357, "encoder_q-layer.3": 8409.7725, "encoder_q-layer.4": 8885.0498, "encoder_q-layer.5": 9455.8691, "encoder_q-layer.6": 10602.5361, "encoder_q-layer.7": 11826.5967, "encoder_q-layer.8": 14916.0791, "encoder_q-layer.9": 13374.6934, "epoch": 0.83, "inbatch_neg_score": 5.5065, "inbatch_pos_score": 6.1055, "learning_rate": 2.0000000000000003e-06, "loss": 3.3312, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.72, "num_token_overlap": 11.6906, "num_token_query": 31.3811, "num_token_union": 65.0802, "num_word_context": 202.3741, "num_word_doc": 49.8201, "num_word_query": 23.3287, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23113.2881, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5078, "query_norm": 2.6378, "queue_k_norm": 2.7183, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3811, "sent_len_1": 66.72, "sent_len_max_0": 127.9875, "sent_len_max_1": 208.5888, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192400 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3339, "doc_norm": 2.7202, "encoder_q-embeddings": 10369.6631, "encoder_q-layer.0": 7696.3525, "encoder_q-layer.1": 8216.7666, "encoder_q-layer.10": 15243.6416, "encoder_q-layer.11": 35899.8281, "encoder_q-layer.2": 9572.835, "encoder_q-layer.3": 9635.4121, "encoder_q-layer.4": 10485.832, "encoder_q-layer.5": 10723.6826, "encoder_q-layer.6": 11863.915, "encoder_q-layer.7": 12638.8379, "encoder_q-layer.8": 14888.4883, "encoder_q-layer.9": 13288.0898, "epoch": 0.84, "inbatch_neg_score": 5.5077, "inbatch_pos_score": 6.1211, "learning_rate": 1.9736842105263157e-06, "loss": 3.3339, "norm_diff": 0.0813, "norm_loss": 0.0, "num_token_doc": 66.8591, "num_token_overlap": 11.6588, "num_token_query": 31.312, "num_token_union": 65.1434, "num_word_context": 202.4104, "num_word_doc": 49.8695, "num_word_query": 23.239, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23236.6954, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5078, "query_norm": 2.6389, "queue_k_norm": 2.7183, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.312, "sent_len_1": 66.8591, "sent_len_max_0": 127.9912, "sent_len_max_1": 210.625, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192500 }, { "accuracy": 53.8086, "active_queue_size": 16384.0, "cl_loss": 3.3447, "doc_norm": 2.7197, "encoder_q-embeddings": 9376.6152, "encoder_q-layer.0": 6403.1118, "encoder_q-layer.1": 6829.7407, "encoder_q-layer.10": 13692.9141, "encoder_q-layer.11": 34908.4375, "encoder_q-layer.2": 7414.8042, "encoder_q-layer.3": 7611.2388, "encoder_q-layer.4": 8161.7476, "encoder_q-layer.5": 8588.1611, "encoder_q-layer.6": 9528.2217, "encoder_q-layer.7": 10791.9434, "encoder_q-layer.8": 13262.5557, "encoder_q-layer.9": 12712.8057, "epoch": 0.84, "inbatch_neg_score": 5.5061, "inbatch_pos_score": 6.1328, "learning_rate": 1.947368421052632e-06, "loss": 3.3447, "norm_diff": 0.0806, "norm_loss": 0.0, "num_token_doc": 66.6751, "num_token_overlap": 11.6293, "num_token_query": 31.3064, "num_token_union": 65.0358, "num_word_context": 202.2643, "num_word_doc": 49.7624, "num_word_query": 23.2612, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21361.553, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5078, "query_norm": 2.6391, "queue_k_norm": 2.7184, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3064, "sent_len_1": 66.6751, "sent_len_max_0": 127.9963, "sent_len_max_1": 207.735, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192600 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.331, "doc_norm": 2.7189, "encoder_q-embeddings": 9545.3135, "encoder_q-layer.0": 6461.4336, "encoder_q-layer.1": 6838.8071, "encoder_q-layer.10": 15241.6367, "encoder_q-layer.11": 35288.9531, "encoder_q-layer.2": 7765.0039, "encoder_q-layer.3": 7976.4126, "encoder_q-layer.4": 8807.7012, "encoder_q-layer.5": 8911.5645, "encoder_q-layer.6": 10306.8828, "encoder_q-layer.7": 12542.9492, "encoder_q-layer.8": 14939.3057, "encoder_q-layer.9": 13255.8447, "epoch": 0.84, "inbatch_neg_score": 5.5087, "inbatch_pos_score": 6.1289, "learning_rate": 1.9210526315789474e-06, "loss": 3.331, "norm_diff": 0.0784, "norm_loss": 0.0, "num_token_doc": 66.8445, "num_token_overlap": 11.6797, "num_token_query": 31.3978, "num_token_union": 65.1831, "num_word_context": 202.329, "num_word_doc": 49.8893, "num_word_query": 23.3186, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22032.8656, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5078, "query_norm": 2.6405, "queue_k_norm": 2.7196, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3978, "sent_len_1": 66.8445, "sent_len_max_0": 127.9925, "sent_len_max_1": 208.7837, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 192700 }, { "accuracy": 52.2949, "active_queue_size": 16384.0, "cl_loss": 3.3466, "doc_norm": 2.7208, "encoder_q-embeddings": 9715.4375, "encoder_q-layer.0": 6599.5371, "encoder_q-layer.1": 7295.4355, "encoder_q-layer.10": 15621.5391, "encoder_q-layer.11": 35813.0195, "encoder_q-layer.2": 7912.1914, "encoder_q-layer.3": 7994.8701, "encoder_q-layer.4": 8831.5781, "encoder_q-layer.5": 8968.8652, "encoder_q-layer.6": 9669.2002, "encoder_q-layer.7": 11414.8857, "encoder_q-layer.8": 14006.7002, "encoder_q-layer.9": 12956.8154, "epoch": 0.84, "inbatch_neg_score": 5.5104, "inbatch_pos_score": 6.125, "learning_rate": 1.8947368421052634e-06, "loss": 3.3466, "norm_diff": 0.0812, "norm_loss": 0.0, "num_token_doc": 66.7351, "num_token_overlap": 11.6825, "num_token_query": 31.3534, "num_token_union": 65.1044, "num_word_context": 202.3111, "num_word_doc": 49.7239, "num_word_query": 23.2723, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22155.0797, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.6395, "queue_k_norm": 2.7187, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3534, "sent_len_1": 66.7351, "sent_len_max_0": 127.9938, "sent_len_max_1": 211.4038, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192800 }, { "accuracy": 52.1973, "active_queue_size": 16384.0, "cl_loss": 3.3205, "doc_norm": 2.722, "encoder_q-embeddings": 9943.3574, "encoder_q-layer.0": 6857.313, "encoder_q-layer.1": 7587.9614, "encoder_q-layer.10": 14308.624, "encoder_q-layer.11": 37271.2422, "encoder_q-layer.2": 8301.8525, "encoder_q-layer.3": 8340.2822, "encoder_q-layer.4": 9292.2832, "encoder_q-layer.5": 9382.5518, "encoder_q-layer.6": 10869.8408, "encoder_q-layer.7": 12033.0732, "encoder_q-layer.8": 14925.8789, "encoder_q-layer.9": 13522.208, "epoch": 0.84, "inbatch_neg_score": 5.5098, "inbatch_pos_score": 6.1328, "learning_rate": 1.868421052631579e-06, "loss": 3.3205, "norm_diff": 0.0817, "norm_loss": 0.0, "num_token_doc": 66.7746, "num_token_overlap": 11.7221, "num_token_query": 31.4851, "num_token_union": 65.1298, "num_word_context": 202.3663, "num_word_doc": 49.8049, "num_word_query": 23.3814, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23541.6808, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5078, "query_norm": 2.6403, "queue_k_norm": 2.7196, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4851, "sent_len_1": 66.7746, "sent_len_max_0": 127.98, "sent_len_max_1": 211.84, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 192900 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3523, "doc_norm": 2.7182, "encoder_q-embeddings": 11407.2627, "encoder_q-layer.0": 8152.0635, "encoder_q-layer.1": 8768.998, "encoder_q-layer.10": 15069.6504, "encoder_q-layer.11": 35107.1367, "encoder_q-layer.2": 10500.9941, "encoder_q-layer.3": 11337.833, "encoder_q-layer.4": 12413.1396, "encoder_q-layer.5": 12083.1914, "encoder_q-layer.6": 12855.4043, "encoder_q-layer.7": 12849.2393, "encoder_q-layer.8": 14665.6748, "encoder_q-layer.9": 12460.4131, "epoch": 0.84, "inbatch_neg_score": 5.5119, "inbatch_pos_score": 6.125, "learning_rate": 1.8421052631578946e-06, "loss": 3.3523, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.6258, "num_token_overlap": 11.6291, "num_token_query": 31.3346, "num_token_union": 65.0575, "num_word_context": 202.0777, "num_word_doc": 49.72, "num_word_query": 23.2714, "postclip_grad_norm": 1.0, "preclip_grad_norm": 23664.9116, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.6345, "queue_k_norm": 2.7179, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3346, "sent_len_1": 66.6258, "sent_len_max_0": 127.9788, "sent_len_max_1": 209.7, "stdk": 0.0495, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 193000 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3376, "doc_norm": 2.72, "encoder_q-embeddings": 9228.2607, "encoder_q-layer.0": 6793.0674, "encoder_q-layer.1": 7196.7705, "encoder_q-layer.10": 13259.6221, "encoder_q-layer.11": 34199.1836, "encoder_q-layer.2": 8061.0215, "encoder_q-layer.3": 7977.0986, "encoder_q-layer.4": 8331.6904, "encoder_q-layer.5": 8487.9082, "encoder_q-layer.6": 9512.1494, "encoder_q-layer.7": 11607.3643, "encoder_q-layer.8": 12996.4746, "encoder_q-layer.9": 12332.8291, "epoch": 0.84, "inbatch_neg_score": 5.5105, "inbatch_pos_score": 6.1289, "learning_rate": 1.8157894736842106e-06, "loss": 3.3376, "norm_diff": 0.0806, "norm_loss": 0.0, "num_token_doc": 66.664, "num_token_overlap": 11.6824, "num_token_query": 31.4413, "num_token_union": 65.0815, "num_word_context": 202.4487, "num_word_doc": 49.7531, "num_word_query": 23.358, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21496.348, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.6394, "queue_k_norm": 2.7193, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4413, "sent_len_1": 66.664, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.7, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193100 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.34, "doc_norm": 2.7183, "encoder_q-embeddings": 9253.6182, "encoder_q-layer.0": 6414.8545, "encoder_q-layer.1": 6967.2012, "encoder_q-layer.10": 14116.4053, "encoder_q-layer.11": 35057.3633, "encoder_q-layer.2": 7793.6162, "encoder_q-layer.3": 7917.0552, "encoder_q-layer.4": 8640.3779, "encoder_q-layer.5": 8620.4375, "encoder_q-layer.6": 9659.3408, "encoder_q-layer.7": 11090.1787, "encoder_q-layer.8": 13471.6611, "encoder_q-layer.9": 13225.0684, "epoch": 0.84, "inbatch_neg_score": 5.5127, "inbatch_pos_score": 6.1172, "learning_rate": 1.7894736842105262e-06, "loss": 3.34, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.7167, "num_token_overlap": 11.6518, "num_token_query": 31.3361, "num_token_union": 65.0675, "num_word_context": 202.1471, "num_word_doc": 49.7606, "num_word_query": 23.2539, "postclip_grad_norm": 1.0, "preclip_grad_norm": 21802.0821, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.6354, "queue_k_norm": 2.7205, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3361, "sent_len_1": 66.7167, "sent_len_max_0": 127.995, "sent_len_max_1": 208.6413, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 193200 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3416, "doc_norm": 2.7216, "encoder_q-embeddings": 9997.9756, "encoder_q-layer.0": 6911.1514, "encoder_q-layer.1": 7505.3755, "encoder_q-layer.10": 15216.1338, "encoder_q-layer.11": 36665.3281, "encoder_q-layer.2": 8371.8799, "encoder_q-layer.3": 8476.0918, "encoder_q-layer.4": 9026.5957, "encoder_q-layer.5": 9470.3408, "encoder_q-layer.6": 10423.7012, "encoder_q-layer.7": 11708.3799, "encoder_q-layer.8": 14118.2129, "encoder_q-layer.9": 13448.4355, "epoch": 0.84, "inbatch_neg_score": 5.5108, "inbatch_pos_score": 6.1289, "learning_rate": 1.7631578947368423e-06, "loss": 3.3416, "norm_diff": 0.0835, "norm_loss": 0.0, "num_token_doc": 66.7587, "num_token_overlap": 11.6789, "num_token_query": 31.343, "num_token_union": 65.0637, "num_word_context": 202.0433, "num_word_doc": 49.8265, "num_word_query": 23.2685, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22987.5094, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.6381, "queue_k_norm": 2.72, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.343, "sent_len_1": 66.7587, "sent_len_max_0": 127.985, "sent_len_max_1": 211.2325, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193300 }, { "accuracy": 51.5625, "active_queue_size": 16384.0, "cl_loss": 3.3323, "doc_norm": 2.7216, "encoder_q-embeddings": 15450.6133, "encoder_q-layer.0": 11555.9424, "encoder_q-layer.1": 13961.1016, "encoder_q-layer.10": 13736.0107, "encoder_q-layer.11": 35836.7812, "encoder_q-layer.2": 17764.2676, "encoder_q-layer.3": 18597.666, "encoder_q-layer.4": 18555.6426, "encoder_q-layer.5": 19783.3379, "encoder_q-layer.6": 18671.7129, "encoder_q-layer.7": 17717.0781, "encoder_q-layer.8": 16182.2402, "encoder_q-layer.9": 13387.2793, "epoch": 0.84, "inbatch_neg_score": 5.5123, "inbatch_pos_score": 6.1328, "learning_rate": 1.7368421052631579e-06, "loss": 3.3323, "norm_diff": 0.0816, "norm_loss": 0.0, "num_token_doc": 66.8716, "num_token_overlap": 11.6727, "num_token_query": 31.3644, "num_token_union": 65.1513, "num_word_context": 202.4732, "num_word_doc": 49.8893, "num_word_query": 23.3027, "postclip_grad_norm": 1.0, "preclip_grad_norm": 28989.4782, "preclip_grad_norm_avg": 0.0003, "q@queue_neg_score": 5.5117, "query_norm": 2.64, "queue_k_norm": 2.7195, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3644, "sent_len_1": 66.8716, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.51, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193400 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3407, "doc_norm": 2.7178, "encoder_q-embeddings": 9743.7256, "encoder_q-layer.0": 7008.5391, "encoder_q-layer.1": 7399.0728, "encoder_q-layer.10": 14867.7471, "encoder_q-layer.11": 35106.2656, "encoder_q-layer.2": 8464.1094, "encoder_q-layer.3": 8579.9873, "encoder_q-layer.4": 9420.8203, "encoder_q-layer.5": 9592.8486, "encoder_q-layer.6": 10353.4258, "encoder_q-layer.7": 11574.6514, "encoder_q-layer.8": 13552.1807, "encoder_q-layer.9": 13056.0488, "epoch": 0.84, "inbatch_neg_score": 5.5127, "inbatch_pos_score": 6.1172, "learning_rate": 1.710526315789474e-06, "loss": 3.3407, "norm_diff": 0.0808, "norm_loss": 0.0, "num_token_doc": 66.7371, "num_token_overlap": 11.6392, "num_token_query": 31.3169, "num_token_union": 65.0955, "num_word_context": 202.2637, "num_word_doc": 49.7901, "num_word_query": 23.2439, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22048.8117, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5117, "query_norm": 2.637, "queue_k_norm": 2.72, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3169, "sent_len_1": 66.7371, "sent_len_max_0": 127.99, "sent_len_max_1": 208.1238, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193500 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.339, "doc_norm": 2.7203, "encoder_q-embeddings": 9802.1963, "encoder_q-layer.0": 6966.293, "encoder_q-layer.1": 7504.4253, "encoder_q-layer.10": 15399.584, "encoder_q-layer.11": 35335.8047, "encoder_q-layer.2": 8424.7305, "encoder_q-layer.3": 8822.457, "encoder_q-layer.4": 9185.7061, "encoder_q-layer.5": 9527.1562, "encoder_q-layer.6": 10956.8574, "encoder_q-layer.7": 12241.4336, "encoder_q-layer.8": 13535.7598, "encoder_q-layer.9": 13091.3438, "epoch": 0.84, "inbatch_neg_score": 5.5154, "inbatch_pos_score": 6.1172, "learning_rate": 1.6842105263157895e-06, "loss": 3.339, "norm_diff": 0.0835, "norm_loss": 0.0, "num_token_doc": 66.7661, "num_token_overlap": 11.6759, "num_token_query": 31.4387, "num_token_union": 65.1704, "num_word_context": 202.1072, "num_word_doc": 49.8229, "num_word_query": 23.3648, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22424.11, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5156, "query_norm": 2.6368, "queue_k_norm": 2.7195, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4387, "sent_len_1": 66.7661, "sent_len_max_0": 127.98, "sent_len_max_1": 208.5613, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193600 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3352, "doc_norm": 2.7195, "encoder_q-embeddings": 10363.4961, "encoder_q-layer.0": 7311.2715, "encoder_q-layer.1": 8016.9331, "encoder_q-layer.10": 15514.7207, "encoder_q-layer.11": 36382.2734, "encoder_q-layer.2": 8647.9922, "encoder_q-layer.3": 8778.2021, "encoder_q-layer.4": 9296.4238, "encoder_q-layer.5": 9551.1602, "encoder_q-layer.6": 10396.4463, "encoder_q-layer.7": 11278.4102, "encoder_q-layer.8": 14294.1328, "encoder_q-layer.9": 13424.8496, "epoch": 0.84, "inbatch_neg_score": 5.5148, "inbatch_pos_score": 6.1289, "learning_rate": 1.6578947368421056e-06, "loss": 3.3352, "norm_diff": 0.0794, "norm_loss": 0.0, "num_token_doc": 66.6543, "num_token_overlap": 11.6497, "num_token_query": 31.3504, "num_token_union": 65.0414, "num_word_context": 202.152, "num_word_doc": 49.7354, "num_word_query": 23.2938, "postclip_grad_norm": 1.0, "preclip_grad_norm": 22911.0897, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5156, "query_norm": 2.6401, "queue_k_norm": 2.7193, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3504, "sent_len_1": 66.6543, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.8113, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 193700 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3347, "doc_norm": 2.7182, "encoder_q-embeddings": 4903.9736, "encoder_q-layer.0": 3513.6079, "encoder_q-layer.1": 3751.5205, "encoder_q-layer.10": 6618.3662, "encoder_q-layer.11": 17887.0957, "encoder_q-layer.2": 4356.0649, "encoder_q-layer.3": 4549.5576, "encoder_q-layer.4": 4802.2285, "encoder_q-layer.5": 4891.2407, "encoder_q-layer.6": 5543.001, "encoder_q-layer.7": 6276.2603, "encoder_q-layer.8": 7174.811, "encoder_q-layer.9": 6631.3125, "epoch": 0.84, "inbatch_neg_score": 5.5158, "inbatch_pos_score": 6.1289, "learning_rate": 1.6315789473684212e-06, "loss": 3.3347, "norm_diff": 0.0819, "norm_loss": 0.0, "num_token_doc": 66.7011, "num_token_overlap": 11.6554, "num_token_query": 31.3467, "num_token_union": 65.09, "num_word_context": 202.3054, "num_word_doc": 49.7751, "num_word_query": 23.2733, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11399.4778, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5156, "query_norm": 2.6363, "queue_k_norm": 2.72, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3467, "sent_len_1": 66.7011, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.4825, "stdk": 0.0494, "stdq": 0.0429, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193800 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3434, "doc_norm": 2.7215, "encoder_q-embeddings": 4741.2451, "encoder_q-layer.0": 3343.9871, "encoder_q-layer.1": 3513.2007, "encoder_q-layer.10": 6960.7783, "encoder_q-layer.11": 17518.7715, "encoder_q-layer.2": 3959.4246, "encoder_q-layer.3": 3943.7869, "encoder_q-layer.4": 4321.3481, "encoder_q-layer.5": 4422.1836, "encoder_q-layer.6": 4916.1318, "encoder_q-layer.7": 5504.2988, "encoder_q-layer.8": 6788.8394, "encoder_q-layer.9": 6500.3945, "epoch": 0.84, "inbatch_neg_score": 5.5153, "inbatch_pos_score": 6.125, "learning_rate": 1.6052631578947368e-06, "loss": 3.3434, "norm_diff": 0.0846, "norm_loss": 0.0, "num_token_doc": 66.6997, "num_token_overlap": 11.6779, "num_token_query": 31.3693, "num_token_union": 65.0489, "num_word_context": 202.2977, "num_word_doc": 49.7806, "num_word_query": 23.3077, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10858.4376, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5156, "query_norm": 2.6369, "queue_k_norm": 2.7205, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3693, "sent_len_1": 66.6997, "sent_len_max_0": 127.9737, "sent_len_max_1": 206.9062, "stdk": 0.0497, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 193900 }, { "accuracy": 52.7344, "active_queue_size": 16384.0, "cl_loss": 3.3381, "doc_norm": 2.7234, "encoder_q-embeddings": 4348.8062, "encoder_q-layer.0": 3112.4841, "encoder_q-layer.1": 3338.417, "encoder_q-layer.10": 7873.7383, "encoder_q-layer.11": 19028.4102, "encoder_q-layer.2": 3722.6577, "encoder_q-layer.3": 3912.1384, "encoder_q-layer.4": 3994.2803, "encoder_q-layer.5": 4190.3052, "encoder_q-layer.6": 4649.2461, "encoder_q-layer.7": 5541.2646, "encoder_q-layer.8": 6801.8501, "encoder_q-layer.9": 6515.876, "epoch": 0.84, "inbatch_neg_score": 5.5159, "inbatch_pos_score": 6.1367, "learning_rate": 1.5789473684210528e-06, "loss": 3.3381, "norm_diff": 0.0886, "norm_loss": 0.0, "num_token_doc": 66.7941, "num_token_overlap": 11.6892, "num_token_query": 31.4085, "num_token_union": 65.1067, "num_word_context": 202.2769, "num_word_doc": 49.8107, "num_word_query": 23.3276, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11243.9906, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5156, "query_norm": 2.6348, "queue_k_norm": 2.7198, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4085, "sent_len_1": 66.7941, "sent_len_max_0": 127.98, "sent_len_max_1": 208.975, "stdk": 0.0498, "stdq": 0.0428, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 194000 }, { "accuracy": 51.709, "active_queue_size": 16384.0, "cl_loss": 3.323, "doc_norm": 2.719, "encoder_q-embeddings": 5311.6309, "encoder_q-layer.0": 3843.0786, "encoder_q-layer.1": 4110.3213, "encoder_q-layer.10": 6670.4351, "encoder_q-layer.11": 17602.6992, "encoder_q-layer.2": 4721.3008, "encoder_q-layer.3": 4807.2275, "encoder_q-layer.4": 5230.7344, "encoder_q-layer.5": 5214.269, "encoder_q-layer.6": 5837.5322, "encoder_q-layer.7": 6446.1562, "encoder_q-layer.8": 7136.6338, "encoder_q-layer.9": 6545.6567, "epoch": 0.84, "inbatch_neg_score": 5.517, "inbatch_pos_score": 6.1328, "learning_rate": 1.5526315789473686e-06, "loss": 3.323, "norm_diff": 0.0808, "norm_loss": 0.0, "num_token_doc": 66.6682, "num_token_overlap": 11.6796, "num_token_query": 31.4158, "num_token_union": 65.1029, "num_word_context": 202.0458, "num_word_doc": 49.7696, "num_word_query": 23.3348, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11418.4615, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5156, "query_norm": 2.6381, "queue_k_norm": 2.7202, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4158, "sent_len_1": 66.6682, "sent_len_max_0": 127.9825, "sent_len_max_1": 208.7113, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194100 }, { "accuracy": 52.0996, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.7234, "encoder_q-embeddings": 4639.3398, "encoder_q-layer.0": 3354.8525, "encoder_q-layer.1": 3683.9028, "encoder_q-layer.10": 6596.626, "encoder_q-layer.11": 17594.9062, "encoder_q-layer.2": 4110.8003, "encoder_q-layer.3": 4227.7681, "encoder_q-layer.4": 4569.2671, "encoder_q-layer.5": 4898.2847, "encoder_q-layer.6": 5271.02, "encoder_q-layer.7": 5961.9097, "encoder_q-layer.8": 7060.9175, "encoder_q-layer.9": 6635.4082, "epoch": 0.84, "inbatch_neg_score": 5.5173, "inbatch_pos_score": 6.1445, "learning_rate": 1.5263157894736842e-06, "loss": 3.3331, "norm_diff": 0.0829, "norm_loss": 0.0, "num_token_doc": 66.8736, "num_token_overlap": 11.6605, "num_token_query": 31.3592, "num_token_union": 65.1401, "num_word_context": 202.0445, "num_word_doc": 49.8684, "num_word_query": 23.276, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10988.129, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5156, "query_norm": 2.6405, "queue_k_norm": 2.7208, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3592, "sent_len_1": 66.8736, "sent_len_max_0": 127.9638, "sent_len_max_1": 210.575, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194200 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3369, "doc_norm": 2.7181, "encoder_q-embeddings": 4617.7837, "encoder_q-layer.0": 3295.5361, "encoder_q-layer.1": 3513.0962, "encoder_q-layer.10": 7407.9297, "encoder_q-layer.11": 18157.1895, "encoder_q-layer.2": 3848.8342, "encoder_q-layer.3": 3886.2449, "encoder_q-layer.4": 4163.9551, "encoder_q-layer.5": 4342.8579, "encoder_q-layer.6": 4942.2344, "encoder_q-layer.7": 5738.0054, "encoder_q-layer.8": 6794.0928, "encoder_q-layer.9": 6686.9609, "epoch": 0.84, "inbatch_neg_score": 5.5189, "inbatch_pos_score": 6.125, "learning_rate": 1.5e-06, "loss": 3.3369, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.9069, "num_token_overlap": 11.6754, "num_token_query": 31.3507, "num_token_union": 65.157, "num_word_context": 202.4229, "num_word_doc": 49.8822, "num_word_query": 23.2797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11045.8118, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5195, "query_norm": 2.6383, "queue_k_norm": 2.7211, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3507, "sent_len_1": 66.9069, "sent_len_max_0": 127.9762, "sent_len_max_1": 210.9837, "stdk": 0.0493, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194300 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3319, "doc_norm": 2.7212, "encoder_q-embeddings": 4745.002, "encoder_q-layer.0": 3276.9695, "encoder_q-layer.1": 3497.4634, "encoder_q-layer.10": 7626.7891, "encoder_q-layer.11": 18152.0176, "encoder_q-layer.2": 3916.9209, "encoder_q-layer.3": 4016.4646, "encoder_q-layer.4": 4132.6426, "encoder_q-layer.5": 4315.8701, "encoder_q-layer.6": 4915.1641, "encoder_q-layer.7": 5657.0181, "encoder_q-layer.8": 6921.7715, "encoder_q-layer.9": 6530.3413, "epoch": 0.84, "inbatch_neg_score": 5.5193, "inbatch_pos_score": 6.125, "learning_rate": 1.4736842105263159e-06, "loss": 3.3319, "norm_diff": 0.0828, "norm_loss": 0.0, "num_token_doc": 66.8363, "num_token_overlap": 11.6839, "num_token_query": 31.4086, "num_token_union": 65.1927, "num_word_context": 202.382, "num_word_doc": 49.8723, "num_word_query": 23.3324, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11117.3003, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5195, "query_norm": 2.6384, "queue_k_norm": 2.7211, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4086, "sent_len_1": 66.8363, "sent_len_max_0": 127.99, "sent_len_max_1": 209.6738, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194400 }, { "accuracy": 53.5156, "active_queue_size": 16384.0, "cl_loss": 3.331, "doc_norm": 2.7194, "encoder_q-embeddings": 13132.6523, "encoder_q-layer.0": 9204.8799, "encoder_q-layer.1": 10400.3848, "encoder_q-layer.10": 7596.9375, "encoder_q-layer.11": 17751.793, "encoder_q-layer.2": 11857.4727, "encoder_q-layer.3": 11954.916, "encoder_q-layer.4": 12716.6445, "encoder_q-layer.5": 13654.1748, "encoder_q-layer.6": 12949.6396, "encoder_q-layer.7": 12864.123, "encoder_q-layer.8": 10891.9141, "encoder_q-layer.9": 7119.8828, "epoch": 0.84, "inbatch_neg_score": 5.519, "inbatch_pos_score": 6.1445, "learning_rate": 1.4473684210526317e-06, "loss": 3.331, "norm_diff": 0.0779, "norm_loss": 0.0, "num_token_doc": 66.7108, "num_token_overlap": 11.7032, "num_token_query": 31.4213, "num_token_union": 65.0904, "num_word_context": 202.2544, "num_word_doc": 49.7915, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 18443.8949, "preclip_grad_norm_avg": 0.0002, "q@queue_neg_score": 5.5195, "query_norm": 2.6414, "queue_k_norm": 2.7195, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4213, "sent_len_1": 66.7108, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.105, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 194500 }, { "accuracy": 51.1719, "active_queue_size": 16384.0, "cl_loss": 3.327, "doc_norm": 2.7239, "encoder_q-embeddings": 4830.6509, "encoder_q-layer.0": 3338.562, "encoder_q-layer.1": 3599.8147, "encoder_q-layer.10": 8531.4277, "encoder_q-layer.11": 19168.6641, "encoder_q-layer.2": 4036.1323, "encoder_q-layer.3": 4210.937, "encoder_q-layer.4": 4482.5151, "encoder_q-layer.5": 4668.748, "encoder_q-layer.6": 5373.6455, "encoder_q-layer.7": 6699.2119, "encoder_q-layer.8": 7486.144, "encoder_q-layer.9": 6981.7471, "epoch": 0.84, "inbatch_neg_score": 5.5198, "inbatch_pos_score": 6.1367, "learning_rate": 1.4210526315789475e-06, "loss": 3.327, "norm_diff": 0.0811, "norm_loss": 0.0, "num_token_doc": 66.8105, "num_token_overlap": 11.6658, "num_token_query": 31.2302, "num_token_union": 65.0478, "num_word_context": 202.0533, "num_word_doc": 49.8691, "num_word_query": 23.1869, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11727.987, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5195, "query_norm": 2.6428, "queue_k_norm": 2.721, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2302, "sent_len_1": 66.8105, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.2363, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194600 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.329, "doc_norm": 2.7209, "encoder_q-embeddings": 5693.1675, "encoder_q-layer.0": 4249.1748, "encoder_q-layer.1": 4486.5742, "encoder_q-layer.10": 7102.3169, "encoder_q-layer.11": 17610.9609, "encoder_q-layer.2": 5151.3638, "encoder_q-layer.3": 4971.8267, "encoder_q-layer.4": 4823.8848, "encoder_q-layer.5": 4632.6387, "encoder_q-layer.6": 5018.3936, "encoder_q-layer.7": 5474.7783, "encoder_q-layer.8": 6677.0732, "encoder_q-layer.9": 6154.5698, "epoch": 0.84, "inbatch_neg_score": 5.5199, "inbatch_pos_score": 6.1367, "learning_rate": 1.3947368421052633e-06, "loss": 3.329, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.7894, "num_token_overlap": 11.6914, "num_token_query": 31.4433, "num_token_union": 65.1477, "num_word_context": 202.3145, "num_word_doc": 49.8519, "num_word_query": 23.3505, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11343.0592, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5195, "query_norm": 2.6401, "queue_k_norm": 2.7206, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4433, "sent_len_1": 66.7894, "sent_len_max_0": 127.995, "sent_len_max_1": 208.8988, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 194700 }, { "accuracy": 53.5645, "active_queue_size": 16384.0, "cl_loss": 3.3301, "doc_norm": 2.7225, "encoder_q-embeddings": 2254.802, "encoder_q-layer.0": 1610.3942, "encoder_q-layer.1": 1682.5248, "encoder_q-layer.10": 3967.217, "encoder_q-layer.11": 8474.5693, "encoder_q-layer.2": 1821.2666, "encoder_q-layer.3": 1877.9747, "encoder_q-layer.4": 2030.7755, "encoder_q-layer.5": 2130.2544, "encoder_q-layer.6": 2415.9207, "encoder_q-layer.7": 2962.5586, "encoder_q-layer.8": 3267.166, "encoder_q-layer.9": 3122.449, "epoch": 0.85, "inbatch_neg_score": 5.52, "inbatch_pos_score": 6.1523, "learning_rate": 1.3684210526315791e-06, "loss": 3.3301, "norm_diff": 0.0799, "norm_loss": 0.0, "num_token_doc": 66.7244, "num_token_overlap": 11.6804, "num_token_query": 31.4362, "num_token_union": 65.1259, "num_word_context": 202.4003, "num_word_doc": 49.7808, "num_word_query": 23.3593, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5271.0463, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5195, "query_norm": 2.6426, "queue_k_norm": 2.7206, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4362, "sent_len_1": 66.7244, "sent_len_max_0": 127.985, "sent_len_max_1": 207.1188, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 194800 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3303, "doc_norm": 2.7208, "encoder_q-embeddings": 2763.428, "encoder_q-layer.0": 2069.2485, "encoder_q-layer.1": 2415.7454, "encoder_q-layer.10": 3650.1321, "encoder_q-layer.11": 8961.5107, "encoder_q-layer.2": 2730.0972, "encoder_q-layer.3": 2938.697, "encoder_q-layer.4": 3181.1819, "encoder_q-layer.5": 3359.488, "encoder_q-layer.6": 3495.0259, "encoder_q-layer.7": 3639.1558, "encoder_q-layer.8": 3799.438, "encoder_q-layer.9": 3360.314, "epoch": 0.85, "inbatch_neg_score": 5.5219, "inbatch_pos_score": 6.1328, "learning_rate": 1.342105263157895e-06, "loss": 3.3303, "norm_diff": 0.0812, "norm_loss": 0.0, "num_token_doc": 66.9112, "num_token_overlap": 11.6776, "num_token_query": 31.3209, "num_token_union": 65.1871, "num_word_context": 202.1668, "num_word_doc": 49.9475, "num_word_query": 23.2643, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6136.7139, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.6396, "queue_k_norm": 2.7215, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3209, "sent_len_1": 66.9112, "sent_len_max_0": 127.9862, "sent_len_max_1": 209.2887, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 194900 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.336, "doc_norm": 2.7178, "encoder_q-embeddings": 2350.4211, "encoder_q-layer.0": 1671.9351, "encoder_q-layer.1": 1795.1361, "encoder_q-layer.10": 3657.6709, "encoder_q-layer.11": 8750.835, "encoder_q-layer.2": 2056.698, "encoder_q-layer.3": 2113.8057, "encoder_q-layer.4": 2252.3442, "encoder_q-layer.5": 2285.7014, "encoder_q-layer.6": 2551.4407, "encoder_q-layer.7": 2920.6172, "encoder_q-layer.8": 3665.4724, "encoder_q-layer.9": 3256.2114, "epoch": 0.85, "inbatch_neg_score": 5.5212, "inbatch_pos_score": 6.1328, "learning_rate": 1.3157894736842106e-06, "loss": 3.336, "norm_diff": 0.0774, "norm_loss": 0.0, "num_token_doc": 66.8423, "num_token_overlap": 11.699, "num_token_query": 31.3934, "num_token_union": 65.1383, "num_word_context": 202.4629, "num_word_doc": 49.8518, "num_word_query": 23.3109, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5520.9756, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5195, "query_norm": 2.6404, "queue_k_norm": 2.7214, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3934, "sent_len_1": 66.8423, "sent_len_max_0": 128.0, "sent_len_max_1": 209.7625, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195000 }, { "accuracy": 52.9785, "active_queue_size": 16384.0, "cl_loss": 3.3344, "doc_norm": 2.7226, "encoder_q-embeddings": 2437.3369, "encoder_q-layer.0": 1753.8756, "encoder_q-layer.1": 1903.2991, "encoder_q-layer.10": 3412.2703, "encoder_q-layer.11": 8673.1895, "encoder_q-layer.2": 2170.4521, "encoder_q-layer.3": 2266.7024, "encoder_q-layer.4": 2299.0103, "encoder_q-layer.5": 2339.9985, "encoder_q-layer.6": 2605.9988, "encoder_q-layer.7": 3061.2524, "encoder_q-layer.8": 3474.769, "encoder_q-layer.9": 3126.9365, "epoch": 0.85, "inbatch_neg_score": 5.5217, "inbatch_pos_score": 6.1484, "learning_rate": 1.2894736842105264e-06, "loss": 3.3344, "norm_diff": 0.081, "norm_loss": 0.0, "num_token_doc": 66.7437, "num_token_overlap": 11.6741, "num_token_query": 31.4404, "num_token_union": 65.1712, "num_word_context": 202.5044, "num_word_doc": 49.8279, "num_word_query": 23.3643, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5500.8991, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.6416, "queue_k_norm": 2.7214, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4404, "sent_len_1": 66.7437, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.02, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195100 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3413, "doc_norm": 2.7197, "encoder_q-embeddings": 2521.9683, "encoder_q-layer.0": 1848.2158, "encoder_q-layer.1": 1953.6929, "encoder_q-layer.10": 3500.8594, "encoder_q-layer.11": 9099.0742, "encoder_q-layer.2": 2282.8018, "encoder_q-layer.3": 2384.2537, "encoder_q-layer.4": 2534.9019, "encoder_q-layer.5": 2799.48, "encoder_q-layer.6": 3112.1226, "encoder_q-layer.7": 3238.9702, "encoder_q-layer.8": 3578.0273, "encoder_q-layer.9": 3385.259, "epoch": 0.85, "inbatch_neg_score": 5.5233, "inbatch_pos_score": 6.1289, "learning_rate": 1.2631578947368422e-06, "loss": 3.3413, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.6771, "num_token_overlap": 11.6658, "num_token_query": 31.3462, "num_token_union": 65.0403, "num_word_context": 201.9806, "num_word_doc": 49.7089, "num_word_query": 23.2715, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5810.6983, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.6401, "queue_k_norm": 2.7202, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3462, "sent_len_1": 66.6771, "sent_len_max_0": 127.9975, "sent_len_max_1": 210.6125, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 195200 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3267, "doc_norm": 2.7222, "encoder_q-embeddings": 2417.3303, "encoder_q-layer.0": 1720.3707, "encoder_q-layer.1": 1889.8608, "encoder_q-layer.10": 3708.936, "encoder_q-layer.11": 8844.7803, "encoder_q-layer.2": 2056.0789, "encoder_q-layer.3": 2158.646, "encoder_q-layer.4": 2264.751, "encoder_q-layer.5": 2305.2642, "encoder_q-layer.6": 2681.9336, "encoder_q-layer.7": 2864.2659, "encoder_q-layer.8": 3457.2637, "encoder_q-layer.9": 3247.0562, "epoch": 0.85, "inbatch_neg_score": 5.5252, "inbatch_pos_score": 6.1367, "learning_rate": 1.236842105263158e-06, "loss": 3.3267, "norm_diff": 0.0822, "norm_loss": 0.0, "num_token_doc": 66.7946, "num_token_overlap": 11.6739, "num_token_query": 31.3313, "num_token_union": 65.1108, "num_word_context": 202.2351, "num_word_doc": 49.8515, "num_word_query": 23.2576, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5581.3735, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.64, "queue_k_norm": 2.7216, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3313, "sent_len_1": 66.7946, "sent_len_max_0": 127.975, "sent_len_max_1": 208.1962, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195300 }, { "accuracy": 53.6621, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.7243, "encoder_q-embeddings": 2256.2217, "encoder_q-layer.0": 1628.615, "encoder_q-layer.1": 1762.2712, "encoder_q-layer.10": 3801.5857, "encoder_q-layer.11": 9057.2695, "encoder_q-layer.2": 1971.938, "encoder_q-layer.3": 2042.4574, "encoder_q-layer.4": 2160.0901, "encoder_q-layer.5": 2266.6948, "encoder_q-layer.6": 2535.8086, "encoder_q-layer.7": 2844.946, "encoder_q-layer.8": 3528.9614, "encoder_q-layer.9": 3448.5632, "epoch": 0.85, "inbatch_neg_score": 5.5241, "inbatch_pos_score": 6.1562, "learning_rate": 1.2105263157894738e-06, "loss": 3.3331, "norm_diff": 0.0791, "norm_loss": 0.0, "num_token_doc": 66.8125, "num_token_overlap": 11.6581, "num_token_query": 31.3183, "num_token_union": 65.1181, "num_word_context": 202.138, "num_word_doc": 49.8368, "num_word_query": 23.2443, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5525.9112, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.6451, "queue_k_norm": 2.7215, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3183, "sent_len_1": 66.8125, "sent_len_max_0": 127.9862, "sent_len_max_1": 208.6138, "stdk": 0.0497, "stdq": 0.0435, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195400 }, { "accuracy": 51.2207, "active_queue_size": 16384.0, "cl_loss": 3.3312, "doc_norm": 2.7206, "encoder_q-embeddings": 2395.3909, "encoder_q-layer.0": 1666.3893, "encoder_q-layer.1": 1799.7493, "encoder_q-layer.10": 3558.3572, "encoder_q-layer.11": 9430.6318, "encoder_q-layer.2": 1993.2627, "encoder_q-layer.3": 2089.3848, "encoder_q-layer.4": 2290.4736, "encoder_q-layer.5": 2300.1875, "encoder_q-layer.6": 2561.7891, "encoder_q-layer.7": 3083.4155, "encoder_q-layer.8": 3718.8379, "encoder_q-layer.9": 3305.8516, "epoch": 0.85, "inbatch_neg_score": 5.5257, "inbatch_pos_score": 6.1406, "learning_rate": 1.1842105263157896e-06, "loss": 3.3312, "norm_diff": 0.0785, "norm_loss": 0.0, "num_token_doc": 66.763, "num_token_overlap": 11.6882, "num_token_query": 31.4662, "num_token_union": 65.176, "num_word_context": 202.2865, "num_word_doc": 49.8339, "num_word_query": 23.3545, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5854.1501, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.6421, "queue_k_norm": 2.722, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4662, "sent_len_1": 66.763, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.0288, "stdk": 0.0494, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195500 }, { "accuracy": 52.9297, "active_queue_size": 16384.0, "cl_loss": 3.336, "doc_norm": 2.7227, "encoder_q-embeddings": 2459.5068, "encoder_q-layer.0": 1683.938, "encoder_q-layer.1": 1898.4496, "encoder_q-layer.10": 3687.9053, "encoder_q-layer.11": 9225.9795, "encoder_q-layer.2": 2179.8801, "encoder_q-layer.3": 2213.9573, "encoder_q-layer.4": 2294.9631, "encoder_q-layer.5": 2387.4524, "encoder_q-layer.6": 2597.3789, "encoder_q-layer.7": 2825.7905, "encoder_q-layer.8": 3479.6072, "encoder_q-layer.9": 3283.613, "epoch": 0.85, "inbatch_neg_score": 5.5248, "inbatch_pos_score": 6.1445, "learning_rate": 1.1578947368421055e-06, "loss": 3.336, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.7875, "num_token_overlap": 11.6907, "num_token_query": 31.3894, "num_token_union": 65.1065, "num_word_context": 202.3504, "num_word_doc": 49.8696, "num_word_query": 23.3163, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5707.0208, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5234, "query_norm": 2.639, "queue_k_norm": 2.7219, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3894, "sent_len_1": 66.7875, "sent_len_max_0": 127.9775, "sent_len_max_1": 209.0462, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195600 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3326, "doc_norm": 2.7188, "encoder_q-embeddings": 2405.8372, "encoder_q-layer.0": 1623.3563, "encoder_q-layer.1": 1775.2893, "encoder_q-layer.10": 3552.0698, "encoder_q-layer.11": 9115.9297, "encoder_q-layer.2": 1940.353, "encoder_q-layer.3": 2049.5908, "encoder_q-layer.4": 2237.6365, "encoder_q-layer.5": 2313.175, "encoder_q-layer.6": 2544.0581, "encoder_q-layer.7": 2890.9192, "encoder_q-layer.8": 3505.8113, "encoder_q-layer.9": 3315.4937, "epoch": 0.85, "inbatch_neg_score": 5.5266, "inbatch_pos_score": 6.1406, "learning_rate": 1.1315789473684213e-06, "loss": 3.3326, "norm_diff": 0.078, "norm_loss": 0.0, "num_token_doc": 66.8064, "num_token_overlap": 11.6909, "num_token_query": 31.3878, "num_token_union": 65.138, "num_word_context": 202.139, "num_word_doc": 49.8368, "num_word_query": 23.2947, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5609.7623, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6408, "queue_k_norm": 2.722, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3878, "sent_len_1": 66.8064, "sent_len_max_0": 127.9975, "sent_len_max_1": 209.4212, "stdk": 0.0493, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 195700 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3295, "doc_norm": 2.722, "encoder_q-embeddings": 2288.5261, "encoder_q-layer.0": 1640.4185, "encoder_q-layer.1": 1783.4307, "encoder_q-layer.10": 3727.6289, "encoder_q-layer.11": 9125.1699, "encoder_q-layer.2": 1955.0856, "encoder_q-layer.3": 1985.785, "encoder_q-layer.4": 2179.5342, "encoder_q-layer.5": 2290.2104, "encoder_q-layer.6": 2522.0354, "encoder_q-layer.7": 2899.1145, "encoder_q-layer.8": 3566.1309, "encoder_q-layer.9": 3477.239, "epoch": 0.85, "inbatch_neg_score": 5.5264, "inbatch_pos_score": 6.1445, "learning_rate": 1.1052631578947369e-06, "loss": 3.3295, "norm_diff": 0.0797, "norm_loss": 0.0, "num_token_doc": 66.8395, "num_token_overlap": 11.6715, "num_token_query": 31.4156, "num_token_union": 65.2063, "num_word_context": 202.4652, "num_word_doc": 49.8703, "num_word_query": 23.3421, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5546.8038, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6423, "queue_k_norm": 2.7215, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4156, "sent_len_1": 66.8395, "sent_len_max_0": 127.9525, "sent_len_max_1": 210.5288, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 195800 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3221, "doc_norm": 2.7227, "encoder_q-embeddings": 2448.1584, "encoder_q-layer.0": 1724.6429, "encoder_q-layer.1": 1864.5674, "encoder_q-layer.10": 3491.6423, "encoder_q-layer.11": 8815.6162, "encoder_q-layer.2": 2024.0188, "encoder_q-layer.3": 2065.7117, "encoder_q-layer.4": 2208.7349, "encoder_q-layer.5": 2333.4719, "encoder_q-layer.6": 2643.2681, "encoder_q-layer.7": 2848.4058, "encoder_q-layer.8": 3516.2334, "encoder_q-layer.9": 3300.0503, "epoch": 0.85, "inbatch_neg_score": 5.5265, "inbatch_pos_score": 6.1445, "learning_rate": 1.0789473684210527e-06, "loss": 3.3221, "norm_diff": 0.0817, "norm_loss": 0.0, "num_token_doc": 66.8703, "num_token_overlap": 11.708, "num_token_query": 31.3827, "num_token_union": 65.1696, "num_word_context": 202.2108, "num_word_doc": 49.9085, "num_word_query": 23.3052, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5521.6331, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.641, "queue_k_norm": 2.7213, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3827, "sent_len_1": 66.8703, "sent_len_max_0": 127.98, "sent_len_max_1": 208.7675, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 195900 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3271, "doc_norm": 2.7237, "encoder_q-embeddings": 2325.6589, "encoder_q-layer.0": 1659.1556, "encoder_q-layer.1": 1784.3965, "encoder_q-layer.10": 3556.374, "encoder_q-layer.11": 8636.3877, "encoder_q-layer.2": 1959.8048, "encoder_q-layer.3": 2096.8867, "encoder_q-layer.4": 2184.4138, "encoder_q-layer.5": 2244.6611, "encoder_q-layer.6": 2513.0696, "encoder_q-layer.7": 2706.573, "encoder_q-layer.8": 3360.8398, "encoder_q-layer.9": 3278.4172, "epoch": 0.85, "inbatch_neg_score": 5.5288, "inbatch_pos_score": 6.1445, "learning_rate": 1.0526315789473685e-06, "loss": 3.3271, "norm_diff": 0.0798, "norm_loss": 0.0, "num_token_doc": 66.7683, "num_token_overlap": 11.7, "num_token_query": 31.3514, "num_token_union": 65.0573, "num_word_context": 202.2407, "num_word_doc": 49.8299, "num_word_query": 23.2823, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5383.2297, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5273, "query_norm": 2.6439, "queue_k_norm": 2.7225, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3514, "sent_len_1": 66.7683, "sent_len_max_0": 127.995, "sent_len_max_1": 206.0775, "stdk": 0.0496, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196000 }, { "accuracy": 50.0488, "active_queue_size": 16384.0, "cl_loss": 3.3345, "doc_norm": 2.7239, "encoder_q-embeddings": 2396.7178, "encoder_q-layer.0": 1692.7253, "encoder_q-layer.1": 1787.6573, "encoder_q-layer.10": 4324.7534, "encoder_q-layer.11": 9399.2852, "encoder_q-layer.2": 2010.6298, "encoder_q-layer.3": 2074.6704, "encoder_q-layer.4": 2263.9043, "encoder_q-layer.5": 2294.5249, "encoder_q-layer.6": 2638.1721, "encoder_q-layer.7": 3015.7219, "encoder_q-layer.8": 3676.4421, "encoder_q-layer.9": 3588.9521, "epoch": 0.85, "inbatch_neg_score": 5.5291, "inbatch_pos_score": 6.1367, "learning_rate": 1.0263157894736843e-06, "loss": 3.3345, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.8475, "num_token_overlap": 11.6847, "num_token_query": 31.3658, "num_token_union": 65.1777, "num_word_context": 202.2667, "num_word_doc": 49.8966, "num_word_query": 23.308, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5782.2965, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6413, "queue_k_norm": 2.7227, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3658, "sent_len_1": 66.8475, "sent_len_max_0": 127.9613, "sent_len_max_1": 208.9487, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196100 }, { "accuracy": 53.0273, "active_queue_size": 16384.0, "cl_loss": 3.3265, "doc_norm": 2.7241, "encoder_q-embeddings": 2668.6819, "encoder_q-layer.0": 1965.1028, "encoder_q-layer.1": 2100.573, "encoder_q-layer.10": 3532.0288, "encoder_q-layer.11": 8641.5312, "encoder_q-layer.2": 2302.1948, "encoder_q-layer.3": 2423.8198, "encoder_q-layer.4": 2553.4072, "encoder_q-layer.5": 2612.7224, "encoder_q-layer.6": 3113.7473, "encoder_q-layer.7": 3317.0286, "encoder_q-layer.8": 3831.3196, "encoder_q-layer.9": 3126.4548, "epoch": 0.85, "inbatch_neg_score": 5.527, "inbatch_pos_score": 6.1484, "learning_rate": 1.0000000000000002e-06, "loss": 3.3265, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.8691, "num_token_overlap": 11.6904, "num_token_query": 31.3216, "num_token_union": 65.1323, "num_word_context": 202.5627, "num_word_doc": 49.9157, "num_word_query": 23.2684, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5672.3301, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6411, "queue_k_norm": 2.7229, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3216, "sent_len_1": 66.8691, "sent_len_max_0": 127.9838, "sent_len_max_1": 207.4038, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196200 }, { "accuracy": 50.7324, "active_queue_size": 16384.0, "cl_loss": 3.3445, "doc_norm": 2.7227, "encoder_q-embeddings": 2436.0049, "encoder_q-layer.0": 1718.8228, "encoder_q-layer.1": 1852.333, "encoder_q-layer.10": 3682.3914, "encoder_q-layer.11": 9197.9219, "encoder_q-layer.2": 2051.9277, "encoder_q-layer.3": 2099.5181, "encoder_q-layer.4": 2256.9202, "encoder_q-layer.5": 2463.4661, "encoder_q-layer.6": 2882.7109, "encoder_q-layer.7": 3058.1973, "encoder_q-layer.8": 3594.647, "encoder_q-layer.9": 3337.603, "epoch": 0.85, "inbatch_neg_score": 5.5282, "inbatch_pos_score": 6.1406, "learning_rate": 9.73684210526316e-07, "loss": 3.3445, "norm_diff": 0.0816, "norm_loss": 0.0, "num_token_doc": 66.7538, "num_token_overlap": 11.6598, "num_token_query": 31.3271, "num_token_union": 65.106, "num_word_context": 202.5489, "num_word_doc": 49.8154, "num_word_query": 23.2612, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5786.9086, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6411, "queue_k_norm": 2.723, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3271, "sent_len_1": 66.7538, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.3375, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196300 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3395, "doc_norm": 2.7222, "encoder_q-embeddings": 2569.5698, "encoder_q-layer.0": 1803.524, "encoder_q-layer.1": 1907.9811, "encoder_q-layer.10": 3469.4517, "encoder_q-layer.11": 8891.6895, "encoder_q-layer.2": 2246.2322, "encoder_q-layer.3": 2327.4595, "encoder_q-layer.4": 2599.6636, "encoder_q-layer.5": 2582.7737, "encoder_q-layer.6": 2872.3169, "encoder_q-layer.7": 3156.5857, "encoder_q-layer.8": 3673.3296, "encoder_q-layer.9": 3313.7078, "epoch": 0.85, "inbatch_neg_score": 5.5292, "inbatch_pos_score": 6.1484, "learning_rate": 9.473684210526317e-07, "loss": 3.3395, "norm_diff": 0.0804, "norm_loss": 0.0, "num_token_doc": 66.6035, "num_token_overlap": 11.6943, "num_token_query": 31.4409, "num_token_union": 65.0037, "num_word_context": 202.2055, "num_word_doc": 49.6842, "num_word_query": 23.3597, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5773.6764, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6419, "queue_k_norm": 2.7233, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4409, "sent_len_1": 66.6035, "sent_len_max_0": 127.9638, "sent_len_max_1": 209.4538, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196400 }, { "accuracy": 53.7598, "active_queue_size": 16384.0, "cl_loss": 3.3186, "doc_norm": 2.7231, "encoder_q-embeddings": 2562.5994, "encoder_q-layer.0": 1821.8744, "encoder_q-layer.1": 1983.2812, "encoder_q-layer.10": 3499.4795, "encoder_q-layer.11": 8695.5273, "encoder_q-layer.2": 2275.8677, "encoder_q-layer.3": 2306.8054, "encoder_q-layer.4": 2511.7605, "encoder_q-layer.5": 2680.8843, "encoder_q-layer.6": 2907.8623, "encoder_q-layer.7": 3027.3079, "encoder_q-layer.8": 3633.1895, "encoder_q-layer.9": 3323.0962, "epoch": 0.85, "inbatch_neg_score": 5.5291, "inbatch_pos_score": 6.1562, "learning_rate": 9.210526315789473e-07, "loss": 3.3186, "norm_diff": 0.0822, "norm_loss": 0.0, "num_token_doc": 66.8135, "num_token_overlap": 11.7221, "num_token_query": 31.465, "num_token_union": 65.1567, "num_word_context": 202.632, "num_word_doc": 49.8417, "num_word_query": 23.3817, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5581.7925, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5273, "query_norm": 2.6409, "queue_k_norm": 2.7229, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.465, "sent_len_1": 66.8135, "sent_len_max_0": 127.9862, "sent_len_max_1": 210.4487, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196500 }, { "accuracy": 51.6113, "active_queue_size": 16384.0, "cl_loss": 3.3379, "doc_norm": 2.7236, "encoder_q-embeddings": 2408.9502, "encoder_q-layer.0": 1662.2725, "encoder_q-layer.1": 1815.8392, "encoder_q-layer.10": 3388.3005, "encoder_q-layer.11": 9089.1357, "encoder_q-layer.2": 2017.152, "encoder_q-layer.3": 2093.3032, "encoder_q-layer.4": 2200.5247, "encoder_q-layer.5": 2297.7952, "encoder_q-layer.6": 2559.4412, "encoder_q-layer.7": 2938.7966, "encoder_q-layer.8": 3504.9077, "encoder_q-layer.9": 3278.428, "epoch": 0.85, "inbatch_neg_score": 5.5296, "inbatch_pos_score": 6.1484, "learning_rate": 8.947368421052631e-07, "loss": 3.3379, "norm_diff": 0.0831, "norm_loss": 0.0, "num_token_doc": 66.6376, "num_token_overlap": 11.6917, "num_token_query": 31.4387, "num_token_union": 65.0703, "num_word_context": 202.2204, "num_word_doc": 49.7663, "num_word_query": 23.3611, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5661.3479, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6405, "queue_k_norm": 2.7237, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4387, "sent_len_1": 66.6376, "sent_len_max_0": 127.995, "sent_len_max_1": 207.5325, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 196600 }, { "accuracy": 51.9043, "active_queue_size": 16384.0, "cl_loss": 3.3342, "doc_norm": 2.7228, "encoder_q-embeddings": 2296.9958, "encoder_q-layer.0": 1594.8464, "encoder_q-layer.1": 1692.7767, "encoder_q-layer.10": 3558.7144, "encoder_q-layer.11": 9008.9521, "encoder_q-layer.2": 1896.2848, "encoder_q-layer.3": 1919.2542, "encoder_q-layer.4": 2099.479, "encoder_q-layer.5": 2174.8872, "encoder_q-layer.6": 2442.3904, "encoder_q-layer.7": 2891.3213, "encoder_q-layer.8": 3223.9897, "encoder_q-layer.9": 3137.4705, "epoch": 0.85, "inbatch_neg_score": 5.5291, "inbatch_pos_score": 6.1406, "learning_rate": 8.684210526315789e-07, "loss": 3.3342, "norm_diff": 0.0847, "norm_loss": 0.0, "num_token_doc": 66.816, "num_token_overlap": 11.6534, "num_token_query": 31.3655, "num_token_union": 65.1359, "num_word_context": 202.2253, "num_word_doc": 49.8467, "num_word_query": 23.2797, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5439.0557, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5312, "query_norm": 2.638, "queue_k_norm": 2.7242, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3655, "sent_len_1": 66.816, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.6825, "stdk": 0.0495, "stdq": 0.0429, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 196700 }, { "accuracy": 51.9531, "active_queue_size": 16384.0, "cl_loss": 3.3354, "doc_norm": 2.7226, "encoder_q-embeddings": 4531.2324, "encoder_q-layer.0": 3303.0117, "encoder_q-layer.1": 3535.9736, "encoder_q-layer.10": 7412.1436, "encoder_q-layer.11": 17820.5059, "encoder_q-layer.2": 3891.0107, "encoder_q-layer.3": 3935.22, "encoder_q-layer.4": 4203.0991, "encoder_q-layer.5": 4451.166, "encoder_q-layer.6": 5053.5557, "encoder_q-layer.7": 5904.3052, "encoder_q-layer.8": 6990.4795, "encoder_q-layer.9": 6298.6104, "epoch": 0.85, "inbatch_neg_score": 5.5311, "inbatch_pos_score": 6.1445, "learning_rate": 8.421052631578948e-07, "loss": 3.3354, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.9157, "num_token_overlap": 11.6469, "num_token_query": 31.2869, "num_token_union": 65.1799, "num_word_context": 202.4449, "num_word_doc": 49.9291, "num_word_query": 23.2376, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11061.7865, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.643, "queue_k_norm": 2.7242, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2869, "sent_len_1": 66.9157, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.0788, "stdk": 0.0495, "stdq": 0.0433, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 196800 }, { "accuracy": 51.6602, "active_queue_size": 16384.0, "cl_loss": 3.3243, "doc_norm": 2.7218, "encoder_q-embeddings": 4722.167, "encoder_q-layer.0": 3445.1736, "encoder_q-layer.1": 3690.646, "encoder_q-layer.10": 6766.7202, "encoder_q-layer.11": 17615.6523, "encoder_q-layer.2": 4237.2393, "encoder_q-layer.3": 4237.6582, "encoder_q-layer.4": 4594.3735, "encoder_q-layer.5": 4574.855, "encoder_q-layer.6": 5020.522, "encoder_q-layer.7": 5724.0791, "encoder_q-layer.8": 6788.3853, "encoder_q-layer.9": 6334.7861, "epoch": 0.85, "inbatch_neg_score": 5.5302, "inbatch_pos_score": 6.1406, "learning_rate": 8.157894736842106e-07, "loss": 3.3243, "norm_diff": 0.0796, "norm_loss": 0.0, "num_token_doc": 66.7977, "num_token_overlap": 11.6646, "num_token_query": 31.3415, "num_token_union": 65.1336, "num_word_context": 202.1323, "num_word_doc": 49.8694, "num_word_query": 23.2746, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10984.2992, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6422, "queue_k_norm": 2.7247, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3415, "sent_len_1": 66.7977, "sent_len_max_0": 127.9912, "sent_len_max_1": 209.1463, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 196900 }, { "accuracy": 52.6367, "active_queue_size": 16384.0, "cl_loss": 3.3348, "doc_norm": 2.7248, "encoder_q-embeddings": 5205.9453, "encoder_q-layer.0": 3721.7205, "encoder_q-layer.1": 4096.9409, "encoder_q-layer.10": 7117.8428, "encoder_q-layer.11": 17345.0332, "encoder_q-layer.2": 4657.1572, "encoder_q-layer.3": 4704.6641, "encoder_q-layer.4": 5049.7954, "encoder_q-layer.5": 5305.5981, "encoder_q-layer.6": 5786.3271, "encoder_q-layer.7": 6096.083, "encoder_q-layer.8": 7254.8579, "encoder_q-layer.9": 6495.6343, "epoch": 0.85, "inbatch_neg_score": 5.5339, "inbatch_pos_score": 6.1562, "learning_rate": 7.894736842105264e-07, "loss": 3.3348, "norm_diff": 0.0841, "norm_loss": 0.0, "num_token_doc": 66.8775, "num_token_overlap": 11.6876, "num_token_query": 31.3242, "num_token_union": 65.1328, "num_word_context": 202.346, "num_word_doc": 49.8937, "num_word_query": 23.2657, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11205.8761, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6407, "queue_k_norm": 2.7238, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3242, "sent_len_1": 66.8775, "sent_len_max_0": 127.99, "sent_len_max_1": 210.8675, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197000 }, { "accuracy": 51.0742, "active_queue_size": 16384.0, "cl_loss": 3.3411, "doc_norm": 2.723, "encoder_q-embeddings": 4476.397, "encoder_q-layer.0": 3390.6909, "encoder_q-layer.1": 3494.8608, "encoder_q-layer.10": 6777.251, "encoder_q-layer.11": 17952.3164, "encoder_q-layer.2": 3826.6626, "encoder_q-layer.3": 3896.3596, "encoder_q-layer.4": 4188.7031, "encoder_q-layer.5": 4352.9771, "encoder_q-layer.6": 4881.3452, "encoder_q-layer.7": 5496.1543, "encoder_q-layer.8": 6635.7354, "encoder_q-layer.9": 6495.8545, "epoch": 0.86, "inbatch_neg_score": 5.5306, "inbatch_pos_score": 6.1562, "learning_rate": 7.631578947368421e-07, "loss": 3.3411, "norm_diff": 0.0788, "norm_loss": 0.0, "num_token_doc": 66.7005, "num_token_overlap": 11.6794, "num_token_query": 31.3727, "num_token_union": 65.0694, "num_word_context": 202.2546, "num_word_doc": 49.7866, "num_word_query": 23.3183, "postclip_grad_norm": 1.0, "preclip_grad_norm": 10946.2815, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6442, "queue_k_norm": 2.7239, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3727, "sent_len_1": 66.7005, "sent_len_max_0": 127.9725, "sent_len_max_1": 209.9025, "stdk": 0.0495, "stdq": 0.0434, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197100 }, { "accuracy": 49.8047, "active_queue_size": 16384.0, "cl_loss": 3.3382, "doc_norm": 2.7237, "encoder_q-embeddings": 4848.7793, "encoder_q-layer.0": 3479.5691, "encoder_q-layer.1": 3869.4712, "encoder_q-layer.10": 7851.3433, "encoder_q-layer.11": 18630.1621, "encoder_q-layer.2": 4220.7383, "encoder_q-layer.3": 4278.8989, "encoder_q-layer.4": 4525.8125, "encoder_q-layer.5": 4693.7192, "encoder_q-layer.6": 5203.0332, "encoder_q-layer.7": 5992.2217, "encoder_q-layer.8": 7115.354, "encoder_q-layer.9": 6906.6128, "epoch": 0.86, "inbatch_neg_score": 5.5336, "inbatch_pos_score": 6.1406, "learning_rate": 7.368421052631579e-07, "loss": 3.3382, "norm_diff": 0.0822, "norm_loss": 0.0, "num_token_doc": 66.7743, "num_token_overlap": 11.6475, "num_token_query": 31.3658, "num_token_union": 65.1425, "num_word_context": 202.3984, "num_word_doc": 49.845, "num_word_query": 23.2997, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11534.5549, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5312, "query_norm": 2.6414, "queue_k_norm": 2.7237, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3658, "sent_len_1": 66.7743, "sent_len_max_0": 128.0, "sent_len_max_1": 208.5725, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197200 }, { "accuracy": 50.1953, "active_queue_size": 16384.0, "cl_loss": 3.3166, "doc_norm": 2.724, "encoder_q-embeddings": 4882.9272, "encoder_q-layer.0": 3460.814, "encoder_q-layer.1": 3675.4285, "encoder_q-layer.10": 7386.8794, "encoder_q-layer.11": 18443.0078, "encoder_q-layer.2": 4098.833, "encoder_q-layer.3": 4123.2656, "encoder_q-layer.4": 4387.6743, "encoder_q-layer.5": 4545.9321, "encoder_q-layer.6": 5212.9087, "encoder_q-layer.7": 6043.3315, "encoder_q-layer.8": 7032.813, "encoder_q-layer.9": 6736.9658, "epoch": 0.86, "inbatch_neg_score": 5.5344, "inbatch_pos_score": 6.1406, "learning_rate": 7.105263157894737e-07, "loss": 3.3166, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.9045, "num_token_overlap": 11.6736, "num_token_query": 31.4087, "num_token_union": 65.2247, "num_word_context": 202.7276, "num_word_doc": 49.9221, "num_word_query": 23.3345, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11415.3705, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6426, "queue_k_norm": 2.7241, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4087, "sent_len_1": 66.9045, "sent_len_max_0": 127.9762, "sent_len_max_1": 209.41, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197300 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3324, "doc_norm": 2.7215, "encoder_q-embeddings": 5420.7944, "encoder_q-layer.0": 3715.8206, "encoder_q-layer.1": 4105.6973, "encoder_q-layer.10": 6687.4229, "encoder_q-layer.11": 17339.2324, "encoder_q-layer.2": 4821.0571, "encoder_q-layer.3": 5114.5952, "encoder_q-layer.4": 5684.5981, "encoder_q-layer.5": 5841.7661, "encoder_q-layer.6": 6353.9507, "encoder_q-layer.7": 7031.4673, "encoder_q-layer.8": 6977.0142, "encoder_q-layer.9": 6237.3345, "epoch": 0.86, "inbatch_neg_score": 5.5341, "inbatch_pos_score": 6.1406, "learning_rate": 6.842105263157896e-07, "loss": 3.3324, "norm_diff": 0.0809, "norm_loss": 0.0, "num_token_doc": 66.8007, "num_token_overlap": 11.6722, "num_token_query": 31.3754, "num_token_union": 65.1606, "num_word_context": 202.4525, "num_word_doc": 49.8677, "num_word_query": 23.3098, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11487.9104, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6406, "queue_k_norm": 2.7244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3754, "sent_len_1": 66.8007, "sent_len_max_0": 127.9675, "sent_len_max_1": 207.59, "stdk": 0.0494, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197400 }, { "accuracy": 48.8281, "active_queue_size": 16384.0, "cl_loss": 3.3303, "doc_norm": 2.725, "encoder_q-embeddings": 4577.1899, "encoder_q-layer.0": 3237.3762, "encoder_q-layer.1": 3511.4492, "encoder_q-layer.10": 7427.811, "encoder_q-layer.11": 18120.4531, "encoder_q-layer.2": 3898.5376, "encoder_q-layer.3": 3988.3186, "encoder_q-layer.4": 4442.3286, "encoder_q-layer.5": 4456.7808, "encoder_q-layer.6": 5163.9238, "encoder_q-layer.7": 5613.0801, "encoder_q-layer.8": 7258.4526, "encoder_q-layer.9": 6499.8003, "epoch": 0.86, "inbatch_neg_score": 5.5356, "inbatch_pos_score": 6.1367, "learning_rate": 6.578947368421053e-07, "loss": 3.3303, "norm_diff": 0.0822, "norm_loss": 0.0, "num_token_doc": 66.7587, "num_token_overlap": 11.6816, "num_token_query": 31.399, "num_token_union": 65.1153, "num_word_context": 202.4703, "num_word_doc": 49.8219, "num_word_query": 23.323, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11163.6739, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6428, "queue_k_norm": 2.7239, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.399, "sent_len_1": 66.7587, "sent_len_max_0": 127.985, "sent_len_max_1": 209.335, "stdk": 0.0496, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197500 }, { "accuracy": 52.832, "active_queue_size": 16384.0, "cl_loss": 3.3331, "doc_norm": 2.726, "encoder_q-embeddings": 4800.98, "encoder_q-layer.0": 3431.5818, "encoder_q-layer.1": 3852.2388, "encoder_q-layer.10": 7877.8384, "encoder_q-layer.11": 18558.4082, "encoder_q-layer.2": 4385.0337, "encoder_q-layer.3": 4337.231, "encoder_q-layer.4": 4522.731, "encoder_q-layer.5": 4725.0137, "encoder_q-layer.6": 5310.5923, "encoder_q-layer.7": 6113.3828, "encoder_q-layer.8": 7028.8232, "encoder_q-layer.9": 6748.27, "epoch": 0.86, "inbatch_neg_score": 5.5341, "inbatch_pos_score": 6.1602, "learning_rate": 6.315789473684211e-07, "loss": 3.3331, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.9334, "num_token_overlap": 11.6646, "num_token_query": 31.4175, "num_token_union": 65.2878, "num_word_context": 202.3764, "num_word_doc": 49.9068, "num_word_query": 23.3262, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11337.2049, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6442, "queue_k_norm": 2.7244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4175, "sent_len_1": 66.9334, "sent_len_max_0": 127.9875, "sent_len_max_1": 209.31, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197600 }, { "accuracy": 50.5859, "active_queue_size": 16384.0, "cl_loss": 3.3364, "doc_norm": 2.7241, "encoder_q-embeddings": 2419.644, "encoder_q-layer.0": 1687.1187, "encoder_q-layer.1": 1784.29, "encoder_q-layer.10": 3972.4475, "encoder_q-layer.11": 10025.2031, "encoder_q-layer.2": 2024.7821, "encoder_q-layer.3": 2106.9915, "encoder_q-layer.4": 2226.5085, "encoder_q-layer.5": 2256.9424, "encoder_q-layer.6": 2571.7112, "encoder_q-layer.7": 3005.7273, "encoder_q-layer.8": 3647.4001, "encoder_q-layer.9": 3394.0798, "epoch": 0.86, "inbatch_neg_score": 5.535, "inbatch_pos_score": 6.1367, "learning_rate": 6.052631578947369e-07, "loss": 3.3364, "norm_diff": 0.0842, "norm_loss": 0.0, "num_token_doc": 66.7503, "num_token_overlap": 11.6954, "num_token_query": 31.5033, "num_token_union": 65.1731, "num_word_context": 202.3418, "num_word_doc": 49.8064, "num_word_query": 23.3943, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6041.0758, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6399, "queue_k_norm": 2.7248, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.5033, "sent_len_1": 66.7503, "sent_len_max_0": 127.9887, "sent_len_max_1": 209.3113, "stdk": 0.0495, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197700 }, { "accuracy": 51.8555, "active_queue_size": 16384.0, "cl_loss": 3.3325, "doc_norm": 2.7225, "encoder_q-embeddings": 2403.7966, "encoder_q-layer.0": 1674.7539, "encoder_q-layer.1": 1812.3143, "encoder_q-layer.10": 3678.478, "encoder_q-layer.11": 8749.8701, "encoder_q-layer.2": 2042.8831, "encoder_q-layer.3": 2091.9951, "encoder_q-layer.4": 2265.7307, "encoder_q-layer.5": 2360.0991, "encoder_q-layer.6": 2616.5046, "encoder_q-layer.7": 2850.1162, "encoder_q-layer.8": 3555.9521, "encoder_q-layer.9": 3159.4849, "epoch": 0.86, "inbatch_neg_score": 5.536, "inbatch_pos_score": 6.1484, "learning_rate": 5.789473684210527e-07, "loss": 3.3325, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.8944, "num_token_overlap": 11.6939, "num_token_query": 31.3201, "num_token_union": 65.1284, "num_word_context": 202.4949, "num_word_doc": 49.911, "num_word_query": 23.2418, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5449.7544, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6407, "queue_k_norm": 2.7249, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3201, "sent_len_1": 66.8944, "sent_len_max_0": 127.9562, "sent_len_max_1": 209.245, "stdk": 0.0494, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 197800 }, { "accuracy": 51.2695, "active_queue_size": 16384.0, "cl_loss": 3.3372, "doc_norm": 2.7254, "encoder_q-embeddings": 2294.793, "encoder_q-layer.0": 1651.3131, "encoder_q-layer.1": 1808.6146, "encoder_q-layer.10": 3553.8345, "encoder_q-layer.11": 9216.1738, "encoder_q-layer.2": 1993.2616, "encoder_q-layer.3": 2089.6553, "encoder_q-layer.4": 2228.0442, "encoder_q-layer.5": 2310.6941, "encoder_q-layer.6": 2640.8579, "encoder_q-layer.7": 3081.9705, "encoder_q-layer.8": 3507.9221, "encoder_q-layer.9": 3284.4417, "epoch": 0.86, "inbatch_neg_score": 5.5363, "inbatch_pos_score": 6.1445, "learning_rate": 5.526315789473684e-07, "loss": 3.3372, "norm_diff": 0.0813, "norm_loss": 0.0, "num_token_doc": 66.8343, "num_token_overlap": 11.6732, "num_token_query": 31.3788, "num_token_union": 65.1728, "num_word_context": 202.4638, "num_word_doc": 49.8593, "num_word_query": 23.3077, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5620.941, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6441, "queue_k_norm": 2.7233, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3788, "sent_len_1": 66.8343, "sent_len_max_0": 127.9788, "sent_len_max_1": 208.7625, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 197900 }, { "accuracy": 49.1211, "active_queue_size": 16384.0, "cl_loss": 3.3412, "doc_norm": 2.725, "encoder_q-embeddings": 2323.7681, "encoder_q-layer.0": 1661.4911, "encoder_q-layer.1": 1793.4238, "encoder_q-layer.10": 4136.1709, "encoder_q-layer.11": 9480.0918, "encoder_q-layer.2": 1990.5126, "encoder_q-layer.3": 2019.969, "encoder_q-layer.4": 2232.9856, "encoder_q-layer.5": 2457.3105, "encoder_q-layer.6": 2730.7395, "encoder_q-layer.7": 3066.2163, "encoder_q-layer.8": 3706.7559, "encoder_q-layer.9": 3457.0244, "epoch": 0.86, "inbatch_neg_score": 5.5364, "inbatch_pos_score": 6.1445, "learning_rate": 5.263157894736843e-07, "loss": 3.3412, "norm_diff": 0.0818, "norm_loss": 0.0, "num_token_doc": 66.7537, "num_token_overlap": 11.6468, "num_token_query": 31.305, "num_token_union": 65.0493, "num_word_context": 202.4052, "num_word_doc": 49.8094, "num_word_query": 23.2548, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5825.3018, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6432, "queue_k_norm": 2.7244, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.305, "sent_len_1": 66.7537, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.1325, "stdk": 0.0496, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198000 }, { "accuracy": 52.5391, "active_queue_size": 16384.0, "cl_loss": 3.3334, "doc_norm": 2.7235, "encoder_q-embeddings": 2277.4927, "encoder_q-layer.0": 1655.7593, "encoder_q-layer.1": 1747.6826, "encoder_q-layer.10": 3410.1438, "encoder_q-layer.11": 8590.9854, "encoder_q-layer.2": 1952.793, "encoder_q-layer.3": 1966.2699, "encoder_q-layer.4": 2095.8481, "encoder_q-layer.5": 2154.071, "encoder_q-layer.6": 2319.301, "encoder_q-layer.7": 2733.5752, "encoder_q-layer.8": 3106.8911, "encoder_q-layer.9": 3016.6431, "epoch": 0.86, "inbatch_neg_score": 5.5371, "inbatch_pos_score": 6.1562, "learning_rate": 5.000000000000001e-07, "loss": 3.3334, "norm_diff": 0.0806, "norm_loss": 0.0, "num_token_doc": 66.8422, "num_token_overlap": 11.6966, "num_token_query": 31.3379, "num_token_union": 65.1236, "num_word_context": 202.5546, "num_word_doc": 49.8761, "num_word_query": 23.2728, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5329.6547, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5391, "query_norm": 2.6429, "queue_k_norm": 2.7248, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3379, "sent_len_1": 66.8422, "sent_len_max_0": 128.0, "sent_len_max_1": 210.455, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198100 }, { "accuracy": 50.1465, "active_queue_size": 16384.0, "cl_loss": 3.3334, "doc_norm": 2.7243, "encoder_q-embeddings": 2784.2239, "encoder_q-layer.0": 1994.2913, "encoder_q-layer.1": 2259.845, "encoder_q-layer.10": 3574.8093, "encoder_q-layer.11": 8860.7959, "encoder_q-layer.2": 2674.8103, "encoder_q-layer.3": 2988.781, "encoder_q-layer.4": 3337.5652, "encoder_q-layer.5": 3489.0381, "encoder_q-layer.6": 3665.6665, "encoder_q-layer.7": 3741.4468, "encoder_q-layer.8": 3864.3047, "encoder_q-layer.9": 3357.5039, "epoch": 0.86, "inbatch_neg_score": 5.5354, "inbatch_pos_score": 6.1523, "learning_rate": 4.7368421052631585e-07, "loss": 3.3334, "norm_diff": 0.0826, "norm_loss": 0.0, "num_token_doc": 66.744, "num_token_overlap": 11.6862, "num_token_query": 31.3926, "num_token_union": 65.087, "num_word_context": 202.1828, "num_word_doc": 49.7963, "num_word_query": 23.3166, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6072.1912, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6417, "queue_k_norm": 2.7238, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3926, "sent_len_1": 66.744, "sent_len_max_0": 127.9788, "sent_len_max_1": 210.625, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 198200 }, { "accuracy": 53.4668, "active_queue_size": 16384.0, "cl_loss": 3.33, "doc_norm": 2.7271, "encoder_q-embeddings": 2302.3796, "encoder_q-layer.0": 1600.3632, "encoder_q-layer.1": 1769.5081, "encoder_q-layer.10": 3421.8074, "encoder_q-layer.11": 9011.5996, "encoder_q-layer.2": 1922.7457, "encoder_q-layer.3": 1968.1863, "encoder_q-layer.4": 2118.5076, "encoder_q-layer.5": 2195.7842, "encoder_q-layer.6": 2467.6807, "encoder_q-layer.7": 2880.8035, "encoder_q-layer.8": 3551.3704, "encoder_q-layer.9": 3305.1101, "epoch": 0.86, "inbatch_neg_score": 5.5362, "inbatch_pos_score": 6.1641, "learning_rate": 4.4736842105263156e-07, "loss": 3.33, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.7563, "num_token_overlap": 11.7018, "num_token_query": 31.4578, "num_token_union": 65.1388, "num_word_context": 202.4689, "num_word_doc": 49.8091, "num_word_query": 23.3602, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5565.0283, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5352, "query_norm": 2.6435, "queue_k_norm": 2.7249, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4578, "sent_len_1": 66.7563, "sent_len_max_0": 127.9813, "sent_len_max_1": 207.7475, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198300 }, { "accuracy": 51.416, "active_queue_size": 16384.0, "cl_loss": 3.3318, "doc_norm": 2.7256, "encoder_q-embeddings": 2345.6472, "encoder_q-layer.0": 1631.0255, "encoder_q-layer.1": 1810.3365, "encoder_q-layer.10": 3668.6982, "encoder_q-layer.11": 9095.4033, "encoder_q-layer.2": 2000.8632, "encoder_q-layer.3": 2037.6725, "encoder_q-layer.4": 2166.4214, "encoder_q-layer.5": 2403.9436, "encoder_q-layer.6": 2818.7581, "encoder_q-layer.7": 3212.0615, "encoder_q-layer.8": 3737.7756, "encoder_q-layer.9": 3297.3062, "epoch": 0.86, "inbatch_neg_score": 5.5371, "inbatch_pos_score": 6.1484, "learning_rate": 4.210526315789474e-07, "loss": 3.3318, "norm_diff": 0.0858, "norm_loss": 0.0, "num_token_doc": 66.7821, "num_token_overlap": 11.6356, "num_token_query": 31.2958, "num_token_union": 65.1061, "num_word_context": 202.289, "num_word_doc": 49.8232, "num_word_query": 23.2489, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5702.2418, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6398, "queue_k_norm": 2.7252, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2958, "sent_len_1": 66.7821, "sent_len_max_0": 127.98, "sent_len_max_1": 209.9888, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198400 }, { "accuracy": 50.6348, "active_queue_size": 16384.0, "cl_loss": 3.3381, "doc_norm": 2.7233, "encoder_q-embeddings": 2233.7051, "encoder_q-layer.0": 1592.3142, "encoder_q-layer.1": 1720.2516, "encoder_q-layer.10": 4770.1313, "encoder_q-layer.11": 9327.3965, "encoder_q-layer.2": 1977.0533, "encoder_q-layer.3": 1976.1709, "encoder_q-layer.4": 2111.9712, "encoder_q-layer.5": 2153.2698, "encoder_q-layer.6": 2483.8167, "encoder_q-layer.7": 2840.2517, "encoder_q-layer.8": 3525.1548, "encoder_q-layer.9": 3406.4446, "epoch": 0.86, "inbatch_neg_score": 5.5372, "inbatch_pos_score": 6.1562, "learning_rate": 3.947368421052632e-07, "loss": 3.3381, "norm_diff": 0.0804, "norm_loss": 0.0, "num_token_doc": 66.8072, "num_token_overlap": 11.6558, "num_token_query": 31.3738, "num_token_union": 65.1382, "num_word_context": 202.3997, "num_word_doc": 49.8117, "num_word_query": 23.2955, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5648.6471, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6429, "queue_k_norm": 2.7247, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3738, "sent_len_1": 66.8072, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.8938, "stdk": 0.0494, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198500 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3346, "doc_norm": 2.7259, "encoder_q-embeddings": 2574.3555, "encoder_q-layer.0": 1794.6917, "encoder_q-layer.1": 1875.34, "encoder_q-layer.10": 3543.6772, "encoder_q-layer.11": 9262.3252, "encoder_q-layer.2": 2120.9487, "encoder_q-layer.3": 2147.3916, "encoder_q-layer.4": 2264.4697, "encoder_q-layer.5": 2382.0886, "encoder_q-layer.6": 2663.1172, "encoder_q-layer.7": 2910.3633, "encoder_q-layer.8": 3424.1797, "encoder_q-layer.9": 3283.9175, "epoch": 0.86, "inbatch_neg_score": 5.5413, "inbatch_pos_score": 6.1484, "learning_rate": 3.6842105263157896e-07, "loss": 3.3346, "norm_diff": 0.0836, "norm_loss": 0.0, "num_token_doc": 66.7967, "num_token_overlap": 11.6046, "num_token_query": 31.2306, "num_token_union": 65.1297, "num_word_context": 202.4739, "num_word_doc": 49.8537, "num_word_query": 23.1951, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5810.0495, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6423, "queue_k_norm": 2.7248, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2306, "sent_len_1": 66.7967, "sent_len_max_0": 127.98, "sent_len_max_1": 206.1838, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198600 }, { "accuracy": 52.0508, "active_queue_size": 16384.0, "cl_loss": 3.3234, "doc_norm": 2.7264, "encoder_q-embeddings": 2328.6982, "encoder_q-layer.0": 1656.4832, "encoder_q-layer.1": 1760.8618, "encoder_q-layer.10": 3276.7825, "encoder_q-layer.11": 8725.2227, "encoder_q-layer.2": 1927.5842, "encoder_q-layer.3": 1961.6178, "encoder_q-layer.4": 2112.4536, "encoder_q-layer.5": 2168.4937, "encoder_q-layer.6": 2462.7815, "encoder_q-layer.7": 2930.49, "encoder_q-layer.8": 3483.7192, "encoder_q-layer.9": 3176.4197, "epoch": 0.86, "inbatch_neg_score": 5.5374, "inbatch_pos_score": 6.1562, "learning_rate": 3.421052631578948e-07, "loss": 3.3234, "norm_diff": 0.0821, "norm_loss": 0.0, "num_token_doc": 66.8002, "num_token_overlap": 11.6944, "num_token_query": 31.4201, "num_token_union": 65.1102, "num_word_context": 202.2301, "num_word_doc": 49.8224, "num_word_query": 23.3295, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5424.7713, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5391, "query_norm": 2.6443, "queue_k_norm": 2.7251, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4201, "sent_len_1": 66.8002, "sent_len_max_0": 127.9737, "sent_len_max_1": 210.6825, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198700 }, { "accuracy": 49.9023, "active_queue_size": 16384.0, "cl_loss": 3.3242, "doc_norm": 2.724, "encoder_q-embeddings": 2503.9365, "encoder_q-layer.0": 1736.0193, "encoder_q-layer.1": 1871.7347, "encoder_q-layer.10": 3577.9629, "encoder_q-layer.11": 9203.6777, "encoder_q-layer.2": 2081.0737, "encoder_q-layer.3": 2152.5632, "encoder_q-layer.4": 2289.3179, "encoder_q-layer.5": 2276.1821, "encoder_q-layer.6": 2539.9941, "encoder_q-layer.7": 2855.4927, "encoder_q-layer.8": 3452.4302, "encoder_q-layer.9": 3194.9827, "epoch": 0.86, "inbatch_neg_score": 5.539, "inbatch_pos_score": 6.1367, "learning_rate": 3.1578947368421055e-07, "loss": 3.3242, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 66.7995, "num_token_overlap": 11.7006, "num_token_query": 31.3551, "num_token_union": 65.0891, "num_word_context": 202.1022, "num_word_doc": 49.8258, "num_word_query": 23.2939, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5604.6161, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6412, "queue_k_norm": 2.7258, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3551, "sent_len_1": 66.7995, "sent_len_max_0": 127.9838, "sent_len_max_1": 208.4825, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 198800 }, { "accuracy": 51.3672, "active_queue_size": 16384.0, "cl_loss": 3.3186, "doc_norm": 2.7275, "encoder_q-embeddings": 2375.083, "encoder_q-layer.0": 1678.9667, "encoder_q-layer.1": 1769.1715, "encoder_q-layer.10": 3780.3247, "encoder_q-layer.11": 9280.4404, "encoder_q-layer.2": 2017.1746, "encoder_q-layer.3": 2112.4785, "encoder_q-layer.4": 2187.4062, "encoder_q-layer.5": 2358.0479, "encoder_q-layer.6": 2661.0864, "encoder_q-layer.7": 2900.2725, "encoder_q-layer.8": 3526.1392, "encoder_q-layer.9": 3330.2998, "epoch": 0.86, "inbatch_neg_score": 5.5377, "inbatch_pos_score": 6.1641, "learning_rate": 2.8947368421052637e-07, "loss": 3.3186, "norm_diff": 0.0824, "norm_loss": 0.0, "num_token_doc": 66.8766, "num_token_overlap": 11.6783, "num_token_query": 31.3364, "num_token_union": 65.1669, "num_word_context": 202.5157, "num_word_doc": 49.91, "num_word_query": 23.2715, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5687.5197, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6451, "queue_k_norm": 2.726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3364, "sent_len_1": 66.8766, "sent_len_max_0": 127.99, "sent_len_max_1": 207.6763, "stdk": 0.0497, "stdq": 0.0434, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 198900 }, { "accuracy": 51.123, "active_queue_size": 16384.0, "cl_loss": 3.3187, "doc_norm": 2.7268, "encoder_q-embeddings": 2653.8484, "encoder_q-layer.0": 1841.3591, "encoder_q-layer.1": 2063.7568, "encoder_q-layer.10": 3955.8933, "encoder_q-layer.11": 9343.4727, "encoder_q-layer.2": 2391.6179, "encoder_q-layer.3": 2559.2771, "encoder_q-layer.4": 2660.3064, "encoder_q-layer.5": 2799.2144, "encoder_q-layer.6": 3045.7102, "encoder_q-layer.7": 3604.1101, "encoder_q-layer.8": 3897.4768, "encoder_q-layer.9": 3625.2192, "epoch": 0.86, "inbatch_neg_score": 5.539, "inbatch_pos_score": 6.1484, "learning_rate": 2.6315789473684213e-07, "loss": 3.3187, "norm_diff": 0.0834, "norm_loss": 0.0, "num_token_doc": 66.7591, "num_token_overlap": 11.6881, "num_token_query": 31.3664, "num_token_union": 65.1073, "num_word_context": 202.3255, "num_word_doc": 49.8063, "num_word_query": 23.2947, "postclip_grad_norm": 1.0, "preclip_grad_norm": 6025.0766, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6434, "queue_k_norm": 2.7245, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3664, "sent_len_1": 66.7591, "sent_len_max_0": 128.0, "sent_len_max_1": 210.4762, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 199000 }, { "accuracy": 51.4648, "active_queue_size": 16384.0, "cl_loss": 3.3401, "doc_norm": 2.7261, "encoder_q-embeddings": 2494.6777, "encoder_q-layer.0": 1747.5205, "encoder_q-layer.1": 1821.0815, "encoder_q-layer.10": 3368.8843, "encoder_q-layer.11": 8688.9727, "encoder_q-layer.2": 2056.5188, "encoder_q-layer.3": 2103.7808, "encoder_q-layer.4": 2233.1221, "encoder_q-layer.5": 2373.0496, "encoder_q-layer.6": 2690.2395, "encoder_q-layer.7": 3034.5803, "encoder_q-layer.8": 3474.7534, "encoder_q-layer.9": 3239.5156, "epoch": 0.86, "inbatch_neg_score": 5.5394, "inbatch_pos_score": 6.1562, "learning_rate": 2.3684210526315792e-07, "loss": 3.3401, "norm_diff": 0.0866, "norm_loss": 0.0, "num_token_doc": 66.8344, "num_token_overlap": 11.6642, "num_token_query": 31.4092, "num_token_union": 65.1927, "num_word_context": 202.5846, "num_word_doc": 49.8626, "num_word_query": 23.3185, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5509.0182, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6395, "queue_k_norm": 2.7263, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4092, "sent_len_1": 66.8344, "sent_len_max_0": 127.9912, "sent_len_max_1": 208.3413, "stdk": 0.0496, "stdq": 0.0429, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 199100 }, { "accuracy": 52.002, "active_queue_size": 16384.0, "cl_loss": 3.3324, "doc_norm": 2.725, "encoder_q-embeddings": 2243.8704, "encoder_q-layer.0": 1580.4078, "encoder_q-layer.1": 1677.2135, "encoder_q-layer.10": 3307.7009, "encoder_q-layer.11": 8719.0723, "encoder_q-layer.2": 1864.208, "encoder_q-layer.3": 1887.0275, "encoder_q-layer.4": 1997.7495, "encoder_q-layer.5": 2082.0139, "encoder_q-layer.6": 2317.5002, "encoder_q-layer.7": 2565.3125, "encoder_q-layer.8": 3178.8518, "encoder_q-layer.9": 3040.4617, "epoch": 0.86, "inbatch_neg_score": 5.5405, "inbatch_pos_score": 6.1641, "learning_rate": 2.105263157894737e-07, "loss": 3.3324, "norm_diff": 0.0827, "norm_loss": 0.0, "num_token_doc": 66.8728, "num_token_overlap": 11.7112, "num_token_query": 31.4206, "num_token_union": 65.1874, "num_word_context": 202.3746, "num_word_doc": 49.8953, "num_word_query": 23.331, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5292.629, "preclip_grad_norm_avg": 0.0, "q@queue_neg_score": 5.5391, "query_norm": 2.6423, "queue_k_norm": 2.726, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4206, "sent_len_1": 66.8728, "sent_len_max_0": 127.9925, "sent_len_max_1": 209.16, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199200 }, { "accuracy": 53.0762, "active_queue_size": 16384.0, "cl_loss": 3.3396, "doc_norm": 2.7271, "encoder_q-embeddings": 2364.4189, "encoder_q-layer.0": 1647.4143, "encoder_q-layer.1": 1745.2113, "encoder_q-layer.10": 3665.4932, "encoder_q-layer.11": 9353.8867, "encoder_q-layer.2": 1960.3953, "encoder_q-layer.3": 2047.4866, "encoder_q-layer.4": 2203.0066, "encoder_q-layer.5": 2237.7258, "encoder_q-layer.6": 2653.1331, "encoder_q-layer.7": 3086.0, "encoder_q-layer.8": 3677.9143, "encoder_q-layer.9": 3398.894, "epoch": 0.86, "inbatch_neg_score": 5.5402, "inbatch_pos_score": 6.1602, "learning_rate": 1.8421052631578948e-07, "loss": 3.3396, "norm_diff": 0.0837, "norm_loss": 0.0, "num_token_doc": 66.7763, "num_token_overlap": 11.6866, "num_token_query": 31.4209, "num_token_union": 65.1503, "num_word_context": 202.2703, "num_word_doc": 49.7842, "num_word_query": 23.3339, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5775.0312, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6434, "queue_k_norm": 2.7262, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.4209, "sent_len_1": 66.7763, "sent_len_max_0": 127.9825, "sent_len_max_1": 210.395, "stdk": 0.0497, "stdq": 0.0432, "stdqueue_k": 0.0497, "stdqueue_q": 0.0, "step": 199300 }, { "accuracy": 49.8535, "active_queue_size": 16384.0, "cl_loss": 3.326, "doc_norm": 2.7254, "encoder_q-embeddings": 2320.261, "encoder_q-layer.0": 1638.218, "encoder_q-layer.1": 1721.6093, "encoder_q-layer.10": 3623.1831, "encoder_q-layer.11": 8983.4688, "encoder_q-layer.2": 1915.7725, "encoder_q-layer.3": 1950.2048, "encoder_q-layer.4": 2172.3967, "encoder_q-layer.5": 2191.8972, "encoder_q-layer.6": 2409.9128, "encoder_q-layer.7": 2831.9194, "encoder_q-layer.8": 3495.3596, "encoder_q-layer.9": 3092.7493, "epoch": 0.87, "inbatch_neg_score": 5.5414, "inbatch_pos_score": 6.1406, "learning_rate": 1.5789473684210527e-07, "loss": 3.326, "norm_diff": 0.0835, "norm_loss": 0.0, "num_token_doc": 66.7205, "num_token_overlap": 11.7035, "num_token_query": 31.4793, "num_token_union": 65.136, "num_word_context": 202.4438, "num_word_doc": 49.7873, "num_word_query": 23.3981, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5547.5417, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6419, "queue_k_norm": 2.7262, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4793, "sent_len_1": 66.7205, "sent_len_max_0": 127.97, "sent_len_max_1": 207.7413, "stdk": 0.0495, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199400 }, { "accuracy": 51.7578, "active_queue_size": 16384.0, "cl_loss": 3.3443, "doc_norm": 2.7267, "encoder_q-embeddings": 2391.9973, "encoder_q-layer.0": 1668.5942, "encoder_q-layer.1": 1817.3557, "encoder_q-layer.10": 3425.4861, "encoder_q-layer.11": 9281.8936, "encoder_q-layer.2": 2036.2379, "encoder_q-layer.3": 2066.4902, "encoder_q-layer.4": 2219.7703, "encoder_q-layer.5": 2222.1729, "encoder_q-layer.6": 2482.8901, "encoder_q-layer.7": 2859.0029, "encoder_q-layer.8": 3493.6682, "encoder_q-layer.9": 3228.3464, "epoch": 0.87, "inbatch_neg_score": 5.5401, "inbatch_pos_score": 6.1562, "learning_rate": 1.3157894736842107e-07, "loss": 3.3443, "norm_diff": 0.086, "norm_loss": 0.0, "num_token_doc": 66.8174, "num_token_overlap": 11.6689, "num_token_query": 31.3351, "num_token_union": 65.1399, "num_word_context": 202.21, "num_word_doc": 49.8563, "num_word_query": 23.2598, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5720.9993, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6407, "queue_k_norm": 2.726, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3351, "sent_len_1": 66.8174, "sent_len_max_0": 127.9813, "sent_len_max_1": 209.1987, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199500 }, { "accuracy": 53.2715, "active_queue_size": 16384.0, "cl_loss": 3.3411, "doc_norm": 2.7279, "encoder_q-embeddings": 2275.011, "encoder_q-layer.0": 1641.5742, "encoder_q-layer.1": 1784.6804, "encoder_q-layer.10": 3569.2017, "encoder_q-layer.11": 9002.7451, "encoder_q-layer.2": 1978.3467, "encoder_q-layer.3": 2098.0444, "encoder_q-layer.4": 2168.9507, "encoder_q-layer.5": 2233.9961, "encoder_q-layer.6": 2507.3308, "encoder_q-layer.7": 2999.447, "encoder_q-layer.8": 3376.2515, "encoder_q-layer.9": 3205.2878, "epoch": 0.87, "inbatch_neg_score": 5.5383, "inbatch_pos_score": 6.1641, "learning_rate": 1.0526315789473685e-07, "loss": 3.3411, "norm_diff": 0.0839, "norm_loss": 0.0, "num_token_doc": 66.8456, "num_token_overlap": 11.681, "num_token_query": 31.3534, "num_token_union": 65.1194, "num_word_context": 202.3054, "num_word_doc": 49.8643, "num_word_query": 23.2705, "postclip_grad_norm": 1.0, "preclip_grad_norm": 5557.5485, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.644, "queue_k_norm": 2.725, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.3534, "sent_len_1": 66.8456, "sent_len_max_0": 127.985, "sent_len_max_1": 209.855, "stdk": 0.0497, "stdq": 0.0433, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199600 }, { "accuracy": 50.9766, "active_queue_size": 16384.0, "cl_loss": 3.3411, "doc_norm": 2.7251, "encoder_q-embeddings": 4762.0166, "encoder_q-layer.0": 3403.2549, "encoder_q-layer.1": 3612.3196, "encoder_q-layer.10": 7599.1611, "encoder_q-layer.11": 17754.2109, "encoder_q-layer.2": 4039.4946, "encoder_q-layer.3": 4230.9243, "encoder_q-layer.4": 4380.2319, "encoder_q-layer.5": 4510.0713, "encoder_q-layer.6": 5156.9497, "encoder_q-layer.7": 5843.8765, "encoder_q-layer.8": 6963.978, "encoder_q-layer.9": 6546.1455, "epoch": 0.87, "inbatch_neg_score": 5.5402, "inbatch_pos_score": 6.1523, "learning_rate": 7.894736842105264e-08, "loss": 3.3411, "norm_diff": 0.0814, "norm_loss": 0.0, "num_token_doc": 66.8085, "num_token_overlap": 11.6545, "num_token_query": 31.3394, "num_token_union": 65.1316, "num_word_context": 202.334, "num_word_doc": 49.882, "num_word_query": 23.2585, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11037.7011, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6437, "queue_k_norm": 2.7262, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3394, "sent_len_1": 66.8085, "sent_len_max_0": 127.9875, "sent_len_max_1": 207.7512, "stdk": 0.0495, "stdq": 0.0432, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199700 }, { "accuracy": 50.8301, "active_queue_size": 16384.0, "cl_loss": 3.3254, "doc_norm": 2.7262, "encoder_q-embeddings": 4866.3242, "encoder_q-layer.0": 3457.8577, "encoder_q-layer.1": 3662.9417, "encoder_q-layer.10": 7447.9932, "encoder_q-layer.11": 18559.291, "encoder_q-layer.2": 4162.1079, "encoder_q-layer.3": 4438.1177, "encoder_q-layer.4": 4705.3184, "encoder_q-layer.5": 4843.1006, "encoder_q-layer.6": 5363.3389, "encoder_q-layer.7": 6319.4644, "encoder_q-layer.8": 7331.0649, "encoder_q-layer.9": 6546.4888, "epoch": 0.87, "inbatch_neg_score": 5.5411, "inbatch_pos_score": 6.1523, "learning_rate": 5.263157894736842e-08, "loss": 3.3254, "norm_diff": 0.0851, "norm_loss": 0.0, "num_token_doc": 66.8667, "num_token_overlap": 11.7114, "num_token_query": 31.4001, "num_token_union": 65.1703, "num_word_context": 202.4246, "num_word_doc": 49.9352, "num_word_query": 23.3273, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11663.1099, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.5391, "query_norm": 2.6411, "queue_k_norm": 2.7256, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.4001, "sent_len_1": 66.8667, "sent_len_max_0": 127.99, "sent_len_max_1": 207.0488, "stdk": 0.0496, "stdq": 0.043, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 199800 }, { "accuracy": 51.8066, "active_queue_size": 16384.0, "cl_loss": 3.3389, "doc_norm": 2.7288, "encoder_q-embeddings": 5055.6025, "encoder_q-layer.0": 3520.6213, "encoder_q-layer.1": 3901.2642, "encoder_q-layer.10": 7153.8223, "encoder_q-layer.11": 17752.8535, "encoder_q-layer.2": 4383.4048, "encoder_q-layer.3": 4609.1069, "encoder_q-layer.4": 4921.8218, "encoder_q-layer.5": 5045.7056, "encoder_q-layer.6": 5471.0273, "encoder_q-layer.7": 6123.9346, "encoder_q-layer.8": 7027.9282, "encoder_q-layer.9": 6578.4336, "epoch": 0.87, "inbatch_neg_score": 5.5413, "inbatch_pos_score": 6.1641, "learning_rate": 2.631578947368421e-08, "loss": 3.3389, "norm_diff": 0.0853, "norm_loss": 0.0, "num_token_doc": 66.5359, "num_token_overlap": 11.6422, "num_token_query": 31.3392, "num_token_union": 64.9922, "num_word_context": 202.0804, "num_word_doc": 49.6532, "num_word_query": 23.3028, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11337.8289, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.543, "query_norm": 2.6435, "queue_k_norm": 2.7243, "queue_ptr": 8192.0, "queue_q_norm": 0.0, "sent_len_0": 31.3392, "sent_len_1": 66.5359, "sent_len_max_0": 127.9825, "sent_len_max_1": 209.2113, "stdk": 0.0498, "stdq": 0.0432, "stdqueue_k": 0.0495, "stdqueue_q": 0.0, "step": 199900 }, { "accuracy": 52.2461, "active_queue_size": 16384.0, "cl_loss": 3.3426, "doc_norm": 2.7269, "encoder_q-embeddings": 4947.875, "encoder_q-layer.0": 3475.3062, "encoder_q-layer.1": 3775.8652, "encoder_q-layer.10": 7358.0522, "encoder_q-layer.11": 17834.2832, "encoder_q-layer.2": 4273.0117, "encoder_q-layer.3": 4471.0601, "encoder_q-layer.4": 4592.1226, "encoder_q-layer.5": 4867.0552, "encoder_q-layer.6": 5191.208, "encoder_q-layer.7": 6038.9585, "encoder_q-layer.8": 6709.2456, "encoder_q-layer.9": 6247.6685, "epoch": 0.87, "inbatch_neg_score": 5.5422, "inbatch_pos_score": 6.1484, "learning_rate": 0.0, "loss": 3.3426, "norm_diff": 0.0851, "norm_loss": 0.0, "num_token_doc": 66.5385, "num_token_overlap": 11.6145, "num_token_query": 31.2047, "num_token_union": 64.9376, "num_word_context": 201.8669, "num_word_doc": 49.6701, "num_word_query": 23.1689, "postclip_grad_norm": 1.0, "preclip_grad_norm": 11044.7652, "preclip_grad_norm_avg": 0.0001, "q@queue_neg_score": 5.543, "query_norm": 2.6418, "queue_k_norm": 2.7255, "queue_ptr": 0.0, "queue_q_norm": 0.0, "sent_len_0": 31.2047, "sent_len_1": 66.5385, "sent_len_max_0": 127.965, "sent_len_max_1": 208.26, "stdk": 0.0496, "stdq": 0.0431, "stdqueue_k": 0.0496, "stdqueue_q": 0.0, "step": 200000 }, { "dev_runtime": 28.6662, "dev_samples_per_second": 1.116, "dev_steps_per_second": 0.035, "epoch": 0.87, "step": 200000, "test_accuracy": 94.32373046875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3210323452949524, "test_doc_norm": 2.710721015930176, "test_inbatch_neg_score": 5.826592445373535, "test_inbatch_pos_score": 6.861468315124512, "test_loss": 0.3210323452949524, "test_loss_align": 0.9824590682983398, "test_loss_unif": -55.07438278198242, "test_loss_unif_q@queue": -55.07438278198242, "test_norm_diff": 0.002459876239299774, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.531642436981201, "test_query_norm": 2.710784912109375, "test_queue_k_norm": 2.7254903316497803, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04428725689649582, "test_stdq": 0.04420718550682068, "test_stdqueue_k": 0.049578290432691574, "test_stdqueue_q": 0.0 }, { "dev_runtime": 28.6662, "dev_samples_per_second": 1.116, "dev_steps_per_second": 0.035, "epoch": 0.87, "eval_beir-arguana_ndcg@10": 0.41208, "eval_beir-arguana_recall@10": 0.69844, "eval_beir-arguana_recall@100": 0.94808, "eval_beir-arguana_recall@20": 0.81437, "eval_beir-avg_ndcg@10": 0.40413475, "eval_beir-avg_recall@10": 0.4766990833333334, "eval_beir-avg_recall@100": 0.6543450833333333, "eval_beir-avg_recall@20": 0.538187, "eval_beir-cqadupstack_ndcg@10": 0.29382749999999996, "eval_beir-cqadupstack_recall@10": 0.3953508333333333, "eval_beir-cqadupstack_recall@100": 0.6354808333333334, "eval_beir-cqadupstack_recall@20": 0.46499000000000007, "eval_beir-fiqa_ndcg@10": 0.26892, "eval_beir-fiqa_recall@10": 0.33482, "eval_beir-fiqa_recall@100": 0.61544, "eval_beir-fiqa_recall@20": 0.41326, "eval_beir-nfcorpus_ndcg@10": 0.32723, "eval_beir-nfcorpus_recall@10": 0.16013, "eval_beir-nfcorpus_recall@100": 0.31693, "eval_beir-nfcorpus_recall@20": 0.20658, "eval_beir-nq_ndcg@10": 0.32677, "eval_beir-nq_recall@10": 0.52134, "eval_beir-nq_recall@100": 0.84849, "eval_beir-nq_recall@20": 0.64569, "eval_beir-quora_ndcg@10": 0.79323, "eval_beir-quora_recall@10": 0.8979, "eval_beir-quora_recall@100": 0.98182, "eval_beir-quora_recall@20": 0.93776, "eval_beir-scidocs_ndcg@10": 0.16768, "eval_beir-scidocs_recall@10": 0.17673, "eval_beir-scidocs_recall@100": 0.3918, "eval_beir-scidocs_recall@20": 0.2363, "eval_beir-scifact_ndcg@10": 0.67807, "eval_beir-scifact_recall@10": 0.804, "eval_beir-scifact_recall@100": 0.92989, "eval_beir-scifact_recall@20": 0.84811, "eval_beir-trec-covid_ndcg@10": 0.5952, "eval_beir-trec-covid_recall@10": 0.642, "eval_beir-trec-covid_recall@100": 0.4574, "eval_beir-trec-covid_recall@20": 0.6, "eval_beir-webis-touche2020_ndcg@10": 0.17834, "eval_beir-webis-touche2020_recall@10": 0.13628, "eval_beir-webis-touche2020_recall@100": 0.41812, "eval_beir-webis-touche2020_recall@20": 0.21481, "eval_senteval-avg_sts": 0.7316156394202711, "eval_senteval-sickr_spearman": 0.7253075282545358, "eval_senteval-stsb_spearman": 0.7379237505860065, "step": 200000, "test_accuracy": 94.32373046875, "test_active_queue_size": 16384.0, "test_cl_loss": 0.3210323452949524, "test_doc_norm": 2.710721015930176, "test_inbatch_neg_score": 5.826592445373535, "test_inbatch_pos_score": 6.861468315124512, "test_loss": 0.3210323452949524, "test_loss_align": 0.9824590682983398, "test_loss_unif": -55.07438278198242, "test_loss_unif_q@queue": -55.07438278198242, "test_norm_diff": 0.002459876239299774, "test_norm_loss": 0.0, "test_q@queue_neg_score": 5.531642436981201, "test_query_norm": 2.710784912109375, "test_queue_k_norm": 2.7254903316497803, "test_queue_ptr": 0.0, "test_queue_q_norm": 0.0, "test_stdk": 0.04428725689649582, "test_stdq": 0.04420718550682068, "test_stdqueue_k": 0.049578290432691574, "test_stdqueue_q": 0.0 }, { "epoch": 0.87, "step": 200000, "total_flos": 0, "train_runtime": 510641.9961, "train_samples_per_second": 0.392 } ], "max_steps": 200000, "num_train_epochs": 1, "total_flos": 0, "trial_name": null, "trial_params": null }