{ "_name_or_path": "distributed/llama-1b", "all_reduce_scores": { "0": "NOT_ALIVE", "1": "NON_PARTICIPATING", "10": "NOT_ALIVE", "100": "NOT_ALIVE", "101": "NOT_ALIVE", "102": "NOT_ALIVE", "103": "NON_PARTICIPATING", "104": "NON_PARTICIPATING", "105": "FAIL", "106": "NOT_ALIVE", "107": "FAIL", "108": "NOT_ALIVE", "109": "NON_PARTICIPATING", "11": "NON_PARTICIPATING", "110": "FAIL", "111": "NON_PARTICIPATING", "112": "NON_PARTICIPATING", "113": "NON_PARTICIPATING", "114": "NON_PARTICIPATING", "115": "NON_PARTICIPATING", "116": "NON_PARTICIPATING", "117": "FAIL", "118": "FAIL", "119": "NON_PARTICIPATING", "12": "NOT_ALIVE", "120": "NON_PARTICIPATING", "121": "NOT_ALIVE", "122": "NON_PARTICIPATING", "123": "FAIL", "124": "NON_PARTICIPATING", "125": "FAIL", "126": "NOT_ALIVE", "127": "NON_PARTICIPATING", "128": "NON_PARTICIPATING", "129": "NON_PARTICIPATING", "13": "NON_PARTICIPATING", "130": "NON_PARTICIPATING", "131": "NOT_ALIVE", "132": "NOT_ALIVE", "133": "NON_PARTICIPATING", "134": "NON_PARTICIPATING", "135": "NON_PARTICIPATING", "136": "NOT_ALIVE", "137": "NOT_ALIVE", "138": "NON_PARTICIPATING", "139": "NOT_ALIVE", "14": "NOT_ALIVE", "140": "NOT_ALIVE", "141": "NON_PARTICIPATING", "142": "NOT_ALIVE", "143": "FAIL", "144": "NON_PARTICIPATING", "145": "NON_PARTICIPATING", "146": "FAIL", "147": "FAIL", "148": "NON_PARTICIPATING", "149": "NOT_ALIVE", "15": "NON_PARTICIPATING", "150": "NON_PARTICIPATING", "151": "NOT_ALIVE", "152": "NOT_ALIVE", "153": "NOT_ALIVE", "154": "NON_PARTICIPATING", "155": "NON_PARTICIPATING", "156": "NON_PARTICIPATING", "157": "NOT_ALIVE", "158": "NON_PARTICIPATING", "159": "FAIL", "16": "NOT_ALIVE", "160": "NOT_ALIVE", "161": "NOT_ALIVE", "162": "NON_PARTICIPATING", "163": "NOT_ALIVE", "164": "NOT_ALIVE", "165": "FAIL", "166": "NON_PARTICIPATING", "167": "NON_PARTICIPATING", "168": "FAIL", "169": "NON_PARTICIPATING", "17": "NON_PARTICIPATING", "170": "FAIL", "171": "NON_PARTICIPATING", "172": "NOT_ALIVE", "173": "FAIL", "174": "NOT_ALIVE", "175": "FAIL", "176": "NOT_ALIVE", "177": "NOT_ALIVE", "178": "NOT_ALIVE", "179": "NOT_ALIVE", "18": "NON_PARTICIPATING", "180": "FAIL", "181": "FAIL", "182": "FAIL", "183": "FAIL", "184": "NON_PARTICIPATING", "185": "FAIL", "186": "NON_PARTICIPATING", "187": "NOT_ALIVE", "188": "NON_PARTICIPATING", "189": "NON_PARTICIPATING", "19": "NON_PARTICIPATING", "190": "NON_PARTICIPATING", "191": "NOT_ALIVE", "192": "NON_PARTICIPATING", "193": "NOT_ALIVE", "194": "NON_PARTICIPATING", "195": "NON_PARTICIPATING", "196": "NOT_ALIVE", "197": "FAIL", "198": "NON_PARTICIPATING", "199": "NON_PARTICIPATING", "2": "NON_PARTICIPATING", "20": "FAIL", "200": "NON_PARTICIPATING", "201": "NOT_ALIVE", "202": "NON_PARTICIPATING", "203": "FAIL", "204": "NOT_ALIVE", "205": "FAIL", "206": "FAIL", "207": "NOT_ALIVE", "208": "NON_PARTICIPATING", "209": "FAIL", "21": "NOT_ALIVE", "210": "NON_PARTICIPATING", "211": "FAIL", "212": "NON_PARTICIPATING", "213": "NON_PARTICIPATING", "214": "FAIL", "215": "NON_PARTICIPATING", "216": "FAIL", "217": "FAIL", "218": "NON_PARTICIPATING", "219": "NOT_ALIVE", "22": "NOT_ALIVE", "220": "NON_PARTICIPATING", "221": "NON_PARTICIPATING", "222": "FAIL", "223": "FAIL", "224": "NOT_ALIVE", "225": "FAIL", "226": "NON_PARTICIPATING", "227": "NOT_ALIVE", "228": "FAIL", "229": "NOT_ALIVE", "23": "NON_PARTICIPATING", "230": "NOT_ALIVE", "231": "NOT_ALIVE", "232": "FAIL", "233": "NON_PARTICIPATING", "234": "NON_PARTICIPATING", "235": "NON_PARTICIPATING", "236": "NOT_ALIVE", "237": "NON_PARTICIPATING", "238": "FAIL", "239": "NON_PARTICIPATING", "24": "FAIL", "240": "NON_PARTICIPATING", "241": "FAIL", "242": "NOT_ALIVE", "243": "NON_PARTICIPATING", "244": "NOT_ALIVE", "245": "NON_PARTICIPATING", "246": "NON_PARTICIPATING", "247": "NOT_ALIVE", "248": "NOT_ALIVE", "249": "NON_PARTICIPATING", "25": "SUCCESS", "250": "NON_PARTICIPATING", "251": "NON_PARTICIPATING", "252": "NON_PARTICIPATING", "253": "NON_PARTICIPATING", "254": "NON_PARTICIPATING", "255": "FAIL", "26": "NOT_ALIVE", "27": "FAIL", "28": "NON_PARTICIPATING", "29": "NON_PARTICIPATING", "3": "NON_PARTICIPATING", "30": "NON_PARTICIPATING", "31": "NON_PARTICIPATING", "32": "NOT_ALIVE", "33": "NOT_ALIVE", "34": "NON_PARTICIPATING", "35": "FAIL", "36": "NON_PARTICIPATING", "37": "NOT_ALIVE", "38": "FAIL", "39": "FAIL", "4": "FAIL", "40": "FAIL", "41": "NOT_ALIVE", "42": "FAIL", "43": "NOT_ALIVE", "44": "NON_PARTICIPATING", "45": "NOT_ALIVE", "46": "FAIL", "47": "NON_PARTICIPATING", "48": "NON_PARTICIPATING", "49": "NON_PARTICIPATING", "5": "NOT_ALIVE", "50": "NON_PARTICIPATING", "51": "FAIL", "52": "NON_PARTICIPATING", "53": "NON_PARTICIPATING", "54": "NOT_ALIVE", "55": "FAIL", "56": "NON_PARTICIPATING", "57": "NON_PARTICIPATING", "58": "NOT_ALIVE", "59": "NON_PARTICIPATING", "6": "NON_PARTICIPATING", "60": "NOT_ALIVE", "61": "FAIL", "62": "NON_PARTICIPATING", "63": "NOT_ALIVE", "64": "NON_PARTICIPATING", "65": "NOT_ALIVE", "66": "NON_PARTICIPATING", "67": "NON_PARTICIPATING", "68": "NON_PARTICIPATING", "69": "NOT_ALIVE", "7": "NON_PARTICIPATING", "70": "NON_PARTICIPATING", "71": "NOT_ALIVE", "72": "NON_PARTICIPATING", "73": "FAIL", "74": "FAIL", "75": "NON_PARTICIPATING", "76": "NON_PARTICIPATING", "77": "NOT_ALIVE", "78": "NON_PARTICIPATING", "79": "NON_PARTICIPATING", "8": "NOT_ALIVE", "80": "NOT_ALIVE", "81": "NOT_ALIVE", "82": "NON_PARTICIPATING", "83": "NON_PARTICIPATING", "84": "NON_PARTICIPATING", "85": "NON_PARTICIPATING", "86": "NON_PARTICIPATING", "87": "FAIL", "88": "NON_PARTICIPATING", "89": "NOT_ALIVE", "9": "NON_PARTICIPATING", "90": "NON_PARTICIPATING", "91": "NON_PARTICIPATING", "92": "FAIL", "93": "NON_PARTICIPATING", "94": "NON_PARTICIPATING", "95": "NOT_ALIVE", "96": "NON_PARTICIPATING", "97": "FAIL", "98": "FAIL", "99": "NON_PARTICIPATING" }, "architectures": [ "LlamaForCausalLM" ], "attention_bias": false, "attention_dropout": 0.0, "block_list": [ 6454519, 6454547 ], "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "inner_step": 72, "intermediate_size": 5632, "last_allreduce_block": 6453451, "max_position_embeddings": 2048, "mlp_bias": false, "model_type": "llama", "num_attention_heads": 32, "num_hidden_layers": 22, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "transformers_version": "4.39.3", "use_cache": false, "vocab_size": 32000 }