lora-llama / all_results.json
Mvp-24's picture
Upload folder using huggingface_hub
b12f229 verified
{
"epoch": 1.0,
"eval_accuracy": 0.9316016518156041,
"max_memory_allocated (GB)": 94.41,
"memory_allocated (GB)": 47.88,
"total_flos": 2.48148778622976e+16,
"total_memory_available (GB)": 94.62,
"train_loss": 3.1766091346740724,
"train_runtime": 857.258,
"train_samples_per_second": 0.933,
"train_steps_per_second": 0.058
}