assignment2-artifacts / part1_dare_results.json
Pritish92's picture
Upload Assignment 2 artifacts
8f75784 verified
{
"results": [
{
"drop_rate_p": 0.1,
"density": 0.9,
"checkpoint_kept": true,
"rougeL": 0.4970292890659862,
"meteor": 0.5402472477605571,
"bleu": 49.617399302698296,
"exact_match": 0.4169941060903733,
"mean_text_metric": 16.88489194650828,
"model_path": "/root/SafeGenAI/work/models/model_sft_dare"
},
{
"drop_rate_p": 0.3,
"density": 0.7,
"checkpoint_kept": true,
"rougeL": 0.49710649238976634,
"meteor": 0.5407666616360554,
"bleu": 49.78977109224581,
"exact_match": 0.4199410609037328,
"mean_text_metric": 16.942548082090543,
"model_path": "/root/SafeGenAI/work/models/model_sft_dare"
},
{
"drop_rate_p": 0.5,
"density": 0.5,
"checkpoint_kept": true,
"rougeL": 0.49744879050858737,
"meteor": 0.5403587306686749,
"bleu": 49.815553566088504,
"exact_match": 0.41895874263261296,
"mean_text_metric": 16.95112036242192,
"model_path": "/root/SafeGenAI/work/models/model_sft_dare"
},
{
"drop_rate_p": 0.7,
"density": 0.3,
"checkpoint_kept": false,
"rougeL": 0.494856579197801,
"meteor": 0.5383071027351187,
"bleu": 49.157171598741805,
"exact_match": 0.4155206286836935,
"mean_text_metric": 16.730111760224908,
"model_path": "temporary merge output removed after evaluation"
}
],
"selected": {
"drop_rate_p": 0.5,
"density": 0.5,
"checkpoint_kept": true,
"rougeL": 0.49744879050858737,
"meteor": 0.5403587306686749,
"bleu": 49.815553566088504,
"exact_match": 0.41895874263261296,
"mean_text_metric": 16.95112036242192,
"model_path": "/root/SafeGenAI/work/models/model_sft_dare"
},
"best_model_path": "/root/SafeGenAI/work/models/model_sft_dare",
"storage_note": "Only the best DARE checkpoint is kept on disk; temporary candidates are deleted after validation."
}