| | import os |
| | import json |
| | from huggingface_hub import HfApi |
| | import glob |
| | from datetime import datetime |
| | from datasets import Dataset |
| |
|
| | TOKEN = os.environ.get("HF_WRITE_TOKEN") |
| | API = HfApi(token=TOKEN) |
| | REPO_ID = "AIEnergyScore/results_debug" |
| | UPLOAD_REPO_ID = 'meg/HUGS_energy' |
| |
|
| | output_directory = API.snapshot_download(repo_id=REPO_ID, repo_type='dataset') |
| | print(output_directory) |
| | |
| | |
| | dataset_results = [] |
| | for task in ['text_generation']: |
| | org_dirs = glob.glob(f"{output_directory}/{task}/*") |
| | print(org_dirs) |
| | for org_dir in org_dirs: |
| | org = org_dir.split("/")[-1] |
| | model_dirs = glob.glob(f"{org_dir}/*") |
| | print(model_dirs) |
| | for model_dir in model_dirs: |
| | model = model_dir.split("/")[-1] |
| | model_runs = glob.glob(f"{model_dir}/*") |
| | dates = [dir.split("/")[-1] for dir in model_runs] |
| | try: |
| | |
| | sorted_dates = sorted( |
| | [datetime.strptime(date, '%Y-%m-%d-%H-%M-%S') for date in |
| | dates]) |
| | |
| | sorted_dates_str = [date.strftime('%Y-%m-%d-%H-%M-%S') for date in |
| | sorted_dates] |
| | last_date = sorted_dates_str[-1] |
| | most_recent_run = f"{model_dir}/{last_date}" |
| | print(most_recent_run) |
| | try: |
| | benchmark_report = json.loads(open(f"{most_recent_run}/benchmark_report.json", "rb+").read()) |
| | print(benchmark_report) |
| | prefill_data = benchmark_report['prefill'] |
| | prefill_energy = prefill_data['energy'] |
| | prefill_efficiency = prefill_data['efficiency'] |
| | decode_data = benchmark_report['decode'] |
| | decode_energy = decode_data['energy'] |
| | decode_efficiency = decode_data['efficiency'] |
| | preprocess_data = benchmark_report['preprocess'] |
| | preprocess_energy = preprocess_data['energy'] |
| | preprocess_efficiency = preprocess_data['efficiency'] |
| | dataset_results += [{'task':task, 'org':org, 'model':model, 'hardware':'a10g-large', |
| | 'date':last_date, 'prefill':{'energy':prefill_energy, |
| | 'efficency':prefill_efficiency}, |
| | 'decode':{'energy':decode_energy, 'efficiency':decode_efficiency}, |
| | 'preprocess': {'energy':preprocess_energy, 'efficiency': preprocess_efficiency}},] |
| |
|
| | except FileNotFoundError: |
| | error_report = open(f"{most_recent_run}/error.log", "rb+").read() |
| | print(error_report) |
| | except ValueError: |
| | |
| | continue |
| |
|
| | hub_dataset_results = Dataset.from_list(dataset_results) |
| | print(hub_dataset_results) |
| | hub_dataset_results.push_to_hub(UPLOAD_REPO_ID, token=TOKEN) |