|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Baseline: MMLU + GSM8K.""" |
|
|
|
|
|
import os, subprocess, glob |
|
|
|
|
|
def main(): |
|
|
hf_token = os.getenv("HF_TOKEN") |
|
|
if hf_token: |
|
|
os.environ.setdefault("HUGGING_FACE_HUB_TOKEN", hf_token) |
|
|
os.environ.setdefault("HF_HUB_TOKEN", hf_token) |
|
|
os.environ["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True" |
|
|
|
|
|
model_args = "model_name=LiquidAI/LFM2.5-1.2B-Instruct,trust_remote_code=True,dtype=float16,max_length=2048" |
|
|
tasks = "leaderboard|mmlu:abstract_algebra|5,leaderboard|mmlu:anatomy|5,leaderboard|mmlu:astronomy|5,leaderboard|mmlu:business_ethics|5,leaderboard|mmlu:clinical_knowledge|5,leaderboard|gsm8k|5" |
|
|
|
|
|
cmd = ["lighteval", "accelerate", model_args, tasks, "--output-dir", "/tmp/results"] |
|
|
print(f"Running: {' '.join(cmd)}") |
|
|
subprocess.run(cmd, check=True) |
|
|
print("DONE") |
|
|
|
|
|
for f in glob.glob("/tmp/results/**/*.json", recursive=True): |
|
|
print(f"\n=== {f} ===") |
|
|
with open(f) as fh: |
|
|
print(fh.read()[:10000]) |
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|