File size: 8,559 Bytes
de561c0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
#!/usr/bin/env bash
# run_benchmark.sh β€” run all llama.cpp benchmarks for 3 LLMs Γ— 4 quants
#
# Produces per-combination result files in the same directory as this script:
#   {prefix}_bench.json    β€” llama-bench throughput + model size
#   {prefix}_vram.log      β€” nvidia-smi dmon (raw, parsed later)
#   {prefix}_ttft.json     β€” TTFT / TPOT from llama-server stream
#   {prefix}_ppl.txt       β€” perplexity on wikitext-2
#   {prefix}_hellaswag.txt β€” HellaSwag accuracy (400 tasks)
#   {prefix}_winogrande.txtβ€” Winogrande accuracy (all tasks)
#
# Run with:  bash results/run_benchmark.sh
# Skip already-done runs: set SKIP_EXISTING=1 (default)
# ─────────────────────────────────────────────────────────────────────────────
set -euo pipefail

SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
ROOT="$(cd "$SCRIPT_DIR/.." && pwd)"

BENCH="$ROOT/build/bin/llama-bench"
PPL="$ROOT/build/bin/llama-perplexity"

WIKI="$ROOT/wikitext-2-raw/wiki.test.raw"
HELLASWAG="$ROOT/hellaswag_val_400.txt"
WINOGRANDE="$ROOT/winogrande-debiased-eval.csv"
TTFT_SCRIPT="$SCRIPT_DIR/measure_ttft.py"

SKIP_EXISTING="${SKIP_EXISTING:-1}"    # set to 0 to re-run everything
SKIP_WINOGRANDE="${SKIP_WINOGRANDE:-0}" # set to 1 to skip Winogrande entirely
DEVICE="${DEVICE:-CUDA0}"              # single GPU to use (CUDA0 = GPU 0)
GPU_IDX=0                              # nvidia-smi GPU index matching DEVICE
NGL=99                                 # GPU layers
PPL_CTX=512                            # context window for perplexity
PPL_CHUNKS=400                         # chunks to evaluate (~488 total); all chunks for full accuracy
PPL_BATCH=8192                         # n_batch for perplexity: n_seq = PPL_BATCH/PPL_CTX
                                       # 8192/512 = 16 chunks processed in parallel per pass
HELLASWAG_TASKS=400                    # subset file already contains exactly 400 tasks

# ── Model paths ───────────────────────────────────────────────────────────────
HF_LLAMA="/home/user1/.cache/huggingface/hub/models--bartowski--Meta-Llama-3.1-8B-Instruct-GGUF/snapshots/bf5b95e96dac0462e2a09145ec66cae9a3f12067"
HF_QWEN="/home/user1/.cache/huggingface/hub/models--bartowski--Qwen2.5-7B-Instruct-GGUF/snapshots/8911e8a47f92bac19d6f5c64a2e2095bd2f7d031"
HF_GEMMA="/home/user1/.cache/huggingface/hub/models--bartowski--gemma-2-9b-it-GGUF/snapshots/d731033f3dc4018261fd39896e50984d398b4ac5"

declare -A MODEL   # MODEL[key] = path
MODEL[llama-f16]="$ROOT/models/llama-3.1-8b-instruct-f16.gguf"
MODEL[llama-Q8_0]="$HF_LLAMA/Meta-Llama-3.1-8B-Instruct-Q8_0.gguf"
MODEL[llama-Q4_K_M]="$ROOT/models/llama-3.1-8b-instruct-Q4_K_M.gguf"
MODEL[llama-Q2_K]="$HF_LLAMA/Meta-Llama-3.1-8B-Instruct-Q2_K.gguf"
MODEL[qwen-f16]="$HF_QWEN/Qwen2.5-7B-Instruct-f16.gguf"
MODEL[qwen-Q8_0]="$HF_QWEN/Qwen2.5-7B-Instruct-Q8_0.gguf"
MODEL[qwen-Q4_K_M]="$HF_QWEN/Qwen2.5-7B-Instruct-Q4_K_M.gguf"
MODEL[qwen-Q2_K]="$HF_QWEN/Qwen2.5-7B-Instruct-Q2_K.gguf"
MODEL[gemma-f16]="$ROOT/models/gemma-2-9b-it-f16.gguf"
MODEL[gemma-Q8_0]="$ROOT/models/gemma-2-9b-it-Q8_0.gguf"
MODEL[gemma-Q4_K_M]="$ROOT/models/gemma-2-9b-it-Q4_K_M.gguf"
MODEL[gemma-Q2_K]="$HF_GEMMA/gemma-2-9b-it-Q2_K.gguf"
# Result file prefix for each key (matches existing naming convention)
declare -A PREFIX
PREFIX[llama-f16]="llama-3.1-8b-instruct-f16"
PREFIX[llama-Q8_0]="llama-3.1-8b-instruct-Q8_0"
PREFIX[llama-Q4_K_M]="llama-3.1-8b-instruct-Q4_K_M"
PREFIX[llama-Q2_K]="llama-3.1-8b-instruct-Q2_K"
PREFIX[qwen-f16]="qwen2.5-7b-instruct-f16"
PREFIX[qwen-Q8_0]="qwen2.5-7b-instruct-Q8_0"
PREFIX[qwen-Q4_K_M]="qwen2.5-7b-instruct-Q4_K_M"
PREFIX[qwen-Q2_K]="qwen2.5-7b-instruct-Q2_K"
PREFIX[gemma-f16]="gemma-2-9b-it-f16"
PREFIX[gemma-Q8_0]="gemma-2-9b-it-Q8_0"
PREFIX[gemma-Q4_K_M]="gemma-2-9b-it-Q4_K_M"
PREFIX[gemma-Q2_K]="gemma-2-9b-it-Q2_K"
KEYS=(
    llama-f16  llama-Q8_0  llama-Q4_K_M  llama-Q2_K
    qwen-f16   qwen-Q8_0   qwen-Q4_K_M   qwen-Q2_K
    gemma-f16  gemma-Q8_0  gemma-Q4_K_M  gemma-Q2_K
)

# ── Helpers ───────────────────────────────────────────────────────────────────
log()  { echo "[$(date '+%H:%M:%S')] $*"; }
skip() { [[ "$SKIP_EXISTING" == "1" && -s "$1" ]]; }

# ── Main loop ─────────────────────────────────────────────────────────────────
for KEY in "${KEYS[@]}"; do
    MODEL_PATH="${MODEL[$KEY]}"
    PFX="$SCRIPT_DIR/${PREFIX[$KEY]}"

    if [[ ! -f "$MODEL_PATH" ]]; then
        echo "WARNING: model not found, skipping $KEY: $MODEL_PATH"
        continue
    fi

    log "═══ $KEY ═══"

    # ── 1. llama-bench (throughput + model size) with VRAM monitoring ─────────
    BENCH_OUT="$PFX"_bench.json
    VRAM_OUT="$PFX"_vram.log
    if skip "$BENCH_OUT"; then
        log "  [bench] skipping (exists)"
    else
        log "  [bench] starting nvidia-smi dmon (GPU $GPU_IDX only)..."
        nvidia-smi dmon -s m -d 1 -i "$GPU_IDX" > "$VRAM_OUT" &
        DMON_PID=$!

        log "  [bench] running llama-bench (single GPU: $DEVICE)..."
        "$BENCH" \
            -m "$MODEL_PATH" \
            -ngl "$NGL" \
            -dev "$DEVICE" \
            -p 512 -n 128 -r 3 \
            -o json \
            > "$BENCH_OUT"

        kill "$DMON_PID" 2>/dev/null && wait "$DMON_PID" 2>/dev/null || true
        log "  [bench] done β†’ $BENCH_OUT"
    fi

    # ── 2. TTFT / TPOT via llama-server ──────────────────────────────────────
    TTFT_OUT="$PFX"_ttft.json
    if skip "$TTFT_OUT"; then
        log "  [ttft] skipping (exists)"
    else
        log "  [ttft] running measure_ttft.py (single GPU: $DEVICE)..."
        python3 "$TTFT_SCRIPT" -m "$MODEL_PATH" --device "$DEVICE" \
            > "$TTFT_OUT" \
            2> "$PFX"_ttft.log
        log "  [ttft] done β†’ $TTFT_OUT"
    fi

    # ── 3. Perplexity (wikitext-2, fixed context for cross-model comparability)
    PPL_OUT="$PFX"_ppl.txt
    if skip "$PPL_OUT"; then
        log "  [ppl] skipping (exists)"
    else
        log "  [ppl] running llama-perplexity (ctx=$PPL_CTX, chunks=$PPL_CHUNKS, batch=$PPL_BATCH, GPU: $DEVICE)..."
        "$PPL" \
            -m "$MODEL_PATH" \
            -ngl "$NGL" \
            -dev "$DEVICE" \
            -f "$WIKI" \
            -c "$PPL_CTX" \
            -b "$PPL_BATCH" \
            --chunks "$PPL_CHUNKS" \
            2>&1 | grep "^Final estimate" | tail -1 > "$PPL_OUT"
        log "  [ppl] done β†’ $PPL_OUT"
    fi

    # ── 4. HellaSwag accuracy ─────────────────────────────────────────────────
    HS_OUT="$PFX"_hellaswag.txt
    if skip "$HS_OUT"; then
        log "  [hellaswag] skipping (exists)"
    else
        log "  [hellaswag] running ($HELLASWAG_TASKS tasks, GPU: $DEVICE)..."
        "$PPL" \
            -m "$MODEL_PATH" \
            -ngl "$NGL" \
            -dev "$DEVICE" \
            -f "$HELLASWAG" \
            --hellaswag \
            --hellaswag-tasks "$HELLASWAG_TASKS" \
            2>&1 | grep -E "^[0-9]+[[:space:]]" | tail -1 > "$HS_OUT"
        log "  [hellaswag] done β†’ $HS_OUT"
    fi

    # ── 5. Winogrande accuracy ────────────────────────────────────────────────
    WG_OUT="$PFX"_winogrande.txt
    if [[ "$SKIP_WINOGRANDE" == "1" ]]; then
        log "  [winogrande] skipped (SKIP_WINOGRANDE=1)"
    elif skip "$WG_OUT"; then
        log "  [winogrande] skipping (exists)"
    else
        log "  [winogrande] running (GPU: $DEVICE)..."
        "$PPL" \
            -m "$MODEL_PATH" \
            -ngl "$NGL" \
            -dev "$DEVICE" \
            -f "$WINOGRANDE" \
            --winogrande \
            2>&1 | grep -E "^[0-9]+[[:space:]]" | tail -1 > "$WG_OUT"
        log "  [winogrande] done β†’ $WG_OUT"
    fi

    log "  βœ“ $KEY complete"
done

log "All benchmarks done. Run: python3 $SCRIPT_DIR/parse_results.py"