llama-cpp-bench / measure_ttft.py
bobchenyx's picture
Upload folder using huggingface_hub
de561c0 verified
#!/usr/bin/env python3
"""Measure TTFT and TPOT by streaming from llama-server."""
import argparse
import json
import subprocess
import sys
import time
import requests
SERVER_BIN = "./build/bin/llama-server"
PORT = 8081
PROMPT = "Explain the difference between machine learning and deep learning in detail." * 8 # ~512 tokens
MAX_TOKENS = 128
RUNS = 3
def wait_for_server(address, timeout=120):
for _ in range(timeout):
try:
r = requests.get(f"{address}/health", timeout=2)
if r.status_code == 200:
return True
except Exception:
pass
time.sleep(1)
return False
def measure_once(address):
payload = {
"messages": [{"role": "user", "content": PROMPT}],
"max_tokens": MAX_TOKENS,
"stream": True,
"temperature": 0.0,
}
t0 = time.perf_counter()
ttft = None
n_tokens = 0
with requests.post(f"{address}/v1/chat/completions", json=payload, stream=True, timeout=120) as resp:
for line in resp.iter_lines():
if not line:
continue
line = line.decode("utf-8")
if not line.startswith("data:"):
continue
data = line[5:].strip()
if data == "[DONE]":
break
try:
chunk = json.loads(data)
delta = chunk["choices"][0]["delta"].get("content", "")
if delta:
if ttft is None:
ttft = time.perf_counter() - t0
n_tokens += 1
except Exception:
continue
t_total = time.perf_counter() - t0
tpot = (t_total - ttft) / max(n_tokens - 1, 1) if n_tokens > 1 else 0
return ttft, tpot, n_tokens, t_total
def main():
parser = argparse.ArgumentParser()
parser.add_argument("-m", "--model", required=True)
parser.add_argument("--runs", type=int, default=RUNS)
parser.add_argument("--device", default=None,
help="GGML device string, e.g. CUDA0 (default: all GPUs)")
args = parser.parse_args()
address = f"http://127.0.0.1:{PORT}"
cmd = [SERVER_BIN, "-m", args.model, "-ngl", "99", "--port", str(PORT), "--log-disable"]
if args.device:
cmd += ["-dev", args.device]
print(f"Starting server: {' '.join(cmd)}", file=sys.stderr)
proc = subprocess.Popen(cmd, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)
try:
if not wait_for_server(address):
print("ERROR: server did not start", file=sys.stderr)
proc.kill()
sys.exit(1)
print("Server ready", file=sys.stderr)
ttfts, tpots = [], []
for i in range(args.runs):
ttft, tpot, n_tokens, t_total = measure_once(address)
ttfts.append(ttft * 1000)
tpots.append(tpot * 1000)
print(f" Run {i+1}: TTFT={ttft*1000:.1f}ms TPOT={tpot*1000:.1f}ms tokens={n_tokens}", file=sys.stderr)
avg_ttft = sum(ttfts) / len(ttfts)
avg_tpot = sum(tpots) / len(tpots)
result = {"ttft_ms": round(avg_ttft, 1), "tpot_ms": round(avg_tpot, 1), "latency_ms": round(avg_ttft + avg_tpot, 1)}
print(json.dumps(result))
finally:
proc.kill()
proc.wait()
if __name__ == "__main__":
main()