File size: 1,663 Bytes
0e13326
 
 
 
 
 
 
 
 
 
 
 
 
39a2a9f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0e13326
 
 
 
 
 
 
 
 
 
 
 
 
39a2a9f
0e13326
39a2a9f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
title: DevOps Metrics Interpreter
emoji: πŸ“Š
colorFrom: blue
colorTo: indigo
sdk: docker
pinned: false
---

# DevOps Metrics Interpreter

Enter your team's DevOps metrics and get a plain-language interpretation compared to DORA benchmark data from the State of DevOps Reports.

## How report generation works

Reports are generated by combining three sources of knowledge, all extracted from the State of DevOps Report PDFs and stored in `data/benchmarks.json`:

1. **Benchmark tiers** β€” elite/high/medium/low bands for the four DORA metrics (deployment frequency, lead time, change failure rate, MTTR)
2. **47 patterns** β€” metric combinations and what they signal (e.g. high deploy frequency + high failure rate β†’ missing release safeguards)
3. **124 key insights** β€” statistics and findings from the reports (e.g. "elite performers deploy 208x more frequently than low performers")

All three are injected into the LLM's system prompt on every request alongside the team's submitted metrics. The LLM is instructed to reference benchmark bands explicitly and ground its improvement recommendations in the extracted knowledge.

Report quality depends on what was captured during extraction. To re-extract with a better model:

```bash
LLM_MODEL=gpt-4o npm run extract
```

## Running locally

```bash
cp .env.local.example .env.local
# Edit .env.local with your settings
npm install
npm run dev
```

## Environment variables (HF Space settings)

| Variable | Value |
|----------|-------|
| `OLLAMA_BASE_URL` | `https://router.huggingface.co/v1` |
| `OPENAI_API_KEY` | Your HuggingFace token (`hf_...`) |
| `LLM_MODEL` | `Qwen/Qwen2.5-72B-Instruct` |