|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
base_model: |
|
|
- meta-llama/Llama-3.2-3B-Instruct |
|
|
--- |
|
|
--- |
|
|
<div align="center"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/distillabs-logo.svg?raw=true" width="40%" alt="distil labs" /> |
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
<table> |
|
|
<tr> |
|
|
<td align="center"> |
|
|
<a href="https://www.distillabs.ai/?utm_source=hugging-face&utm_medium=referral&utm_campaign=distil-resume-roast"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-distillabs-home.svg?raw=true" alt="Homepage"/> |
|
|
</a> |
|
|
</td> |
|
|
<td align="center"> |
|
|
<a href="https://github.com/distil-labs"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-github.svg?raw=true" alt="GitHub"/> |
|
|
</a> |
|
|
</td> |
|
|
<td align="center"> |
|
|
<a href="https://huggingface.co/distil-labs"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-huggingface.svg?raw=true" alt="Hugging Face"/> |
|
|
</a> |
|
|
</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center"> |
|
|
<a href="https://www.linkedin.com/company/distil-labs/"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-linkedin.svg?raw=true" alt="LinkedIn"/> |
|
|
</a> |
|
|
</td> |
|
|
<td align="center"> |
|
|
<a href="https://distil-labs-community.slack.com/join/shared_invite/zt-36zqj87le-i3quWUn2bjErRq22xoE58g"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-slack.svg?raw=true" alt="Slack"/> |
|
|
</a> |
|
|
</td> |
|
|
<td align="center"> |
|
|
<a href="https://x.com/distil_labs"> |
|
|
<img src="https://github.com/distil-labs/badges/blob/main/badge-twitter.svg?raw=true" alt="Twitter"/> |
|
|
</a> |
|
|
</td> |
|
|
</tr> |
|
|
</table> |
|
|
</div> |
|
|
|
|
|
# Resume Roaster AI |
|
|
|
|
|
We trained an SLM (Small Language Model) assistant for automatic resume critique — a Llama-3.2-3B parameter model that generates "Roast Mode" feedback and professional improvement suggestions. |
|
|
Run it locally to keep your personal data private, or deploy it for instant feedback! |
|
|
|
|
|
|
|
|
### **1. Install Dependencies** |
|
|
|
|
|
First, install **[Ollama](http://ollama.com/)** from their official website. |
|
|
Then set up your Python environment: |
|
|
|
|
|
```bash |
|
|
# Create a virtual environment |
|
|
python -m venv .venv |
|
|
source .venv/bin/activate # Windows: .venv\Scripts\activate |
|
|
|
|
|
# Install required tools |
|
|
pip install huggingface_hub ollama rich pymupdf |
|
|
``` |
|
|
Available models hosted on HuggingFace: |
|
|
|
|
|
- **[distil-labs/Distil-Rost-Resume-Llama-3.2-3B-Instruct](https://huggingface.co/distil-labs/Distil-Rost-Resume-Llama-3.2-3B-Instruct)** |
|
|
|
|
|
### **2. Setup the Model** |
|
|
|
|
|
Download your fine-tuned GGUF model and register it with Ollama. |
|
|
|
|
|
```bash |
|
|
hf download distil-labs/Distil-Rost-Resume-Llama-3.2-3B-Instruct --local-dir distil-model |
|
|
|
|
|
cd distil-model |
|
|
# Create the Ollama model from the Modelfile |
|
|
ollama create roast_master -f Modelfile |
|
|
``` |
|
|
|
|
|
### **3. Usage** |
|
|
|
|
|
Now you can roast any resume PDF instantly from your terminal. |
|
|
|
|
|
```bash |
|
|
# Syntax: python roast.py <path_to_resume.pdf> |
|
|
python roast.py my_resume.pdf |
|
|
``` |
|
|
|
|
|
## ✨ Features |
|
|
|
|
|
The assistant is trained to analyze resumes and output structured JSON containing: |
|
|
|
|
|
- **💀 Roast Critique** |
|
|
A sarcastic, humorous paragraph quoting specific problematic parts of the resume (typos, clichés, gaps). |
|
|
|
|
|
- **✨ Professional Suggestions** |
|
|
A list of **exactly 3** constructive, actionable tips to improve the resume. |
|
|
|
|
|
- **📊 Rating** |
|
|
An integer score **(1–10)** based on overall resume quality. |
|
|
|
|
|
## 📊 Model Evaluation & Fine-Tuning Results |
|
|
|
|
|
To validate the necessity of fine-tuning, we performed a strict **A/B Test** comparing the **Base Model** (Llama-3.2-3B-Instruct) against our **Fine-Tuned Student** (Llama-3.2-3B-Instruct). |
|
|
|
|
|
### 1. The Engineering Challenge |
|
|
We needed the model to satisfy three conflicting requirements simultaneously: |
|
|
1. **Strict JSON Schema:** Output *only* valid JSON (no Markdown wrappers like ` ```json `, no conversational filler). |
|
|
2. **Persona Shift:** Move from the base model's "Helpful Assistant" tone to a "Ruthless Roaster" persona. |
|
|
3. **Context Awareness:** Cite specific details from the resume rather than giving generic advice. |
|
|
|
|
|
### 2. Quantitative Results |
|
|
|
|
|
| Metric | 🤖 Base Model (Llama-3.2-1B) | 👨🏫 Teacher Model (gpt-oss-120b) | 🔥 Fine-Tuned Student (Custom) | |
|
|
| :--- | :--- | :--- | :--- | |
|
|
| **JSON Valid Rate** | **70% (Failed)** | **100% (Passed)** | **100% (Passed)** <br>Matches Teacher performance. | |
|
|
| **Persona Score** | **4/10 (Generic Snark)** <br>Often polite or apologetic. | **10/10 (Benchmark)** | **10/10 (Highly Contextual)** <br>Successfully mimicked the Teacher's tone. | |
|
|
| **Hallucination** | **High** <br>(Invented formatting wrappers) | **None** | **None** <br>Strict adherence. | |
|
|
|
|
|
> **Methodology:** A human evaluator reviewed outputs. |
|
|
|
|
|
### **Evaluation Metrics** |
|
|
|
|
|
| **Metric** | **Teacher** | **Student-Base (Untrained)** | **Student-Tuned (Ours)** | |
|
|
|--------------------|-------------------|-------------------------------|---------------------------| |
|
|
| **ROUGE-L** | 0.44 | 0.10 | **0.43 ⬆** | |
|
|
| **METEOR** | 0.36 | 0.17 | **0.35 ⬆** | |
|
|
|
|
|
|
|
|
### 3. Qualitative Comparison (Side-by-Side) |
|
|
|
|
|
We fed the same resume ("Paritosh") to both models. |
|
|
|
|
|
#### ❌ Base Model Output (Failure) |
|
|
* **Status:** **CRASH** (JSON Parsing Error) |
|
|
* **Why:** The model "hallucinated" a conversational intro and added a disclaimer at the end, breaking the JSON parser. |
|
|
```text |
|
|
Here is the critique for the resume: |
|
|
{ |
|
|
"roast_critique": "The resume is okay but lacks specific details..." |
|
|
} |
|
|
Note: I tried to be critical but maintained a professional tone. |
|
|
``` |
|
|
#### ✅ Fine-Tuned Model Output (Success) |
|
|
**Status:** PASS (Valid JSON) |
|
|
|
|
|
**Why:** The model output raw JSON immediately. The tone was specific to the user's *"Income Inequality"* project. |
|
|
#### JSON |
|
|
```json |
|
|
{ |
|
|
"roast_critique": "PARITOSH, the future of computer science is you... in a really long time. Your resume is a jumbled mess of buzzwords. 'Integrated Unstructured data' sounds like something a 5-year-old would say, and don't even get me started on the 'Income Inequality Analysis' - who doesn't love a good spreadsheet? Oh, and please stop using Comic Sans.", |
|
|
"professional_suggestions": [ |
|
|
"Provide specific examples of real-world problems you solved instead of listing generic tech stacks.", |
|
|
"Quantify accomplishments. Instead of 'reduced response time', say 'improved response time from 2s to 0.4s'.", |
|
|
"Get rid of the 'Certifications' section and replace it with real project experience." |
|
|
], |
|
|
"rating": 6 |
|
|
} |
|
|
``` |
|
|
|
|
|
#### Example 1 |
|
|
<p align="center"> |
|
|
<img src="https://github.com/distil-labs/distil-resume-roast/blob/main/examples/rr-1.png?raw=true" width="550" alt="Example 1" /> |
|
|
</p> |
|
|
|
|
|
--- |
|
|
|
|
|
#### Example 2 |
|
|
<p align="center"> |
|
|
<img src="https://github.com/distil-labs/distil-resume-roast/blob/main/examples/rr-2.png?raw=true" width="550" alt="Example 2" /> |
|
|
</p> |
|
|
|
|
|
--- |
|
|
|
|
|
#### Example 3 |
|
|
<p align="center"> |
|
|
<img src="https://github.com/distil-labs/distil-resume-roast/blob/main/examples/rr-3.png?raw=true" width="550" alt="Example 3" /> |
|
|
</p> |
|
|
|
|
|
|
|
|
### **Training Config** |
|
|
|
|
|
- **Student:** Llama-3.2-3B-Instruct |
|
|
- **Teacher:** openai.gpt-oss-120b |
|
|
- **Dataset:** 10,000 synthetic examples |
|
|
|
|
|
### 4. Conclusion |
|
|
|
|
|
The fine-tuning process **successfully eliminated the formatting hallucinations** present in the base model and **significantly enhanced the "Roaster" persona**, making the outputs more structured, consistent, and aligned with the intended tone. |
|
|
|
|
|
|
|
|
|
|
|
## ❓ FAQ |
|
|
|
|
|
--- |
|
|
|
|
|
<details> |
|
|
<summary><strong>Q: Why not just use ChatGPT or Claude?</strong></summary> |
|
|
|
|
|
**Privacy and cost.** |
|
|
Resumes contain sensitive personal data (PII). Sending them to cloud APIs risks exposure. |
|
|
Our model runs **fully locally**, ensuring zero data leaks and costs **nothing** to run. |
|
|
</details> |
|
|
|
|
|
--- |
|
|
|
|
|
<details> |
|
|
<summary><strong>Q: How accurate is a 3B model compared to GPT-4?</strong></summary> |
|
|
|
|
|
Surprisingly good for this specific task! |
|
|
Because it’s fine-tuned on **6,000+ high-quality roast-style examples**, it performs far better than a generic prompt to GPT-4. |
|
|
It captures the **roast persona** more consistently and is extremely fast. |
|
|
</details> |
|
|
|
|
|
--- |
|
|
|
|
|
<details> |
|
|
<summary><strong>Q: Can I use this for serious resume reviews?</strong></summary> |
|
|
|
|
|
Yes! |
|
|
The **Professional Suggestions** section is trained on real career guidance data. |
|
|
You can ignore the roast and only use the actionable tips. |
|
|
</details> |
|
|
|
|
|
--- |
|
|
|
|
|
<details> |
|
|
<summary><strong>Q: The model is too mean! Can I change it?</strong></summary> |
|
|
|
|
|
The model is intentionally “brutally honest.” |
|
|
But since it outputs **structured JSON**, you can simply hide the `roast` field and show only the suggestions. |
|
|
</details> |
|
|
|
|
|
--- |
|
|
|
|
|
<details> |
|
|
<summary><strong>Q: What hardware do I need?</strong></summary> |
|
|
|
|
|
**Minimum:** |
|
|
- 8GB RAM (CPU Mode) |
|
|
- Works well on modern laptops (Mac M1/M2/M3 recommended) |
|
|
|
|
|
**Recommended:** |
|
|
- NVIDIA GPU with **4GB+ VRAM** for 2–5s inference |
|
|
</details> |
|
|
|