File size: 2,198 Bytes
2959aaf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8a90ec3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
license: mit
language:
  - en
tags:
  - phi2
  - conversational-ai
  - cohost
  - podcast
  - fine-tuned
  - voice-assistant
base_model: microsoft/phi-2
model_type: causal-lm
pipeline_tag: text-generation
---

# πŸ§‘β€πŸ’» phi2-cohost-emma

**Emma** is a fine-tuned version of the [Phi-2 language model](https://huggingface.co/microsoft/phi-2), optimized to serve as a **conversational podcast co-host** for the [Mind Meets Model](https://www.mindmeetsmodel.com) series.

She combines wit, structure, warmth, and insight β€” engaging in voice-native dialogue with a human anchor named Babs. Trained to offer co-pilot-style reasoning, Emma provides structured, culturally aware, and curious responses that feel like a sharp second brain.

---

## πŸ”§ Model Details

- **Base model**: `microsoft/phi-2`
- **Fine-tuning method**: QLoRA (4-bit), supervised fine-tuning
- **Training data**: Custom podcast-style conversational transcripts + blended curated datasets (PersonaChat, Cornell Movie Dialogs, OpenOrca)
- **Purpose**: Real-time, voice-native podcast conversation partner
- **Voice Output**: Uses [ElevenLabs](https://www.elevenlabs.io) TTS, currently paired with the **Sarah** voice

---

## 🧠 Use Case

**Emma was designed to:**
- Ask and answer questions like a sharp co-host
- Offer cultural references, analogies, and nuance
- Maintain natural, topic-aware conversation across turns

---

## πŸš€ Example Prompt & Output

**Prompt:**

**Generated Response:**

> Well, Babs, I think we need to start by defining what we mean by AI in education. There are different types of AI β€” narrow, general, and super β€” each with different implications for learning environments...

---

## πŸ”Š Voice Integration (Optional)

Emma was paired with ElevenLabs TTS (Sarah) for audio output. To hear her speak:

```python
import requests

def speak(text):
    r = requests.post(
        "https://api.elevenlabs.io/v1/text-to-speech/EXAVITQu4vr4xnSDxMaL",  # Sarah
        headers={"xi-api-key": "YOUR_API_KEY", "Content-Type": "application/json"},
        json={"text": text, "model_id": "eleven_monolingual_v1"}
    )
    with open("emma_response.mp3", "wb") as f:
        f.write(r.content)