File size: 4,412 Bytes
56589d3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
# Deploy to HuggingFace Spaces - Quick Start

## βœ… Issue Fixed
**The `quote_extractor` import error has been fixed!** The app will now work even if the file is missing.



---



## πŸš€ Option 1: Automated Preparation (Recommended)



Run this script to prepare a clean deployment package:



```bash

python prepare_for_spaces.py

```



This will:

- Create a `spaces_deployment/` directory

- Copy only the required files

- Remove any .env or test files

- Show you a summary of what's included



Then upload everything from `spaces_deployment/` to your Space.



---



## πŸ“‹ Option 2: Manual Upload



Upload these files to your HuggingFace Space:



### Required Files (Must have)

```

app.py

llm.py

extractors.py

tagging.py

chunking.py

validation.py

reporting.py

dashboard.py

production_logger.py

quote_extractor.py

requirements.txt

```



### Optional Files

```

README.md

HUGGINGFACE_SPACES_SETUP.md

```



**DO NOT upload:**

- `.env` file

- `test_*.py` files

- `logs/` directory

- `outputs/` directory



---



## πŸ”§ Space Configuration



### 1. Create Space

- Go to https://huggingface.co/new-space

- Name: `transcriptor-ai` (or your choice)

- SDK: **Gradio**

- Hardware: **GPU (T4 or better)** ← Important!

### 2. Upload Files
- Drag and drop all files from the list above
- OR connect a Git repository

### 3. Configure (Optional)
Go to **Settings β†’ Variables** and add:

| Variable | Value | When to Use |
|----------|-------|-------------|
| `DEBUG_MODE` | `True` | To see detailed logs |
| `LOCAL_MODEL` | `TinyLlama/TinyLlama-1.1B-Chat-v1.0` | For faster (but lower quality) processing |
| `LLM_TEMPERATURE` | `0.5` | For more deterministic outputs |

**Note:** All settings have defaults - you don't need to configure anything!

---

## ⏱️ First Deployment

### What to Expect
1. **Build time:** 2-5 minutes (installing dependencies)
2. **Model download:** 2-5 minutes (first time only - downloads Phi-3-mini)
3. **Subsequent starts:** 30-60 seconds

### Watch the Logs
Click **Logs** tab to see:
```

βœ… Configuration loaded for HuggingFace Spaces

πŸš€ TranscriptorAI Enterprise - LLM Backend: local

[Local Model] Loading microsoft/Phi-3-mini-4k-instruct...

Downloading (…)lve/main/config.json: 100%

[Local Model] βœ… Model loaded on cuda:0

Running on local URL:  http://0.0.0.0:7860

```

---

## πŸ§ͺ Test Your Space

1. Wait for "Running on local URL" message
2. Upload a sample transcript (DOCX or PDF)
3. Select "HCP" as interviewee type
4. Click "Analyze Transcripts"

**Expected:**
- Processing time: 5-10 minutes (depending on transcript length)
- Quality score: 0.7-1.0
- CSV and PDF downloads available

---

## πŸ› Troubleshooting

### Error: `ModuleNotFoundError: No module named 'quote_extractor'`

**Status:** βœ… FIXED - This is now optional



### Error: `ModuleNotFoundError: No module named 'xyz'`

**Solution:** Upload the missing `xyz.py` file



### Error: `CUDA out of memory`

**Solution:**

- Change model: Add Variable `LOCAL_MODEL=TinyLlama/TinyLlama-1.1B-Chat-v1.0`
- OR upgrade to larger GPU

### Error: Very slow processing
**Check:**
- Is GPU hardware selected? (Not CPU)
- Look for "Model loaded on cuda:0" in logs
- If you see "cpu", upgrade to GPU tier

### Quality Score still 0.00
**Debug:**
1. Set `DEBUG_MODE=True` in Variables
2. Check logs for "[Local Model] βœ… Generated X characters"
3. Look for "[LLM Debug] Successfully extracted JSON"
4. If you see `[Error]` messages, share them

---

## πŸ’‘ Tips

### Reduce Costs
- Space sleeps after 48h inactivity (free)
- Only pays for GPU time when active
- ~$0.60/hour for T4 GPU

### Improve Speed
- Use smaller model (TinyLlama)
- Reduce max tokens (edit llm.py line 410)
- Process fewer chunks

### Improve Quality
- Use larger model (Mistral-7B)
- Increase temperature for creative outputs
- Keep default Phi-3-mini for best balance

---

## πŸ“ž Need Help?

1. **Check logs first** - Most issues show clear error messages
2. **Read HUGGINGFACE_SPACES_SETUP.md** - Detailed troubleshooting
3. **Test locally first** - Run `python test_local_model.py`

---

## ✨ You're Ready!

Run the preparation script:
```bash

python prepare_for_spaces.py

```

Then upload to HuggingFace Spaces and you're done! πŸŽ‰

---

**Last Updated:** October 2025