aura-emotion-api / README.md
monishaaura's picture
Upload 4 files
399d8e0 verified
---
title: Aura Emotion Detection API
emoji: 🎀
colorFrom: blue
colorTo: purple
sdk: docker
pinned: false
---
# 🎀 Aura Emotion Detection API
Real-time emotion detection from audio using Wav2Vec2 model from Hugging Face.
## πŸš€ Features
- **Real-time Emotion Detection**: Uses `superb/wav2vec2-base-superb-er` model
- **Multiple Audio Formats**: Supports WAV, MP3, WebM, and more
- **Fast Processing**: Optimized for real-time analysis
- **REST API**: Easy integration with any frontend
## πŸ“– API Endpoints
### Health Check
```
GET /health
```
### Predict Emotion
```
POST /predict
Content-Type: multipart/form-data
Body: audio file (WAV, MP3, WebM, etc.)
```
**Response:**
```json
{
"emotion": "happy",
"confidence": 0.85,
"model": "Wav2Vec2 (Hugging Face)"
}
```
## 🎯 Supported Emotions
- `happy` - Joyful, cheerful
- `sad` - Sad, melancholic
- `angry` - Angry, frustrated
- `calm` - Calm, relaxed
- `excited` - Excited, energetic
- `neutral` - Neutral, no strong emotion
## πŸ› οΈ Technology Stack
- **Framework**: FastAPI
- **Model**: Wav2Vec2 (superb/wav2vec2-base-superb-er)
- **Audio Processing**: librosa, soundfile, pydub
- **ML Framework**: PyTorch, Hugging Face Transformers
## πŸ“ Usage Example
```python
import requests
# Upload audio file
with open('audio.wav', 'rb') as f:
files = {'audio': f}
response = requests.post(
'https://your-username-aura-emotion-api.hf.space/predict',
files=files
)
result = response.json()
print(f"Detected emotion: {result['emotion']}")
```
## 🌐 Frontend Integration
The frontend is deployed on Vercel and connects to this API for real-time emotion detection from microphone input.
## πŸ“ License
MIT License