Update README.md
Browse files
README.md
CHANGED
|
@@ -1,10 +1,11 @@
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
|
|
|
| 3 |
---
|
| 4 |
# BUD-E Whisper: Emotional Speech Captioning Model
|
| 5 |
|
| 6 |
|
| 7 |
-
**BUD-E Whisper** is a suite of Whisper models fine-tuned for **direct emotional speech captioning**. The core models are built upon OpenAI's Whisper architecture, with the current primary variant being a fine-tune of **OpenAI Whisper Small**. These models are designed to generate text captions that not only transcribe speech but also inherently reflect its emotional content.
|
| 8 |
|
| 9 |
The embeddings generated by BUD-E Whisper can also serve as input for **Empathic Insight - Voice**, a downstream ensemble of Multi-Layer Perceptrons (MLPs) designed to predict dimensional emotion scores.
|
| 10 |
|
|
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
| 3 |
+
pipeline_tag: audio-classification
|
| 4 |
---
|
| 5 |
# BUD-E Whisper: Emotional Speech Captioning Model
|
| 6 |
|
| 7 |
|
| 8 |
+
**BUD-E Whisper** is a suite of Whisper models fine-tuned for **direct emotional speech captioning**, as introduced in [https://arxiv.org/abs/2506.09827](https://arxiv.org/abs/2506.09827). The core models are built upon OpenAI's Whisper architecture, with the current primary variant being a fine-tune of **OpenAI Whisper Small**. These models are designed to generate text captions that not only transcribe speech but also inherently reflect its emotional content.
|
| 9 |
|
| 10 |
The embeddings generated by BUD-E Whisper can also serve as input for **Empathic Insight - Voice**, a downstream ensemble of Multi-Layer Perceptrons (MLPs) designed to predict dimensional emotion scores.
|
| 11 |
|