--- license: mit language: - en metrics: - accuracy - f1 - precision base_model: - bhadresh-savani/distilbert-base-uncased-emotion pipeline_tag: audio-classification tags: - speech - emotion - ser - classification --- # TuBERT: Multimodal Speech Emotion Recognition For Real-Time Avatar Control *This project was developed for my senior thesis at Princeton University. Paper being published soon.* ## About TuBERT is a multimodal speech emotion recognition model that runs in real-time and on-device. I designed it with PNGTubers in mind, but there are plenty of other applications for it as well! To test the model for yourself using a GUI, see the [GitHub repository](https://github.com/YacoubKahkajian/TuBERT) for installation instructions. ## Usage `tubert.pt` is the base TuBERT model trained on MELD, described in the paper and used by default for evaluation. `tubert_iemocap.pt` is the version of the model fine-tuned on IEMOCAP.