DiCoW_v3_2 / README.md
Lakoc's picture
Update README.md
00a09c6 verified
---
library_name: transformers
tags:
- speech
- automatic-speech-recognition
- whisper
- multilingual
- speaker-diarization
- meeting-transcription
- DiCoW
- BUT-FIT
pipeline_tag: automatic-speech-recognition
license: cc-by-4.0
datasets:
- microsoft/NOTSOFAR
- edinburghcstr/ami
---
# 🧠 DiCoW\_v3.2 — BUT-FIT Model for MT-ASR
This repository hosts the **DiCoW\_v3.2** model developed by [BUT Speech@FIT](https://github.com/BUTSpeechFIT), tailored for **multi-talker automatic speech recognition (MT-ASR)**.
This model is available under the terms of CC BY 4.0. It incorporates an MIT-licensed base model and CC BY 4.0 licensed training data.
## 🔧 Key Improvements over DiCoW v1
* **FDDT (Frame-Level Diarization Dependent Transformation)** before positional embeddings
* **Less strict suppressive initialization** to ease early training dynamics
* **Enhanced sequential decoding** with fallback seeking
* **Frozen decoder** during fine-tuning to retain language modeling capabilities
### 🧪 Augmentations
* Random **STNO** noise injection
* Segment-wise random class flipping of **STNO tokens**
* **SpecAugment**
* **MUSAN** noise mixing
### ⚙️ Optimization & Inference Enhancements
* Updated **learning schedule**
* Improved **hallucination detection & mitigation** during inference
---
## 🛠️ Model Usage
```python
from transformers import AutoModelForSpeechSeq2Seq
MODEL_NAME = "BUT-FIT/DiCoW_v3_2"
dicow = AutoModelForSpeechSeq2Seq.from_pretrained(MODEL_NAME, trust_remote_code=True)
```
➡️ For detailed inference pipelines, see: [**DiCoW GitHub (Inference)**](https://github.com/BUTSpeechFIT/DiCoW)
---
## 🏆 Performance
See how **DiCoW_v3.2** performs on our multi-talker ASR benchmark:
- 🔗 [**EMMA-MT ASR Leaderboard**](https://huggingface.co/spaces/BUT-FIT/EMMA_leaderboard)
---
## 📦 Model Details
* **Base Model:** Whisper large-v3-turbo
* **Training Datasets:**
* [NOTSOFAR-1](https://github.com/microsoft/NOTSOFAR1-Challenge)
* [AMI Meeting Corpus](http://groups.inf.ed.ac.uk/ami/corpus/)
* [Libri2Mix](https://github.com/JorisCos/LibriMix)
---
## 🧬 Source Repositories
* 🔧 [Training Code: TS-ASR-Whisper](https://github.com/BUTSpeechFIT/TS-ASR-Whisper)
* 🚀 [Inference](https://github.com/BUTSpeechFIT/DiCoW)
---
## 📚 Related Publications
* 📰 **Journal Paper:**
*DiCoW: Diarization-Conditioned Whisper for Target Speaker Automatic Speech Recognition*
[Computer Speech & Language, 2026](https://www.sciencedirect.com/science/article/pii/S088523082500066X)
* 📰 **ICASSP 2025:**
*Target Speaker ASR with Whisper*
[IEEE ICASSP 2025](https://doi.org/10.1109/ICASSP49660.2025.10887683)
* 📰 **CHiME-8 System Description:**
*BUT/JHU System Description for CHiME-8 NOTSOFAR-1 Challenge*
[CHiME 2024 Proceedings](https://doi.org/10.21437/CHiME.2024-4)
* 📰 **MLC-SLM Challenge Submission:**
*BUT System for the MLC-SLM Challenge*
[arXiv:2506.13414](https://arxiv.org/abs/2506.13414)
---
## 📝 Citation
If you use this model, please cite the following works:
```bibtex
@article{POLOK2026101841,
title = {DiCoW: Diarization-conditioned Whisper for target speaker automatic speech recognition},
journal = {Computer Speech & Language},
volume = {95},
pages = {101841},
year = {2026},
issn = {0885-2308},
doi = {https://doi.org/10.1016/j.csl.2025.101841},
url = {https://www.sciencedirect.com/science/article/pii/S088523082500066X},
author = {Alexander Polok and Dominik Klement and Martin Kocour and Jiangyu Han and Federico Landini and Bolaji Yusuf and Matthew Wiesner and Sanjeev Khudanpur and Jan Černocký and Lukáš Burget},
keywords = {Diarization-conditioned Whisper, Target-speaker ASR, Speaker diarization, Long-form ASR, Whisper adaptation},
}
@INPROCEEDINGS{10887683,
author={Polok, Alexander and Klement, Dominik and Wiesner, Matthew and Khudanpur, Sanjeev and Černocký, Jan and Burget, Lukáš},
booktitle={ICASSP 2025 - 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
title={Target Speaker ASR with Whisper},
year={2025},
volume={},
number={},
pages={1-5},
keywords={Transforms;Signal processing;Transformers;Acoustics;Speech processing;target-speaker ASR;diarization conditioning;multi-speaker ASR;Whisper},
doi={10.1109/ICASSP49660.2025.10887683}
}
```
---
## 📬 Contact
For questions or collaboration inquiries:
📧 **Email:** [ipoloka@fit.vut.cz](mailto:ipoloka@fit.vut.cz)
🏢 **Affiliation:** [BUT Speech@FIT](https://github.com/BUTSpeechFIT), Brno University of Technology
🔗 **GitHub:** [BUTSpeechFIT](https://github.com/BUTSpeechFIT)