Update README.md
Browse files
README.md
CHANGED
|
@@ -1,7 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
## Model Card: Wav2vec_Classroom_FT
|
| 2 |
|
| 3 |
### Model Overview
|
| 4 |
-
**Model Name:**Wav2vec_Classroom_FT
|
| 5 |
**Version:** 1.0
|
| 6 |
**Developed By:** Ahmed Adel Attia (University of Maryland and Stanford University)
|
| 7 |
**Date:** 2025
|
|
@@ -11,6 +18,8 @@ NCTE-Baseline-ASR is an automatic speech recognition (ASR) model trained for cla
|
|
| 11 |
|
| 12 |
This model is adapted from **[Wav2vec-Classroom](https://huggingface.co/aadel4/Wav2vec_Classroom)**, which was trained using continued pretraining (CPT) on large-scale unlabeled classroom speech data. The adaptation involves direct fine-tuning on a limited transcribed dataset.
|
| 13 |
|
|
|
|
|
|
|
| 14 |
**Use Case:**
|
| 15 |
- Speech-to-text transcription for classroom environments.
|
| 16 |
- ASR applications requiring high precision with limited data.
|
|
@@ -43,5 +52,4 @@ This model is adapted from **[Wav2vec-Classroom](https://huggingface.co/aadel4/W
|
|
| 43 |
### Usage Request
|
| 44 |
If you use the NCTE-Baseline-ASR model in your research, please acknowledge this work and refer to the original paper submitted to Interspeech 2025.
|
| 45 |
|
| 46 |
-
For inquiries or collaborations, please contact the authors of the original paper.
|
| 47 |
-
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
base_model:
|
| 4 |
+
- aadel4/Wav2vec_Classroom
|
| 5 |
+
- facebook/wav2vec2-large-robust
|
| 6 |
+
pipeline_tag: automatic-speech-recognition
|
| 7 |
+
---
|
| 8 |
## Model Card: Wav2vec_Classroom_FT
|
| 9 |
|
| 10 |
### Model Overview
|
| 11 |
+
**Model Name:** Wav2vec_Classroom_FT
|
| 12 |
**Version:** 1.0
|
| 13 |
**Developed By:** Ahmed Adel Attia (University of Maryland and Stanford University)
|
| 14 |
**Date:** 2025
|
|
|
|
| 18 |
|
| 19 |
This model is adapted from **[Wav2vec-Classroom](https://huggingface.co/aadel4/Wav2vec_Classroom)**, which was trained using continued pretraining (CPT) on large-scale unlabeled classroom speech data. The adaptation involves direct fine-tuning on a limited transcribed dataset.
|
| 20 |
|
| 21 |
+
This model was originally trained using the fairseq library then ported into Huggingface.
|
| 22 |
+
|
| 23 |
**Use Case:**
|
| 24 |
- Speech-to-text transcription for classroom environments.
|
| 25 |
- ASR applications requiring high precision with limited data.
|
|
|
|
| 52 |
### Usage Request
|
| 53 |
If you use the NCTE-Baseline-ASR model in your research, please acknowledge this work and refer to the original paper submitted to Interspeech 2025.
|
| 54 |
|
| 55 |
+
For inquiries or collaborations, please contact the authors of the original paper.
|
|
|