File size: 2,234 Bytes
3fce4cb
2aee800
 
 
 
 
 
 
 
5bf0275
2aee800
 
 
 
 
 
 
4d2d927
5bf0275
2aee800
5bf0275
 
2aee800
5bf0275
 
 
3fce4cb
 
2aee800
3fce4cb
2aee800
3fce4cb
 
 
2aee800
 
 
 
3fce4cb
2aee800
3fce4cb
2aee800
 
 
3fce4cb
2aee800
 
 
3fce4cb
2aee800
 
3fce4cb
2aee800
 
3fce4cb
2aee800
 
 
 
3fce4cb
2aee800
 
3fce4cb
 
 
2aee800
 
 
 
3fce4cb
2aee800
3fce4cb
2aee800
 
 
3fce4cb
2aee800
3fce4cb
2aee800
3fce4cb
2aee800
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
---
language: en
license: apache-2.0
tags:
- whisper
- automatic-speech-recognition
- speech
- audio
datasets:
- your-dataset-name  # Replace with your actual dataset
metrics:
- wer
- cer
model-index:
- name: AfroLogicInsect/whisper-finetuned-float32
  results:
  - task:
      name: Automatic Speech Recognition
      type: automatic-speech-recognition
    dataset:
      name: Your Dataset Name  # Replace
      type: your-dataset-type  # Replace
    metrics:
    - name: WER
      type: wer
      value: "your-wer-score"  # Replace with actual score
---

# AfroLogicInsect/whisper-finetuned-float32

Fine-tuned Whisper model (float32 version) for speech recognition

## Model Details

- **Model Type**: Whisper (Fine-tuned)
- **Language**: English
- **Data Type**: float32
- **Use Cases**: Speech-to-text transcription

## Usage

```python
from transformers import WhisperProcessor, WhisperForConditionalGeneration
import librosa

# Load model and processor
processor = WhisperProcessor.from_pretrained("AfroLogicInsect/whisper-finetuned-float32")
model = WhisperForConditionalGeneration.from_pretrained("AfroLogicInsect/whisper-finetuned-float32")

# Load audio
audio, sr = librosa.load("path/to/audio.wav", sr=16000)

# Process
input_features = processor(audio, sampling_rate=16000, return_tensors="pt").input_features

# Generate transcription
with torch.no_grad():
    predicted_ids = model.generate(input_features)
    transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)[0]

print(transcription)
```

## Training Details

- **Base Model**: OpenAI Whisper
- **Training Dataset**: [Add your dataset details]
- **Training Parameters**: [Add your training parameters]
- **Evaluation Metrics**: [Add your evaluation results]

## Limitations and Biases

- This model may have biases present in the training data
- Performance may vary on different accents or audio qualities
- Recommended for English speech recognition tasks

## Citation

If you use this model, please cite:

```bibtex
@misc{whisper-finetuned,
  author = {Daniel AMAH},
  title = {Fine-tuned Whisper Model},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/AfroLogicInsect/whisper-finetuned-float32}
}
```