dungca commited on
Commit
c86e069
·
verified ·
1 Parent(s): af980d3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -21
README.md CHANGED
@@ -11,48 +11,75 @@ tags:
11
  library_name: transformers
12
  base_model: openai/whisper-tiny
13
  pipeline_tag: automatic-speech-recognition
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
 
16
  # Whisper Tiny JA LoRA (ReazonSpeech)
17
 
18
  LoRA adapter fine-tuned from `openai/whisper-tiny` for Japanese ASR.
19
 
20
- - Base model: `openai/whisper-tiny`
21
- - Training method: LoRA (`q_proj`, `v_proj`)
22
- - Dataset: `reazon-research/reazonspeech` (gated)
23
- - Language: Japanese (`ja`)
24
-
25
  ## Model Type
26
 
27
  This repository contains **LoRA adapter weights only**.
28
- To run inference, load this adapter on top of `openai/whisper-tiny`.
 
 
 
 
 
29
 
30
  ## Training Setup
31
 
32
- - Epochs: `3`
33
  - Learning rate: `1e-5`
 
34
  - LoRA r / alpha / dropout: `16 / 32 / 0.05`
35
- - Batch size: `32` (or your actual value)
36
  - Framework: `transformers`, `peft`
 
37
 
38
- ## Intended Use
39
 
40
- - Japanese speech-to-text transcription
41
- - Lightweight adaptation with small trainable parameter count
 
 
 
 
 
42
 
43
- ## Limitations
44
 
45
- - Performance may degrade on domain/audio conditions not covered by ReazonSpeech
46
- - Not evaluated for safety-critical use cases
47
- - Dataset access requires accepted terms on Hugging Face
48
 
49
- ## Evaluation
 
50
 
51
- Fill with your real metrics:
52
 
53
- - CER: `TODO`
54
- - WER: `TODO`
55
- - Eval split: `TODO`
56
 
57
  ## Load Adapter
58
 
@@ -61,7 +88,7 @@ from transformers import WhisperForConditionalGeneration, WhisperProcessor
61
  from peft import PeftModel
62
 
63
  base_model_id = "openai/whisper-tiny"
64
- adapter_id = "dungca/whisper-tiny-ja-lora" # replace if needed
65
 
66
  processor = WhisperProcessor.from_pretrained(base_model_id)
67
  base_model = WhisperForConditionalGeneration.from_pretrained(base_model_id)
 
11
  library_name: transformers
12
  base_model: openai/whisper-tiny
13
  pipeline_tag: automatic-speech-recognition
14
+ datasets:
15
+ - reazon-research/reazonspeech
16
+ metrics:
17
+ - cer
18
+ - loss
19
+ model-index:
20
+ - name: whisper-tiny-ja-lora
21
+ results:
22
+ - task:
23
+ type: automatic-speech-recognition
24
+ name: Automatic Speech Recognition
25
+ dataset:
26
+ name: japanese-asr/ja_asr.reazonspeech_test
27
+ type: japanese-asr/ja_asr.reazonspeech_test
28
+ split: test
29
+ metrics:
30
+ - type: cer
31
+ name: Character Error Rate (CER)
32
+ value: 0.52497
33
+ - type: loss
34
+ name: Eval Loss
35
+ value: 1.17656
36
  ---
37
 
38
  # Whisper Tiny JA LoRA (ReazonSpeech)
39
 
40
  LoRA adapter fine-tuned from `openai/whisper-tiny` for Japanese ASR.
41
 
 
 
 
 
 
42
  ## Model Type
43
 
44
  This repository contains **LoRA adapter weights only**.
45
+ Use it on top of `openai/whisper-tiny`.
46
+
47
+ - Base model: `openai/whisper-tiny`
48
+ - Language: Japanese (`ja`)
49
+ - Training method: LoRA (`q_proj`, `v_proj`)
50
+ - Dataset: `reazon-research/reazonspeech` (gated)
51
 
52
  ## Training Setup
53
 
54
+ - Epochs (configured): `3`
55
  - Learning rate: `1e-5`
56
+ - Batch size: `32`
57
  - LoRA r / alpha / dropout: `16 / 32 / 0.05`
 
58
  - Framework: `transformers`, `peft`
59
+ - Runtime: Kaggle GPU P100
60
 
61
+ ## Evaluation (Latest W&B Run)
62
 
63
+ - `eval/cer`: **0.52497** (52.50%)
64
+ - `eval/loss`: **1.17656**
65
+ - `eval/runtime`: **162.422 s**
66
+ - `eval/samples_per_second`: **12.314**
67
+ - `eval/steps_per_second`: **0.77**
68
+ - `train/global_step`: **3000**
69
+ - `train/epoch`: **1.54719**
70
 
71
+ > Note: WER was not logged in this run.
72
 
73
+ ## Intended Use
 
 
74
 
75
+ - Japanese speech-to-text transcription
76
+ - Lightweight adapter training and deployment
77
 
78
+ ## Limitations
79
 
80
+ - Quality depends on domain/audio condition match with training data
81
+ - Not validated for safety-critical production use
82
+ - Requires accepted access to gated dataset when reproducing training
83
 
84
  ## Load Adapter
85
 
 
88
  from peft import PeftModel
89
 
90
  base_model_id = "openai/whisper-tiny"
91
+ adapter_id = "dungca/whisper-tiny-ja-lora"
92
 
93
  processor = WhisperProcessor.from_pretrained(base_model_id)
94
  base_model = WhisperForConditionalGeneration.from_pretrained(base_model_id)