YaoyaoChang
commited on
Commit
Β·
2d0b945
1
Parent(s):
4c769d9
update README
Browse files
README.md
CHANGED
|
@@ -29,14 +29,15 @@ library_name: transformers
|
|
| 29 |
|
| 30 |
## π₯ Key Features
|
| 31 |
|
| 32 |
-
- **π 60-
|
| 33 |
-
Unlike conventional ASR models that slice audio into short chunks (often losing global context), VibeVoice ASR accepts up to **60 minutes** of continuous audio input within 64K length. This ensures consistent speaker tracking and semantic coherence across the entire hour.
|
| 34 |
|
| 35 |
-
- **π€
|
| 36 |
-
Users can provide customized
|
| 37 |
|
| 38 |
- **π Rich Transcription (Who, When, What)**:
|
| 39 |
-
The model performs ASR,
|
|
|
|
| 40 |
|
| 41 |
|
| 42 |
|
|
|
|
| 29 |
|
| 30 |
## π₯ Key Features
|
| 31 |
|
| 32 |
+
- **π 60-minute Single-Pass Processing**:
|
| 33 |
+
Unlike conventional ASR models that slice audio into short chunks (often losing global context), VibeVoice ASR accepts up to **60 minutes** of continuous audio input within 64K token length. This ensures consistent speaker tracking and semantic coherence across the entire hour.
|
| 34 |
|
| 35 |
+
- **π€ Customized Hotwords**:
|
| 36 |
+
Users can provide customized hotwords (e.g., specific names, technical terms, or background info) to guide the recognition process, significantly improving accuracy on domain-specific content.
|
| 37 |
|
| 38 |
- **π Rich Transcription (Who, When, What)**:
|
| 39 |
+
The model jointly performs ASR, diarization, and timestamping, producing a structured output that indicates *who* said *what* and *when*.
|
| 40 |
+
|
| 41 |
|
| 42 |
|
| 43 |
|