moinbach7 commited on
Commit
a7e2179
·
verified ·
1 Parent(s): 5d4c68e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -54
README.md CHANGED
@@ -1,54 +1,57 @@
1
- ---
2
- license: gpl-3.0
3
- language:
4
- - en
5
- pipeline_tag: text2text-generation
6
- tags:
7
- - code
8
- - asr
9
- - inverse text normalization
10
- datasets:
11
- - text-normalization-challenge-english-language
12
- ---
13
-
14
- ---
15
-
16
- ---
17
-
18
- # asr_inverse_text_normalization
19
-
20
- Finetuned a facebook/bart-base Pretrained model on the ASR inverse text normalization dataset by treating it as a seq2seq task. Other approaches which may be considered is by considering it as a TokenClassification task and the one mentioned here https://machinelearning.apple.com/research/inverse-text-normal.
21
-
22
- ## Model description
23
-
24
- BART (Bidirectional and Auto-Regressive Transformers) is a pre-trained transformer-based neural network model developed by Facebook AI Research (FAIR) for various natural language processing (NLP) tasks
25
-
26
- The BART architecture is based on the Transformer model, which is a type of neural network architecture that processes sequential input data, such as text, by applying self-attention mechanisms to capture the relationships between different words in the input sequence.
27
- BART includes both auto-regressive and bidirectional encoder-decoder transformer architectures, which enable it to perform both generation and prediction tasks
28
-
29
- BART was trained on a diverse range of NLP tasks, including machine translation, summarization, and question answering, and has shown strong performance across multiple benchmarks.
30
- Its training process involves corrupting text with different types of noise and training the model to reconstruct the original text, which has been shown to improve the model's ability to generalize to new tasks and outperform other pre-trained language models like GPT and BERT
31
-
32
- The model flavour which was chosen is that of "facebook/bart-base" and columns "after" is used as the source while "before" column is used as the targets.
33
-
34
- ## Intended uses & limitations
35
-
36
- This model can be used as an out-of-the-box solution to the invesrse text normalization which can convert ASR generated un-normalized text such as
37
- "my c v v for my card is five six seven and it expires on november twenty three" -> "my CVV for my card is 567 and it expires on November 23"
38
-
39
- The model needs to be explored for various min and max length setting at the time of generation for your specific usecase
40
-
41
- ### How to use
42
-
43
- ```python
44
-
45
- >>> from transformers import pipeline
46
- >>> generator = pipeline(model="pavanBuduguppa/asr_inverse_text_normalization")
47
-
48
- >>> generator("my c v v for my card is five six seven and it expires on november twenty three")
49
-
50
- ```
51
-
52
- ## Training data
53
-
54
- All credits and rights for the training data belongs to Google. The data was merely obtained and processed for this model and the original data can be found here https://www.kaggle.com/competitions/text-normalization-challenge-english-language/data
 
 
 
 
1
+ ---
2
+ license: gpl-3.0
3
+ language:
4
+ - en
5
+ pipeline_tag: text2text-generation
6
+ tags:
7
+ - code
8
+ - asr
9
+ - inverse text normalization
10
+ datasets:
11
+ - text-normalization-challenge-english-language
12
+ ---
13
+
14
+ # ASR Inverse Text Normalization
15
+
16
+ This repository provides a **fine-tuned BART model** for the task of **ASR Inverse Text Normalization (ITN)**.
17
+ The goal is to transform raw, unnormalized ASR transcripts into properly formatted text.
18
+
19
+ ---
20
+
21
+ ## Model Overview
22
+
23
+ **BART (Bidirectional and Auto-Regressive Transformers)** is a transformer-based model introduced by Facebook AI Research.
24
+ It is designed for both text understanding and generation tasks.
25
+
26
+ - **Architecture**: Encoder–Decoder Transformer with self-attention.
27
+ - **Pretraining objective**: Reconstruct original text from corrupted/noisy versions.
28
+ - **Applications**: Summarization, machine translation, question answering, and text normalization.
29
+
30
+ For this project:
31
+ - Base model: `facebook/bart-base`
32
+ - Training setup: Treated as a **sequence-to-sequence** problem
33
+ - Dataset: [Text Normalization Challenge - English Language (Kaggle)](https://www.kaggle.com/competitions/text-normalization-challenge-english-language/data)
34
+ - Columns used:
35
+ - Input: `"after"` (ASR-like text)
36
+ - Target: `"before"` (normalized text)
37
+
38
+ ---
39
+
40
+ ## Intended Use
41
+
42
+ The model can be applied directly to **normalize ASR outputs** in speech-to-text pipelines.
43
+
44
+ ---
45
+
46
+ ## Quickstart
47
+
48
+ ```python
49
+ from transformers import pipeline
50
+
51
+ # Load pipeline
52
+ generator = pipeline(model="pavanBuduguppa/asr_inverse_text_normalization")
53
+
54
+ # Run inference
55
+ result = generator("my c v v for my card is five six seven and it expires on november twenty three")
56
+ print(result)
57
+