File size: 5,479 Bytes
b778085
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
from transformers import Wav2Vec2FeatureExtractor, Wav2Vec2Model, Wav2Vec2Processor, Wav2Vec2CTCTokenizer
import torchaudio
import torch
from huggingface_hub import notebook_login
from datasets import load_dataset
import re
from transformers import Wav2Vec2ForCTC
from torch.utils.data import DataLoader

organization_name = "ASR-Erzya-Final-Project"
#organization_name = "zmmccormick3"
dataset_name = "asr_erzya_final_data"

dataset = load_dataset(f"{organization_name}/{dataset_name}")

processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
model = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-960h")

#://huggingface.co/datasets/ASR-Erzya-Final-Project/asr_erzya_final_data/tree/main

def preprocess_data(batch):
    inputs = processor(batch["audio"], return_tensors="pt", sampling_rate=16000)
    targets = processor(batch["text"], return_tensors="pt", padding=True)

    # Ensure that target tensor is of shape (batch_size, sequence_length)
    targets["input_ids"] = targets["input_ids"].unsqueeze(0)

    return inputs, targets

# Create DataLoader for training data
train_data_loader = DataLoader(dataset["train"], batch_size=4, collate_fn=preprocess_data)

# Training loop
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
optimizer = torch.optim.AdamW(model.parameters(), lr=1e-4)

for epoch in range(5):  # Set the number of epochs you want to train for
    model.train()
    for batch in train_data_loader:
        inputs, targets = batch
        inputs = {key: value.to(device) for key, value in inputs.items()}
        targets = {key: value.to(device) for key, value in targets.items()}

        optimizer.zero_grad()
        outputs = model(**inputs, labels=targets["input_ids"])
        loss = outputs.loss
        loss.backward()
        optimizer.step()

# Save the trained model
model.save_pretrained("your-organization/your-wav2vec-ctc-model")
processor.save_pretrained("your-organization/your-wav2vec-ctc-model")


#common_voice_train = load_dataset(f"{organization_name}/{dataset_name}", split="train")
#common_voice_test = load_dataset(f"{organization_name}/{dataset_name}", split="test")

# notebook_login()

common_voice_train = load_dataset("common_voice", "myv", split="train")
common_voice_test = load_dataset("common_voice", "myv", split="test")

# Remove unnecessary columns
common_voice_train = common_voice_train.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
common_voice_test = common_voice_test.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])

chars_to_remove_regex = '[\,\?\.\!\-\;\:\"\“\%\‘\”\�\']'

def remove_special_characters(batch):
    batch["sentence"] = re.sub(chars_to_remove_regex, '', batch["sentence"]).lower()
    return batch

common_voice_train = common_voice_train.map(remove_special_characters)
common_voice_test = common_voice_test.map(remove_special_characters)

def replace_hatted_characters(batch):
    batch["sentence"] = re.sub('[â]', 'a', batch["sentence"])
    # Add similar lines for other hat characters if needed
    return batch

common_voice_train = common_voice_train.map(replace_hatted_characters)
common_voice_test = common_voice_test.map(replace_hatted_characters)

def extract_all_chars(batch):
    all_text = " ".join(batch["sentence"])
    vocab = list(set(all_text))
    return {"vocab": [vocab], "all_text": [all_text]}

vocab_train = common_voice_train.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_train.column_names)
vocab_test = common_voice_test.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_test.column_names)

vocab_list = list(set(vocab_train["vocab"][0]) | set(vocab_test["vocab"][0]))
vocab_dict = {v: k for k, v in enumerate(sorted(vocab_list))}
vocab_dict["|"] = vocab_dict[" "]
del vocab_dict[" "]
vocab_dict["[UNK]"] = len(vocab_dict)
vocab_dict["[PAD]"] = len(vocab_dict)

import json
with open('vocab.json', 'w') as vocab_file:
    json.dump(vocab_dict, vocab_file)

tokenizer = Wav2Vec2CTCTokenizer.from_pretrained("./", unk_token="[UNK]", pad_token="[PAD]", word_delimiter_token="|")

repo_name = "wav2vec2-large-xls-r-300m-tr-colab" # TK repository name
tokenizer.push_to_hub(repo_name)

# tokenizer = Wav2Vec2CTCTokenizer.from_pretrained() # TK
# feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h")


def downsample(batch):
    resample = torchaudio.transforms.Resample(orig_freq=48_000, new_freq=16_000)
    batch["audio"] = resample(batch["audio"])
    return batch


common_voice_train = common_voice_train.map(downsample)
common_voice_test = common_voice_test.map(downsample)


feature_extractor = Wav2Vec2FeatureExtractor(feature_size=1, sampling_rate=16000, padding_value=0.0, do_normalize=True, return_attention_mask=True)
# processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
processor = Wav2Vec2Processor(feature_extractor=feature_extractor, tokenizer=tokenizer)

'''
model = Wav2Vec2Model.from_pretrained("Link/to/HuggingfaceModel")
array, fs = torchaudio.load("/Local/link/to/your/audio.wav"
input_input = processor(array.squeeze(), sampling_rate=fs, return_tensors="pt")
with torch.no_grad():
    outputs = model(**input_input)
print(f"Hidden state shape: {outputs.last_hidden_state.numpy().shape}")
'''