File size: 5,479 Bytes
b778085 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 |
from transformers import Wav2Vec2FeatureExtractor, Wav2Vec2Model, Wav2Vec2Processor, Wav2Vec2CTCTokenizer
import torchaudio
import torch
from huggingface_hub import notebook_login
from datasets import load_dataset
import re
from transformers import Wav2Vec2ForCTC
from torch.utils.data import DataLoader
organization_name = "ASR-Erzya-Final-Project"
#organization_name = "zmmccormick3"
dataset_name = "asr_erzya_final_data"
dataset = load_dataset(f"{organization_name}/{dataset_name}")
processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
model = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-960h")
#://huggingface.co/datasets/ASR-Erzya-Final-Project/asr_erzya_final_data/tree/main
def preprocess_data(batch):
inputs = processor(batch["audio"], return_tensors="pt", sampling_rate=16000)
targets = processor(batch["text"], return_tensors="pt", padding=True)
# Ensure that target tensor is of shape (batch_size, sequence_length)
targets["input_ids"] = targets["input_ids"].unsqueeze(0)
return inputs, targets
# Create DataLoader for training data
train_data_loader = DataLoader(dataset["train"], batch_size=4, collate_fn=preprocess_data)
# Training loop
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
optimizer = torch.optim.AdamW(model.parameters(), lr=1e-4)
for epoch in range(5): # Set the number of epochs you want to train for
model.train()
for batch in train_data_loader:
inputs, targets = batch
inputs = {key: value.to(device) for key, value in inputs.items()}
targets = {key: value.to(device) for key, value in targets.items()}
optimizer.zero_grad()
outputs = model(**inputs, labels=targets["input_ids"])
loss = outputs.loss
loss.backward()
optimizer.step()
# Save the trained model
model.save_pretrained("your-organization/your-wav2vec-ctc-model")
processor.save_pretrained("your-organization/your-wav2vec-ctc-model")
#common_voice_train = load_dataset(f"{organization_name}/{dataset_name}", split="train")
#common_voice_test = load_dataset(f"{organization_name}/{dataset_name}", split="test")
# notebook_login()
common_voice_train = load_dataset("common_voice", "myv", split="train")
common_voice_test = load_dataset("common_voice", "myv", split="test")
# Remove unnecessary columns
common_voice_train = common_voice_train.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
common_voice_test = common_voice_test.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
chars_to_remove_regex = '[\,\?\.\!\-\;\:\"\“\%\‘\”\�\']'
def remove_special_characters(batch):
batch["sentence"] = re.sub(chars_to_remove_regex, '', batch["sentence"]).lower()
return batch
common_voice_train = common_voice_train.map(remove_special_characters)
common_voice_test = common_voice_test.map(remove_special_characters)
def replace_hatted_characters(batch):
batch["sentence"] = re.sub('[â]', 'a', batch["sentence"])
# Add similar lines for other hat characters if needed
return batch
common_voice_train = common_voice_train.map(replace_hatted_characters)
common_voice_test = common_voice_test.map(replace_hatted_characters)
def extract_all_chars(batch):
all_text = " ".join(batch["sentence"])
vocab = list(set(all_text))
return {"vocab": [vocab], "all_text": [all_text]}
vocab_train = common_voice_train.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_train.column_names)
vocab_test = common_voice_test.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_test.column_names)
vocab_list = list(set(vocab_train["vocab"][0]) | set(vocab_test["vocab"][0]))
vocab_dict = {v: k for k, v in enumerate(sorted(vocab_list))}
vocab_dict["|"] = vocab_dict[" "]
del vocab_dict[" "]
vocab_dict["[UNK]"] = len(vocab_dict)
vocab_dict["[PAD]"] = len(vocab_dict)
import json
with open('vocab.json', 'w') as vocab_file:
json.dump(vocab_dict, vocab_file)
tokenizer = Wav2Vec2CTCTokenizer.from_pretrained("./", unk_token="[UNK]", pad_token="[PAD]", word_delimiter_token="|")
repo_name = "wav2vec2-large-xls-r-300m-tr-colab" # TK repository name
tokenizer.push_to_hub(repo_name)
# tokenizer = Wav2Vec2CTCTokenizer.from_pretrained() # TK
# feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h")
def downsample(batch):
resample = torchaudio.transforms.Resample(orig_freq=48_000, new_freq=16_000)
batch["audio"] = resample(batch["audio"])
return batch
common_voice_train = common_voice_train.map(downsample)
common_voice_test = common_voice_test.map(downsample)
feature_extractor = Wav2Vec2FeatureExtractor(feature_size=1, sampling_rate=16000, padding_value=0.0, do_normalize=True, return_attention_mask=True)
# processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
processor = Wav2Vec2Processor(feature_extractor=feature_extractor, tokenizer=tokenizer)
'''
model = Wav2Vec2Model.from_pretrained("Link/to/HuggingfaceModel")
array, fs = torchaudio.load("/Local/link/to/your/audio.wav"
input_input = processor(array.squeeze(), sampling_rate=fs, return_tensors="pt")
with torch.no_grad():
outputs = model(**input_input)
print(f"Hidden state shape: {outputs.last_hidden_state.numpy().shape}")
'''
|