zmmccormick3's picture
Upload main.py
b778085
from transformers import Wav2Vec2FeatureExtractor, Wav2Vec2Model, Wav2Vec2Processor, Wav2Vec2CTCTokenizer
import torchaudio
import torch
from huggingface_hub import notebook_login
from datasets import load_dataset
import re
from transformers import Wav2Vec2ForCTC
from torch.utils.data import DataLoader
organization_name = "ASR-Erzya-Final-Project"
#organization_name = "zmmccormick3"
dataset_name = "asr_erzya_final_data"
dataset = load_dataset(f"{organization_name}/{dataset_name}")
processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
model = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-960h")
#://huggingface.co/datasets/ASR-Erzya-Final-Project/asr_erzya_final_data/tree/main
def preprocess_data(batch):
inputs = processor(batch["audio"], return_tensors="pt", sampling_rate=16000)
targets = processor(batch["text"], return_tensors="pt", padding=True)
# Ensure that target tensor is of shape (batch_size, sequence_length)
targets["input_ids"] = targets["input_ids"].unsqueeze(0)
return inputs, targets
# Create DataLoader for training data
train_data_loader = DataLoader(dataset["train"], batch_size=4, collate_fn=preprocess_data)
# Training loop
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
optimizer = torch.optim.AdamW(model.parameters(), lr=1e-4)
for epoch in range(5): # Set the number of epochs you want to train for
model.train()
for batch in train_data_loader:
inputs, targets = batch
inputs = {key: value.to(device) for key, value in inputs.items()}
targets = {key: value.to(device) for key, value in targets.items()}
optimizer.zero_grad()
outputs = model(**inputs, labels=targets["input_ids"])
loss = outputs.loss
loss.backward()
optimizer.step()
# Save the trained model
model.save_pretrained("your-organization/your-wav2vec-ctc-model")
processor.save_pretrained("your-organization/your-wav2vec-ctc-model")
#common_voice_train = load_dataset(f"{organization_name}/{dataset_name}", split="train")
#common_voice_test = load_dataset(f"{organization_name}/{dataset_name}", split="test")
# notebook_login()
common_voice_train = load_dataset("common_voice", "myv", split="train")
common_voice_test = load_dataset("common_voice", "myv", split="test")
# Remove unnecessary columns
common_voice_train = common_voice_train.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
common_voice_test = common_voice_test.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "segment", "up_votes"])
chars_to_remove_regex = '[\,\?\.\!\-\;\:\"\“\%\‘\”\�\']'
def remove_special_characters(batch):
batch["sentence"] = re.sub(chars_to_remove_regex, '', batch["sentence"]).lower()
return batch
common_voice_train = common_voice_train.map(remove_special_characters)
common_voice_test = common_voice_test.map(remove_special_characters)
def replace_hatted_characters(batch):
batch["sentence"] = re.sub('[â]', 'a', batch["sentence"])
# Add similar lines for other hat characters if needed
return batch
common_voice_train = common_voice_train.map(replace_hatted_characters)
common_voice_test = common_voice_test.map(replace_hatted_characters)
def extract_all_chars(batch):
all_text = " ".join(batch["sentence"])
vocab = list(set(all_text))
return {"vocab": [vocab], "all_text": [all_text]}
vocab_train = common_voice_train.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_train.column_names)
vocab_test = common_voice_test.map(extract_all_chars, batched=True, batch_size=-1, keep_in_memory=True, remove_columns=common_voice_test.column_names)
vocab_list = list(set(vocab_train["vocab"][0]) | set(vocab_test["vocab"][0]))
vocab_dict = {v: k for k, v in enumerate(sorted(vocab_list))}
vocab_dict["|"] = vocab_dict[" "]
del vocab_dict[" "]
vocab_dict["[UNK]"] = len(vocab_dict)
vocab_dict["[PAD]"] = len(vocab_dict)
import json
with open('vocab.json', 'w') as vocab_file:
json.dump(vocab_dict, vocab_file)
tokenizer = Wav2Vec2CTCTokenizer.from_pretrained("./", unk_token="[UNK]", pad_token="[PAD]", word_delimiter_token="|")
repo_name = "wav2vec2-large-xls-r-300m-tr-colab" # TK repository name
tokenizer.push_to_hub(repo_name)
# tokenizer = Wav2Vec2CTCTokenizer.from_pretrained() # TK
# feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h")
def downsample(batch):
resample = torchaudio.transforms.Resample(orig_freq=48_000, new_freq=16_000)
batch["audio"] = resample(batch["audio"])
return batch
common_voice_train = common_voice_train.map(downsample)
common_voice_test = common_voice_test.map(downsample)
feature_extractor = Wav2Vec2FeatureExtractor(feature_size=1, sampling_rate=16000, padding_value=0.0, do_normalize=True, return_attention_mask=True)
# processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h")
processor = Wav2Vec2Processor(feature_extractor=feature_extractor, tokenizer=tokenizer)
'''
model = Wav2Vec2Model.from_pretrained("Link/to/HuggingfaceModel")
array, fs = torchaudio.load("/Local/link/to/your/audio.wav"
input_input = processor(array.squeeze(), sampling_rate=fs, return_tensors="pt")
with torch.no_grad():
outputs = model(**input_input)
print(f"Hidden state shape: {outputs.last_hidden_state.numpy().shape}")
'''