| import os |
| os.environ["CUDA_VISIBLE_DEVICES"] = "0,1" |
|
|
| from transformers import CamembertTokenizer, CamembertForSequenceClassification, CamembertConfig |
| from transformers import Trainer, TrainingArguments |
|
|
| import pandas as pd |
| import numpy as np |
|
|
| from loadDataSet import loadData, labels_to_numeric |
| from helpers import compute_max_sent_length, get_device, set_seed |
| from bert_utils import ( |
| FrenchDataset, |
| compute_metrics, |
| ) |
|
|
| from nltk.tokenize import sent_tokenize |
|
|
| set_seed(1) |
|
|
| if __name__ == "__main__": |
| |
| device = get_device() |
|
|
| |
| base_path = "../code/" |
| train_path = base_path + "train_slices.txt" |
| val_path = base_path + "val_slices.txt" |
|
|
| |
| trainSamples, trainLabels = loadData("train", train_path) |
| valSamples, valLabels = loadData("validation", val_path) |
|
|
| print("Initial train size: %d" % len(trainSamples)) |
| print("Val size: %d" % len(valSamples)) |
|
|
| |
| print("Loading CamemBERT tokenizer...") |
| tokenizer = CamembertTokenizer.from_pretrained("camembert-base") |
|
|
| |
| |
| |
| |
|
|
|
|
| |
| max_len = 128 |
|
|
| |
| |
| trainLabels = labels_to_numeric(trainLabels) |
| valLabels = labels_to_numeric(valLabels) |
|
|
|
|
| |
| train_encodings = tokenizer(trainSamples, truncation=True, padding=True, max_length=max_len) |
| |
| valid_encodings = tokenizer(valSamples, truncation=True, padding=True, max_length=max_len) |
| |
| |
| train_dataset = FrenchDataset(train_encodings, trainLabels) |
| valid_dataset = FrenchDataset(valid_encodings, valLabels) |
|
|
| |
| config = CamembertConfig.from_pretrained("camembert-base", output_hidden_states=True) |
| model = CamembertForSequenceClassification.from_pretrained("camembert-base", num_labels=4).to(device) |
|
|
| |
| training_args = TrainingArguments( |
| output_dir="./bert_models_saved/out_fold", |
| num_train_epochs=30, |
| per_device_train_batch_size=32, |
| per_device_eval_batch_size=32, |
| warmup_steps=500, |
| weight_decay=0.01, |
| logging_dir='./logs', |
| load_best_model_at_end=True, |
| |
| logging_steps=250, |
| eval_steps=250, |
| |
| save_total_limit=5, |
| save_strategy="steps", |
| evaluation_strategy="steps", |
| ) |
|
|
| trainer = Trainer( |
| model=model, |
| args=training_args, |
| train_dataset=train_dataset, |
| eval_dataset=valid_dataset, |
| compute_metrics=compute_metrics, |
| ) |
|
|
| |
| trainer.train() |
|
|
| |
| trainer.save_model("./bert_models_saved/out_fold") |
|
|
| |
| trainer.evaluate() |
|
|
| |
| model.save_pretrained("./bert_models_saved/best_model/") |
| tokenizer.save_pretrained("./bert_models_saved/best_model/") |
|
|
|
|
|
|