import pandas as pd import torch import torch.nn as nn import torch.optim as optim from gensim.models.fasttext import load_facebook_model from torch.utils.data import DataLoader, Dataset import numpy as np FASTTEXT_PATH = "FastText.bin" TRAIN_PATH = "TRAIN.tsv" TEST_PATH = "Test-1.tsv" print("Učitavanje FastText modela...") ft_model = load_facebook_model(FASTTEXT_PATH) embedding_dim = ft_model.vector_size def tokenize(text): return text.lower().split() class FastTextDataset(Dataset): def __init__(self, texts, labels): self.texts = [tokenize(text) for text in texts] self.labels = labels def __len__(self): return len(self.labels) def __getitem__(self, idx): tokens = self.texts[idx] vectors = [ft_model.wv[token] if token in ft_model.wv else np.zeros(embedding_dim) for token in tokens] max_len = 50 if len(vectors) > max_len: vectors = vectors[:max_len] else: vectors += [np.zeros(embedding_dim)] * (max_len - len(vectors)) return torch.tensor(vectors, dtype=torch.float32), torch.tensor(self.labels[idx], dtype=torch.long) class CNNClassifier(nn.Module): def __init__(self, embedding_dim, num_classes): super(CNNClassifier, self).__init__() self.conv1 = nn.Conv1d(embedding_dim, 100, kernel_size=3, padding=1) self.relu = nn.ReLU() self.pool = nn.AdaptiveMaxPool1d(1) self.dropout = nn.Dropout(0.5) self.fc = nn.Linear(100, num_classes) def forward(self, x): x = x.permute(0, 2, 1) x = self.relu(self.conv1(x)) x = self.pool(x).squeeze(2) x = self.dropout(x) return self.fc(x) class LSTMClassifier(nn.Module): def __init__(self, embedding_dim, hidden_dim, num_classes): super(LSTMClassifier, self).__init__() self.lstm = nn.LSTM(embedding_dim, hidden_dim, batch_first=True, bidirectional=True) self.dropout = nn.Dropout(0.5) self.fc = nn.Linear(hidden_dim * 2, num_classes) def forward(self, x): _, (hn, _) = self.lstm(x) x = torch.cat((hn[-2], hn[-1]), dim=1) # concatenate forward and backward hidden states x = self.dropout(x) return self.fc(x) print("Učitavanje podataka...") train_df = pd.read_csv(TRAIN_PATH, sep="\t").rename(columns={"Sentence": "text", "Label": "label"}) test_df = pd.read_csv(TEST_PATH, sep="\t").rename(columns={"Sentence": "text", "Label": "label"}) train_df["label"] = train_df["label"].astype(int) test_df["label"] = test_df["label"].astype(int) num_classes = train_df["label"].nunique() train_dataset = FastTextDataset(train_df["text"].tolist(), train_df["label"].tolist()) test_dataset = FastTextDataset(test_df["text"].tolist(), test_df["label"].tolist()) train_loader = DataLoader(train_dataset, batch_size=32, shuffle=True) test_loader = DataLoader(test_dataset, batch_size=32) device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def train(model, loader, optimizer, criterion): model.train() total_loss = 0 for x, y in loader: x, y = x.to(device), y.to(device) optimizer.zero_grad() output = model(x) loss = criterion(output, y) loss.backward() optimizer.step() total_loss += loss.item() return total_loss / len(loader) def evaluate(model, loader): model.eval() correct = total = 0 with torch.no_grad(): for x, y in loader: x, y = x.to(device), y.to(device) output = model(x) preds = torch.argmax(output, dim=1) correct += (preds == y).sum().item() total += y.size(0) return correct / total for model_type in ["LSTM", "CNN"]: print(f"\n==============================") print(f"Treniramo model: {model_type}") print(f"==============================") if model_type == "LSTM": model = LSTMClassifier(embedding_dim, hidden_dim=256, num_classes=num_classes) else: model = CNNClassifier(embedding_dim, num_classes=num_classes) model = model.to(device) optimizer = optim.Adam(model.parameters(), lr=1e-3) criterion = nn.CrossEntropyLoss() for epoch in range(1, 11): train_loss = train(model, train_loader, optimizer, criterion) test_acc = evaluate(model, test_loader) print(f"{model_type} | Epoch {epoch} | Loss: {train_loss:.4f} | Test Accuracy: {test_acc:.4f}") model_path = f"fasttext_{model_type.lower()}.pt" torch.save(model.state_dict(), model_path) print(f"{model_type} model spremljen kao: {model_path}")