| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
| from base.graph_recommender import GraphRecommender |
| from util.conf import OptionConf |
| from util.sampler import next_batch_pairwise |
| from util.loss_torch import l2_reg_loss, InfoNCE, batch_softmax_loss |
|
|
| |
|
|
| """ |
| Note: This version of code conducts feature dropout on the item embeddings |
| because items features are not always available in many academic datasets. |
| """ |
|
|
|
|
| class SSL4Rec(GraphRecommender): |
| def __init__(self, conf, training_set, test_set): |
| super(SSL4Rec, self).__init__(conf, training_set, test_set) |
| args = OptionConf(self.config['SSL4Rec']) |
| self.cl_rate = float(args['-alpha']) |
| self.tau = float(args['-tau']) |
| self.drop_rate = float(args['-drop']) |
| self.model = DNN_Encoder(self.data, self.emb_size, self.drop_rate, self.tau) |
|
|
| def train(self): |
| model = self.model.cuda() |
| optimizer = torch.optim.Adam(model.parameters(), lr=self.lRate) |
| for epoch in range(self.maxEpoch): |
| for n, batch in enumerate(next_batch_pairwise(self.data, self.batch_size)): |
| query_idx, item_idx, _neg = batch |
| model.train() |
| query_emb, item_emb = model(query_idx, item_idx) |
| rec_loss = batch_softmax_loss(query_emb, item_emb, self.tau) |
| cl_loss = self.cl_rate * model.cal_cl_loss(item_idx) |
| batch_loss = rec_loss + l2_reg_loss(self.reg, query_emb, item_emb) + cl_loss |
| |
| optimizer.zero_grad() |
| batch_loss.backward() |
| optimizer.step() |
| if n % 100 == 0: |
| print('training:', epoch + 1, 'batch', n, 'rec_loss:', rec_loss.item(), 'cl_loss', cl_loss.item()) |
| model.eval() |
| with torch.no_grad(): |
| self.query_emb, self.item_emb = self.model(list(range(self.data.user_num)),list(range(self.data.item_num))) |
| self.fast_evaluation(epoch) |
| self.query_emb, self.item_emb = self.best_query_emb, self.best_item_emb |
|
|
| def save(self): |
| with torch.no_grad(): |
| self.best_query_emb, self.best_item_emb = self.model.forward(list(range(self.data.user_num)),list(range(self.data.item_num))) |
|
|
| def predict(self, u): |
| u = self.data.get_user_id(u) |
| score = torch.matmul(self.query_emb[u], self.item_emb.transpose(0, 1)) |
| return score.cpu().numpy() |
|
|
|
|
| class DNN_Encoder(nn.Module): |
| def __init__(self, data, emb_size, drop_rate, temperature): |
| super(DNN_Encoder, self).__init__() |
| self.data = data |
| self.emb_size = emb_size |
| self.tau = temperature |
| self.user_tower = nn.Sequential( |
| nn.Linear(self.emb_size, 1024), |
| nn.ReLU(True), |
| nn.Linear(1024, 128), |
| nn.Sigmoid() |
| ) |
| self.item_tower = nn.Sequential( |
| nn.Linear(self.emb_size, 1024), |
| nn.ReLU(True), |
| nn.Linear(1024, 128), |
| nn.Sigmoid() |
| ) |
| self.dropout = nn.Dropout(drop_rate) |
| initializer = nn.init.xavier_uniform_ |
| self.initial_user_emb = nn.Parameter(initializer(torch.empty(self.data.user_num, self.emb_size))) |
| self.initial_item_emb = nn.Parameter(initializer(torch.empty(self.data.item_num, self.emb_size))) |
|
|
| def forward(self, q, x): |
| q_emb = self.initial_user_emb[q] |
| i_emb = self.initial_item_emb[x] |
|
|
| q_emb = self.user_tower(q_emb) |
| i_emb = self.item_tower(i_emb) |
|
|
| return q_emb, i_emb |
|
|
| def item_encoding(self, x): |
| i_emb = self.initial_item_emb[x] |
| i1_emb = self.dropout(i_emb) |
| i2_emb = self.dropout(i_emb) |
|
|
| i1_emb = self.item_tower(i1_emb) |
| i2_emb = self.item_tower(i2_emb) |
|
|
| return i1_emb, i2_emb |
|
|
| def cal_cl_loss(self, idx): |
| item_view1, item_view_2 = self.item_encoding(idx) |
| item_view1, item_view_2 = F.normalize(item_view1, dim=1), F.normalize(item_view_2, dim=1) |
| cl_loss = InfoNCE(item_view1, item_view_2, self.tau) |
| return cl_loss |
|
|