aarath97 commited on
Commit
3f32f3d
·
verified ·
1 Parent(s): 44abecb

Upload final_code.py

Browse files
Files changed (1) hide show
  1. final_code.py +173 -0
final_code.py ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from torch.utils.data import Dataset, DataLoader
3
+ from transformers import MT5ForConditionalGeneration, MT5Tokenizer, AdamW
4
+ from transformers import AutoModel, AutoTokenizer
5
+ from sklearn.metrics.pairwise import cosine_similarity
6
+ import pandas as pd
7
+ import matplotlib.pyplot as plt
8
+ import numpy as np
9
+ from huggingface_hub import HfApi, HfFolder, Repository, notebook_login, create_repo, upload_folder
10
+ import os
11
+ import shutil
12
+
13
+ # ========== CONFIG ==========
14
+ HF_USERNAME = "aarath97"
15
+ HF_REPO = "mt5-dogri-translation"
16
+ MODEL_NAME = "google/mt5-large"
17
+ BATCH_SIZE = 2
18
+ LR = 1e-5
19
+ DPO_STEPS = 100
20
+ HGRL_STEPS = 100
21
+ COMBINED_STEPS = 50
22
+ GAMMA = 3.5
23
+ ALPHA = 0.5
24
+ BETA = 0.5
25
+
26
+ # ========== LOAD DATA ==========
27
+ df = pd.read_excel("dogri_train.xlsx")
28
+ train_data = list(zip(df['Dogri'], df['English'], df['Unpreffered']))
29
+
30
+ # ========== TOKENIZERS ==========
31
+ tokenizer = MT5Tokenizer.from_pretrained(MODEL_NAME)
32
+ sbert = AutoModel.from_pretrained("sentence-transformers/all-MiniLM-L6-v2")
33
+ sbert_tokenizer = AutoTokenizer.from_pretrained("sentence-transformers/all-MiniLM-L6-v2")
34
+
35
+ # ========== UTILITIES ==========
36
+ def compute_similarity(sent1, sent2):
37
+ emb1 = sbert(**sbert_tokenizer(sent1, return_tensors='pt')).last_hidden_state.mean(1)
38
+ emb2 = sbert(**sbert_tokenizer(sent2, return_tensors='pt')).last_hidden_state.mean(1)
39
+ return cosine_similarity(emb1.detach().numpy(), emb2.detach().numpy())[0][0]
40
+
41
+ def hyper_gamma_reward(rho):
42
+ return rho * np.exp(-GAMMA * (1 - rho))
43
+
44
+ # ========== DATASET ==========
45
+ class DogriDataset(Dataset):
46
+ def __init__(self, data):
47
+ self.data = data
48
+
49
+ def __len__(self):
50
+ return len(self.data)
51
+
52
+ def __getitem__(self, idx):
53
+ return self.data[idx]
54
+
55
+ dataloader = DataLoader(DogriDataset(train_data), batch_size=BATCH_SIZE, shuffle=True)
56
+
57
+ # ========== TRAINING ==========
58
+ model = MT5ForConditionalGeneration.from_pretrained(MODEL_NAME).to("cuda")
59
+ optimizer = AdamW(model.parameters(), lr=LR)
60
+
61
+ dpo_losses, hgrl_losses, final_losses = [], [], []
62
+
63
+ # ---- DPO Training ----
64
+ for step in range(DPO_STEPS):
65
+ batch = next(iter(dataloader))
66
+ loss_batch = []
67
+ for src, ref, unpref in zip(*batch):
68
+ input_ids = tokenizer(src, return_tensors='pt', truncation=True, padding=True).input_ids.to("cuda")
69
+ ref_ids = tokenizer(ref, return_tensors='pt', truncation=True, padding=True).input_ids.to("cuda")
70
+ unpref_ids = tokenizer(unpref, return_tensors='pt', truncation=True, padding=True).input_ids.to("cuda")
71
+
72
+ ref_logprob = model(input_ids=input_ids, labels=ref_ids).loss
73
+ unpref_logprob = model(input_ids=input_ids, labels=unpref_ids).loss
74
+
75
+ logit_diff = -ref_logprob.item() + unpref_logprob.item()
76
+ beta = 1.0
77
+ loss = -torch.log(torch.sigmoid(torch.tensor(beta * logit_diff)))
78
+ loss_batch.append(loss)
79
+
80
+ loss_val = torch.stack(loss_batch).mean()
81
+ loss_val.backward()
82
+ optimizer.step()
83
+ optimizer.zero_grad()
84
+ dpo_losses.append(loss_val.item())
85
+
86
+ # ---- HGRL Training ----
87
+ for step in range(HGRL_STEPS):
88
+ batch = next(iter(dataloader))
89
+ loss_batch = []
90
+ for src, ref, _ in zip(*batch):
91
+ input_ids = tokenizer(src, return_tensors='pt').input_ids.to("cuda")
92
+ gen_ids = model.generate(input_ids)
93
+ gen_text = tokenizer.decode(gen_ids[0], skip_special_tokens=True)
94
+
95
+ rho = compute_similarity(gen_text, ref)
96
+ reward = hyper_gamma_reward(rho)
97
+
98
+ labels = tokenizer(gen_text, return_tensors='pt').input_ids.to("cuda")
99
+ logprob = model(input_ids=input_ids, labels=labels).loss
100
+
101
+ loss = -reward * logprob
102
+ loss_batch.append(loss)
103
+
104
+ loss_val = torch.stack(loss_batch).mean()
105
+ loss_val.backward()
106
+ optimizer.step()
107
+ optimizer.zero_grad()
108
+ hgrl_losses.append(loss_val.item())
109
+
110
+ # ---- Combined Training ----
111
+ for step in range(COMBINED_STEPS):
112
+ batch = next(iter(dataloader))
113
+ loss_dpo_batch, loss_hgrl_batch = [], []
114
+ for src, ref, unpref in zip(*batch):
115
+ input_ids = tokenizer(src, return_tensors='pt').input_ids.to("cuda")
116
+ ref_ids = tokenizer(ref, return_tensors='pt').input_ids.to("cuda")
117
+ unpref_ids = tokenizer(unpref, return_tensors='pt').input_ids.to("cuda")
118
+
119
+ logprob_ref = model(input_ids=input_ids, labels=ref_ids).loss
120
+ logprob_unpref = model(input_ids=input_ids, labels=unpref_ids).loss
121
+ dpo_loss = -torch.log(torch.sigmoid(torch.tensor(logprob_unpref.item() - logprob_ref.item())))
122
+ loss_dpo_batch.append(dpo_loss)
123
+
124
+ gen_ids = model.generate(input_ids)
125
+ gen_text = tokenizer.decode(gen_ids[0], skip_special_tokens=True)
126
+ rho = compute_similarity(gen_text, ref)
127
+ reward = hyper_gamma_reward(rho)
128
+
129
+ labels = tokenizer(gen_text, return_tensors='pt').input_ids.to("cuda")
130
+ logprob = model(input_ids=input_ids, labels=labels).loss
131
+ hgrl_loss = -reward * logprob
132
+ loss_hgrl_batch.append(hgrl_loss)
133
+
134
+ loss_dpo_mean = torch.stack(loss_dpo_batch).mean()
135
+ loss_hgrl_mean = torch.stack(loss_hgrl_batch).mean()
136
+ combined_loss = ALPHA * loss_dpo_mean + BETA * loss_hgrl_mean
137
+ combined_loss.backward()
138
+ optimizer.step()
139
+ optimizer.zero_grad()
140
+ final_losses.append(combined_loss.item())
141
+
142
+ # ========== SAVE OUTPUTS ==========
143
+ plt.plot(dpo_losses, label="DPO")
144
+ plt.plot(hgrl_losses, label="HGRL")
145
+ plt.plot(final_losses, label="Combined")
146
+ plt.xlabel("Steps")
147
+ plt.ylabel("Loss")
148
+ plt.legend()
149
+ plt.savefig("loss_curve.png")
150
+
151
+ with open("loss_report.txt", "w") as f:
152
+ f.write("DPO Final Loss: {:.4f}\n".format(dpo_losses[-1]))
153
+ f.write("HGRL Final Loss: {:.4f}\n".format(hgrl_losses[-1]))
154
+ f.write("Combined Final Loss: {:.4f}\n".format(final_losses[-1]))
155
+
156
+ # ========== TEST AND SAVE TRANSLATIONS ==========
157
+ test_df = pd.read_excel("in22conv.xlsx")
158
+ test_outputs = []
159
+ for line in test_df.iloc[:, 0].tolist():
160
+ input_ids = tokenizer(line, return_tensors='pt').input_ids.to("cuda")
161
+ outputs = model.generate(input_ids)
162
+ translation = tokenizer.decode(outputs[0], skip_special_tokens=True)
163
+ test_outputs.append(translation)
164
+
165
+ output_df = pd.DataFrame({"Dogri": test_df.iloc[:, 0], "English": test_outputs})
166
+ output_df.to_excel("translated_output.xlsx", index=False)
167
+
168
+ # ========== PUSH TO HUGGING FACE ==========
169
+ model.save_pretrained("mt5-dogri")
170
+ tokenizer.save_pretrained("mt5-dogri")
171
+ create_repo(f"{HF_USERNAME}/{HF_REPO}", private=False, exist_ok=True)
172
+ upload_folder(repo_id=f"{HF_USERNAME}/{HF_REPO}", folder_path="mt5-dogri")
173
+ print("Model uploaded successfully!")