GeneratingQuestions / HVU_QA /generate_question.py
DANGDOCAO's picture
HVU_QA
4b4a869 verified
import re
import torch
from difflib import SequenceMatcher
from transformers import T5Tokenizer, T5ForConditionalGeneration
from transformers.utils import logging as hf_logging
hf_logging.set_verbosity_error()
MODEL_DIR = "t5-viet-qg-finetuned"
PATTERN = re.compile(
r"""
“([^”]{3,120})”
| "([^"]{3,120})"
| \b(?:là|gồm|do|theo)\s+([^,.;:\n]{3,120})
| \b\d{4}\b
| \b(?:Điều|Khoản)\s+\d+\b
""",
re.VERBOSE | re.IGNORECASE,
)
def norm(s: str) -> str:
return re.sub(r"\s+", " ", s).strip()
def is_dup(q: str, qs: list[str], thr: float = 0.85) -> bool:
ql = q.lower()
for x in qs:
if SequenceMatcher(None, ql, x.lower()).ratio() >= thr:
return True
return False
def extract_answers(ctx: str, max_n: int = 60) -> list[str]:
ctx = norm(ctx)
answers, seen = [], set()
for m in PATTERN.finditer(ctx):
for g in m.groups():
if not g:
continue
g = norm(g)
k = g.lower()
if 3 <= len(g) <= 120 and k not in seen:
seen.add(k)
answers.append(g)
if len(answers) >= max_n:
return answers
if len(answers) < 8:
for i in range(0, min(len(ctx), 500), 60):
ch = norm(ctx[i : i + 60])
k = ch.lower()
if len(ch) >= 15 and k not in seen:
seen.add(k)
answers.append(ch)
if len(answers) >= max_n:
break
if not answers and ctx:
answers = [ctx[:120]]
return answers
def load_model():
tok = T5Tokenizer.from_pretrained(MODEL_DIR)
mdl = T5ForConditionalGeneration.from_pretrained(MODEL_DIR)
dev = "cuda" if torch.cuda.is_available() else "cpu"
try:
mdl = mdl.to(dev)
except RuntimeError:
dev = "cpu"
try:
torch.cuda.empty_cache()
except Exception:
pass
mdl = mdl.to(dev)
mdl.eval()
return tok, mdl, dev
tokenizer, model, device = load_model()
def generate_questions(context: str, n: int = 20) -> list[str]:
ctx = norm(context)
answers = extract_answers(ctx, max_n=80)
questions = []
gen_cfg = dict(
do_sample=True,
top_k=80,
top_p=0.98,
temperature=1.05,
max_new_tokens=72,
no_repeat_ngram_size=3,
repetition_penalty=1.08,
)
num_ret = 8 if n <= 20 else 10
def run_prompt(ans: str, rounds: int):
nonlocal gen_cfg
prompt = f"answer: {ans}\ncontext: {ctx}\nquestion:"
inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=512).to(device)
for _ in range(rounds):
outs = model.generate(**inputs, num_return_sequences=num_ret, **gen_cfg)
added = 0
for o in outs:
q = norm(tokenizer.decode(o, skip_special_tokens=True))
if not q:
continue
if not q.endswith("?"):
q += "?"
if len(q) >= 6 and not is_dup(q, questions, thr=0.85):
questions.append(q)
added += 1
if len(questions) >= n:
return
if added == 0:
gen_cfg["temperature"] = min(1.25, gen_cfg["temperature"] + 0.05)
gen_cfg["top_p"] = min(0.995, gen_cfg["top_p"] + 0.005)
with torch.inference_mode():
for ans in answers:
if len(questions) >= n:
break
run_prompt(ans, rounds=6)
if len(questions) < n:
run_prompt(ctx[:120], rounds=12)
return questions[:n]
if __name__ == "__main__":
ctx = input("\nNhập đoạn văn bản:\n").strip()
try:
n = int((input("\nNhập số lượng câu hỏi cần sinh: ").strip() or ""))
except ValueError:
n = 20
n = max(1, min(n, 200))
qs = generate_questions(ctx, n)
print("\nCác câu hỏi sinh ra:")
for i, q in enumerate(qs, 1):
print(f"{i}. {q}")