from transformers import AutoModelForCausalLM, AutoTokenizer import torch import gradio as gr # Set up the model and tokenizer device = "cuda" if torch.cuda.is_available() else "cpu" model = AutoModelForCausalLM.from_pretrained( "betterdataai/PII_DETECTION_MODEL", trust_remote_code=True ).to(device) tokenizer = AutoTokenizer.from_pretrained( "betterdataai/PII_DETECTION_MODEL", trust_remote_code=True ) classes_list = ['','','','','','','','','','','','','','','','','','','','','','','','','','','','