| import gradio as gr | |
| import time | |
| import torch | |
| from transformers import BertTokenizer, BertForSequenceClassification | |
| label_dict = {"Urgency": 0, "Not Dark Pattern": 1, "Scarcity": 2, "Misdirection": 3, "Social Proof": 4, "Obstruction": 5, "Sneaking": 6, "Forced Action": 7} | |
| model = BertForSequenceClassification.from_pretrained("bert-base-uncased", num_labels=len(label_dict)) | |
| fine_tuned_model_path = "models/finetuned_BERT_5k_epoch_5.model" | |
| model.load_state_dict(torch.load(fine_tuned_model_path, map_location=torch.device('cpu'))) | |
| tokenizer = BertTokenizer.from_pretrained('bert-base-uncased', do_lower_case=True) | |
| def get_dark_pattern_name(label): | |
| reverse_label_dict = {v: k for k, v in label_dict.items()} | |
| return reverse_label_dict[label] | |
| def find_dark_pattern(text_predict): | |
| encoded_text = tokenizer.encode_plus( | |
| text_predict, | |
| add_special_tokens=True, | |
| return_attention_mask=True, | |
| pad_to_max_length=True, | |
| max_length=256, | |
| return_tensors='pt' | |
| ) | |
| model.eval() | |
| with torch.no_grad(): | |
| inputs = { | |
| 'input_ids': encoded_text['input_ids'], | |
| 'attention_mask': encoded_text['attention_mask'] | |
| } | |
| outputs = model(**inputs) | |
| predictions = outputs.logits | |
| probabilities = torch.nn.functional.softmax(predictions, dim=1) | |
| predicted_label = torch.argmax(probabilities, dim=1).item() | |
| return get_dark_pattern_name(predicted_label) | |
| def predict(text_to_predict): | |
| start_time = time.time() | |
| print("Predicting Dark Pattern...") | |
| for i in range(10): | |
| predicted_darkp = find_dark_pattern(text_to_predict) | |
| time.sleep(0.5) | |
| end_time = time.time() | |
| total_time = end_time - start_time | |
| return predicted_darkp | |
| demo = gr.Interface(fn=predict, inputs="text", outputs="text") | |
| demo.launch(share=True) | |