Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,6 +3,10 @@ import streamlit as st
|
|
| 3 |
import torch
|
| 4 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 5 |
import json
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
st.set_page_config(page_title="👮 PromptPolice", layout="centered")
|
| 8 |
st.title("👮 PromptPolice: Prompt Evaluator")
|
|
@@ -28,6 +32,8 @@ Respond ONLY in this JSON format:
|
|
| 28 |
}}
|
| 29 |
}}"""
|
| 30 |
|
|
|
|
|
|
|
| 31 |
# --- Load Model and Tokenizer ---
|
| 32 |
@st.cache_resource
|
| 33 |
def load_model():
|
|
|
|
| 3 |
import torch
|
| 4 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 5 |
import json
|
| 6 |
+
from huggingface_hub import login
|
| 7 |
+
import os
|
| 8 |
+
|
| 9 |
+
login(os.environ["HF_PROJECT_TOKEN"])
|
| 10 |
|
| 11 |
st.set_page_config(page_title="👮 PromptPolice", layout="centered")
|
| 12 |
st.title("👮 PromptPolice: Prompt Evaluator")
|
|
|
|
| 32 |
}}
|
| 33 |
}}"""
|
| 34 |
|
| 35 |
+
|
| 36 |
+
|
| 37 |
# --- Load Model and Tokenizer ---
|
| 38 |
@st.cache_resource
|
| 39 |
def load_model():
|