Update app.py
Browse files
app.py
CHANGED
|
@@ -3,13 +3,11 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
|
|
| 3 |
from peft import PeftModel, PeftConfig
|
| 4 |
import gradio as gr
|
| 5 |
|
| 6 |
-
# Replace this with your Hugging Face model repo ID
|
| 7 |
model_repo = "nambn0321/LLM_model"
|
| 8 |
|
| 9 |
-
# Load LoRA adapter config
|
| 10 |
peft_config = PeftConfig.from_pretrained(model_repo)
|
| 11 |
|
| 12 |
-
# Load base model (will automatically download from Hub)
|
| 13 |
bnb_config = BitsAndBytesConfig(
|
| 14 |
load_in_4bit=True,
|
| 15 |
bnb_4bit_quant_type="nf4",
|
|
|
|
| 3 |
from peft import PeftModel, PeftConfig
|
| 4 |
import gradio as gr
|
| 5 |
|
|
|
|
| 6 |
model_repo = "nambn0321/LLM_model"
|
| 7 |
|
| 8 |
+
# Load LoRA adapter config
|
| 9 |
peft_config = PeftConfig.from_pretrained(model_repo)
|
| 10 |
|
|
|
|
| 11 |
bnb_config = BitsAndBytesConfig(
|
| 12 |
load_in_4bit=True,
|
| 13 |
bnb_4bit_quant_type="nf4",
|