Arthur Samuel Galego Panucci FIgueiredo commited on
Commit
da49834
Β·
verified Β·
1 Parent(s): 7b5af8b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -2
app.py CHANGED
@@ -1,19 +1,26 @@
 
1
  import torch
2
  import gradio as gr
3
  from transformers import AutoTokenizer, AutoModelForCausalLM
4
  from peft import PeftModel
5
 
6
  BASE_MODEL = "google/gemma-3-270m-it"
7
- LORA_MODEL = "loboGOAT/DogeAI-v1.0-instruct" # seu LoRA
 
 
8
 
9
  device = "cuda" if torch.cuda.is_available() else "cpu"
10
 
11
  print("πŸ”„ Loading tokenizer...")
12
- tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
 
 
 
13
 
14
  print("πŸ”„ Loading base model...")
15
  model = AutoModelForCausalLM.from_pretrained(
16
  BASE_MODEL,
 
17
  torch_dtype=torch.float16 if device == "cuda" else torch.float32,
18
  device_map="auto"
19
  )
@@ -22,6 +29,7 @@ print("πŸ”„ Applying LoRA...")
22
  model = PeftModel.from_pretrained(model, LORA_MODEL)
23
  model.eval()
24
 
 
25
  def chat(user_input):
26
  prompt = (
27
  "<bos>\n"
 
1
+ import os
2
  import torch
3
  import gradio as gr
4
  from transformers import AutoTokenizer, AutoModelForCausalLM
5
  from peft import PeftModel
6
 
7
  BASE_MODEL = "google/gemma-3-270m-it"
8
+ LORA_MODEL = "loboGOAT/DogeAI-v1.0-instruct"
9
+
10
+ HF_TOKEN = os.environ.get("HF_TOKEN")
11
 
12
  device = "cuda" if torch.cuda.is_available() else "cpu"
13
 
14
  print("πŸ”„ Loading tokenizer...")
15
+ tokenizer = AutoTokenizer.from_pretrained(
16
+ BASE_MODEL,
17
+ token=HF_TOKEN
18
+ )
19
 
20
  print("πŸ”„ Loading base model...")
21
  model = AutoModelForCausalLM.from_pretrained(
22
  BASE_MODEL,
23
+ token=HF_TOKEN,
24
  torch_dtype=torch.float16 if device == "cuda" else torch.float32,
25
  device_map="auto"
26
  )
 
29
  model = PeftModel.from_pretrained(model, LORA_MODEL)
30
  model.eval()
31
 
32
+
33
  def chat(user_input):
34
  prompt = (
35
  "<bos>\n"