ybelkada commited on
Commit
1373e5d
Β·
verified Β·
1 Parent(s): f8d7a7f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -5,7 +5,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
5
  import gradio as gr
6
  from threading import Thread
7
 
8
- MODEL = "tiiuae/Falcon-EB-3B-Instruct"
9
 
10
  TITLE = "<h1><center>Falcon-E-3B-Instruct playground</center></h1>"
11
  SUB_TITLE = """<center>This interface has been created for quick validation purposes, do not use it for production.</center>"""
@@ -29,11 +29,10 @@ END_MESSAGE = """
29
 
30
  device = "cuda" # for GPU usage or "cpu" for CPU usage
31
 
32
- tokenizer = AutoTokenizer.from_pretrained(MODEL, token=os.getenv("HF_TOKEN"))
33
  model = AutoModelForCausalLM.from_pretrained(
34
  MODEL,
35
- torch_dtype=torch.bfloat16,
36
- token=os.getenv("HF_TOKEN")
37
  ).to(device)
38
 
39
  model = torch.compile(model)
 
5
  import gradio as gr
6
  from threading import Thread
7
 
8
+ MODEL = "tiiuae/Falcon-E-3B-Instruct"
9
 
10
  TITLE = "<h1><center>Falcon-E-3B-Instruct playground</center></h1>"
11
  SUB_TITLE = """<center>This interface has been created for quick validation purposes, do not use it for production.</center>"""
 
29
 
30
  device = "cuda" # for GPU usage or "cpu" for CPU usage
31
 
32
+ tokenizer = AutoTokenizer.from_pretrained(MODEL)
33
  model = AutoModelForCausalLM.from_pretrained(
34
  MODEL,
35
+ torch_dtype=torch.bfloat16
 
36
  ).to(device)
37
 
38
  model = torch.compile(model)