qqqcat commited on
Commit
b71ebe4
·
verified ·
1 Parent(s): d50fdea

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -1
app.py CHANGED
@@ -4,7 +4,16 @@ import gradio as gr
4
 
5
  model_name = "RUC-DataLab/DeepAnalyze-8B"
6
 
7
- tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
 
 
 
 
 
 
 
 
 
8
  model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True)
9
 
10
  def chat_fn(message, history):
 
4
 
5
  model_name = "RUC-DataLab/DeepAnalyze-8B"
6
 
7
+ # 先尝试 fast tokenizer,失败就退回到 slow tokenizer(use_fast=False)
8
+ from transformers import AutoTokenizer
9
+
10
+ try:
11
+ tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True, use_fast=True)
12
+ except Exception as e:
13
+ print("Fast tokenizer failed:", e)
14
+ print("Falling back to slow tokenizer (use_fast=False).")
15
+ tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True, use_fast=False)
16
+
17
  model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True)
18
 
19
  def chat_fn(message, history):