tosei0000 commited on
Commit
589fde9
·
verified ·
1 Parent(s): 29840c0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -6
app.py CHANGED
@@ -1,23 +1,19 @@
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  import torch
3
 
4
- # 设置模型路径(替换为你上传后显示的路径名)
5
- model_path = "tosei0000/code" # 修改为你上传的文件夹名
6
 
7
  # 加载模型和分词器
8
  tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
9
  model = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True)
10
 
11
- # 如果你有 GPU(Kaggle 支持 GPU),把模型放到 GPU
12
  device = "cuda" if torch.cuda.is_available() else "cpu"
13
  model.to(device)
14
 
15
- # 推理函数
16
  def chat(prompt, max_new_tokens=100):
17
  inputs = tokenizer(prompt, return_tensors="pt").to(device)
18
  outputs = model.generate(**inputs, max_new_tokens=max_new_tokens)
19
  return tokenizer.decode(outputs[0], skip_special_tokens=True)
20
 
21
- # 测试
22
- response = chat("你好,请介绍一下你自己。")
23
  print(response)
 
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  import torch
3
 
4
+ model_path = "tosei0000/code"
 
5
 
6
  # 加载模型和分词器
7
  tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
8
  model = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True)
9
 
 
10
  device = "cuda" if torch.cuda.is_available() else "cpu"
11
  model.to(device)
12
 
 
13
  def chat(prompt, max_new_tokens=100):
14
  inputs = tokenizer(prompt, return_tensors="pt").to(device)
15
  outputs = model.generate(**inputs, max_new_tokens=max_new_tokens)
16
  return tokenizer.decode(outputs[0], skip_special_tokens=True)
17
 
18
+ response = chat("こんにちは!")
 
19
  print(response)