adiom commited on
Commit
1be7a42
·
1 Parent(s): 37284d1
Files changed (2) hide show
  1. app.py +4 -5
  2. requirements.txt +2 -3
app.py CHANGED
@@ -1,10 +1,10 @@
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
3
 
4
- # Загружаем Mistral-7B
5
- model_name = "mistralai/Mistral-7B-Instruct-v0.3"
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
8
 
9
  def chat(message):
10
  inputs = tokenizer(message, return_tensors="pt").to("cpu")
@@ -12,6 +12,5 @@ def chat(message):
12
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
13
  return response
14
 
15
- iface = gr.Interface(fn=chat, inputs="text", outputs="text", title="Mistral-7B Chatbot")
16
  iface.launch()
17
-
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
+ import torch
4
 
5
+ model_name = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
 
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
8
 
9
  def chat(message):
10
  inputs = tokenizer(message, return_tensors="pt").to("cpu")
 
12
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
13
  return response
14
 
15
+ iface = gr.Interface(fn=chat, inputs="text", outputs="text", title="TinyLlama Chatbot")
16
  iface.launch()
 
requirements.txt CHANGED
@@ -1,6 +1,5 @@
 
1
  transformers
2
  gradio
3
- torch
4
- bitsandbytes
5
  sentencepiece
6
- accelerate>=0.26.0
 
1
+ torch
2
  transformers
3
  gradio
4
+ accelerate
 
5
  sentencepiece