Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,6 +3,7 @@ import gradio as gr
|
|
| 3 |
import transformers
|
| 4 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 5 |
import torch
|
|
|
|
| 6 |
|
| 7 |
title = """
|
| 8 |
# Welcome to 🌟Tonic's🫡Command-R
|
|
@@ -24,6 +25,7 @@ model = AutoModelForCausalLM.from_pretrained(model_id, quantization_config=bnb_c
|
|
| 24 |
|
| 25 |
@spaces.GPU
|
| 26 |
def generate_response(user_input, max_new_tokens, temperature):
|
|
|
|
| 27 |
messages = [{"role": "user", "content": user_input}]
|
| 28 |
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
| 29 |
input_ids = input_ids.to(model.device)
|
|
|
|
| 3 |
import transformers
|
| 4 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 5 |
import torch
|
| 6 |
+
import os
|
| 7 |
|
| 8 |
title = """
|
| 9 |
# Welcome to 🌟Tonic's🫡Command-R
|
|
|
|
| 25 |
|
| 26 |
@spaces.GPU
|
| 27 |
def generate_response(user_input, max_new_tokens, temperature):
|
| 28 |
+
os.system('nvidia-smi')
|
| 29 |
messages = [{"role": "user", "content": user_input}]
|
| 30 |
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
| 31 |
input_ids = input_ids.to(model.device)
|