FlawedLLM commited on
Update app.py
Browse files
app.py
CHANGED
|
@@ -60,12 +60,11 @@ from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
|
| 60 |
# # torch_dtype =torch.float16
|
| 61 |
# model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_9",config=config, ignore_mismatched_sizes=True).to('cuda')
|
| 62 |
# Load model directly
|
| 63 |
-
@spaces.GPU(duration=300)
|
| 64 |
|
| 65 |
tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini_00")
|
| 66 |
model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_00", load_in_4bit=True).to('cuda')
|
| 67 |
|
| 68 |
-
|
| 69 |
def chunk_it(input_command):
|
| 70 |
alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
| 71 |
|
|
|
|
| 60 |
# # torch_dtype =torch.float16
|
| 61 |
# model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_9",config=config, ignore_mismatched_sizes=True).to('cuda')
|
| 62 |
# Load model directly
|
|
|
|
| 63 |
|
| 64 |
tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini_00")
|
| 65 |
model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_00", load_in_4bit=True).to('cuda')
|
| 66 |
|
| 67 |
+
@spaces.GPU(duration=300)
|
| 68 |
def chunk_it(input_command):
|
| 69 |
alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
| 70 |
|