arjunanand13 commited on
Commit
1a36d4e
·
verified ·
1 Parent(s): 7777eff

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -4,7 +4,7 @@ import subprocess
4
  import torch
5
  from peft import LoraConfig
6
  from huggingface_hub import InferenceApi
7
- # from transformers import BitsAndBytesConfig
8
 
9
  subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
10
  DEVICE = "cuda:0"
 
4
  import torch
5
  from peft import LoraConfig
6
  from huggingface_hub import InferenceApi
7
+ from transformers import BitsAndBytesConfig
8
 
9
  subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
10
  DEVICE = "cuda:0"