akshit-g commited on
Commit
b0dd6ae
·
verified ·
1 Parent(s): a77f3d8

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -8,16 +8,16 @@ from PIL import ImageDraw
8
  from torchvision.transforms.v2 import Resize
9
 
10
  import subprocess
11
- # subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
12
 
13
  model_id = "vikhyatk/moondream2"
14
  revision = "2024-05-20"
15
  tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
16
  moondream = AutoModelForCausalLM.from_pretrained(
17
  model_id, trust_remote_code=True, revision=revision,
18
- # torch_dtype=torch.bfloat16, device_map={"": "cuda"}
19
- torch_dtype=torch.float32, device_map="cpu"
20
- # attn_implementation="flash_attention_2"
21
  )
22
  moondream.eval()
23
 
 
8
  from torchvision.transforms.v2 import Resize
9
 
10
  import subprocess
11
+ subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
12
 
13
  model_id = "vikhyatk/moondream2"
14
  revision = "2024-05-20"
15
  tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
16
  moondream = AutoModelForCausalLM.from_pretrained(
17
  model_id, trust_remote_code=True, revision=revision,
18
+ torch_dtype=torch.bfloat16, device_map={"": "cuda"},
19
+ # torch_dtype=torch.float32, device_map="cpu"
20
+ attn_implementation="flash_attention_2"
21
  )
22
  moondream.eval()
23