ericjedha commited on
Commit
a1cae3c
·
verified ·
1 Parent(s): 65a0071

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -3
app.py CHANGED
@@ -5,14 +5,12 @@ import re
5
  import time
6
  import torch
7
  import spaces
8
- import subprocess
9
- subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
10
 
11
  from io import BytesIO
12
 
13
  processor = AutoProcessor.from_pretrained("HuggingFaceTB/SmolVLM2-2.2B-Instruct")
14
  model = AutoModelForImageTextToText.from_pretrained("HuggingFaceTB/SmolVLM2-2.2B-Instruct",
15
- _attn_implementation="flash_attention_2",
16
  torch_dtype=torch.bfloat16).to("cuda:0")
17
 
18
 
 
5
  import time
6
  import torch
7
  import spaces
8
+
 
9
 
10
  from io import BytesIO
11
 
12
  processor = AutoProcessor.from_pretrained("HuggingFaceTB/SmolVLM2-2.2B-Instruct")
13
  model = AutoModelForImageTextToText.from_pretrained("HuggingFaceTB/SmolVLM2-2.2B-Instruct",
 
14
  torch_dtype=torch.bfloat16).to("cuda:0")
15
 
16