Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -42,15 +42,6 @@ from transformers import (
|
|
| 42 |
from PIL import ImageDraw
|
| 43 |
from torchvision.transforms.v2 import Resize
|
| 44 |
|
| 45 |
-
if IN_SPACES:
|
| 46 |
-
import subprocess
|
| 47 |
-
|
| 48 |
-
subprocess.run(
|
| 49 |
-
"pip install flash-attn --no-build-isolation",
|
| 50 |
-
env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
|
| 51 |
-
shell=True,
|
| 52 |
-
)
|
| 53 |
-
|
| 54 |
os.environ["HF_TOKEN"] = os.environ.get("TOKEN_FROM_SECRET") or True
|
| 55 |
tokenizer = AutoTokenizer.from_pretrained("vikhyatk/moondream-next", revision=REVISION)
|
| 56 |
moondream = AutoModelForCausalLM.from_pretrained(
|
|
@@ -58,7 +49,6 @@ moondream = AutoModelForCausalLM.from_pretrained(
|
|
| 58 |
trust_remote_code=True,
|
| 59 |
torch_dtype=torch.float16,
|
| 60 |
device_map={"": "cuda"},
|
| 61 |
-
attn_implementation="flash_attention_2",
|
| 62 |
revision=REVISION
|
| 63 |
)
|
| 64 |
|
|
|
|
| 42 |
from PIL import ImageDraw
|
| 43 |
from torchvision.transforms.v2 import Resize
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
os.environ["HF_TOKEN"] = os.environ.get("TOKEN_FROM_SECRET") or True
|
| 46 |
tokenizer = AutoTokenizer.from_pretrained("vikhyatk/moondream-next", revision=REVISION)
|
| 47 |
moondream = AutoModelForCausalLM.from_pretrained(
|
|
|
|
| 49 |
trust_remote_code=True,
|
| 50 |
torch_dtype=torch.float16,
|
| 51 |
device_map={"": "cuda"},
|
|
|
|
| 52 |
revision=REVISION
|
| 53 |
)
|
| 54 |
|