Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -42,15 +42,6 @@ from transformers import (
|
|
42 |
from PIL import ImageDraw
|
43 |
from torchvision.transforms.v2 import Resize
|
44 |
|
45 |
-
if IN_SPACES:
|
46 |
-
import subprocess
|
47 |
-
|
48 |
-
subprocess.run(
|
49 |
-
"pip install flash-attn --no-build-isolation",
|
50 |
-
env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
|
51 |
-
shell=True,
|
52 |
-
)
|
53 |
-
|
54 |
os.environ["HF_TOKEN"] = os.environ.get("TOKEN_FROM_SECRET") or True
|
55 |
tokenizer = AutoTokenizer.from_pretrained("vikhyatk/moondream-next", revision=REVISION)
|
56 |
moondream = AutoModelForCausalLM.from_pretrained(
|
@@ -58,7 +49,6 @@ moondream = AutoModelForCausalLM.from_pretrained(
|
|
58 |
trust_remote_code=True,
|
59 |
torch_dtype=torch.float16,
|
60 |
device_map={"": "cuda"},
|
61 |
-
attn_implementation="flash_attention_2",
|
62 |
revision=REVISION
|
63 |
)
|
64 |
|
|
|
42 |
from PIL import ImageDraw
|
43 |
from torchvision.transforms.v2 import Resize
|
44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
os.environ["HF_TOKEN"] = os.environ.get("TOKEN_FROM_SECRET") or True
|
46 |
tokenizer = AutoTokenizer.from_pretrained("vikhyatk/moondream-next", revision=REVISION)
|
47 |
moondream = AutoModelForCausalLM.from_pretrained(
|
|
|
49 |
trust_remote_code=True,
|
50 |
torch_dtype=torch.float16,
|
51 |
device_map={"": "cuda"},
|
|
|
52 |
revision=REVISION
|
53 |
)
|
54 |
|