Update app.py
Browse files
app.py
CHANGED
@@ -11,11 +11,11 @@ import os
|
|
11 |
from moviepy.editor import VideoFileClip
|
12 |
|
13 |
# Install flash-attn
|
14 |
-
|
15 |
|
16 |
# Model and Processor Loading (Done once at startup)
|
17 |
MODEL_ID = "Qwen/Qwen2-VL-7B-Instruct"
|
18 |
-
model = Qwen2VLForConditionalGeneration.from_pretrained(MODEL_ID, trust_remote_code=True, torch_dtype=torch.
|
19 |
processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
|
20 |
|
21 |
DESCRIPTION = "[Qwen2-VL-7B Demo](https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct)"
|
|
|
11 |
from moviepy.editor import VideoFileClip
|
12 |
|
13 |
# Install flash-attn
|
14 |
+
subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
15 |
|
16 |
# Model and Processor Loading (Done once at startup)
|
17 |
MODEL_ID = "Qwen/Qwen2-VL-7B-Instruct"
|
18 |
+
model = Qwen2VLForConditionalGeneration.from_pretrained(MODEL_ID, trust_remote_code=True, torch_dtype=torch.float16).to("cuda").eval()
|
19 |
processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
|
20 |
|
21 |
DESCRIPTION = "[Qwen2-VL-7B Demo](https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct)"
|