sayakpaul HF Staff commited on
Commit
de4cfc8
·
verified ·
1 Parent(s): 945785e

Update qwenimage/qwen_fa3_processor.py

Browse files
Files changed (1) hide show
  1. qwenimage/qwen_fa3_processor.py +1 -0
qwenimage/qwen_fa3_processor.py CHANGED
@@ -17,6 +17,7 @@ def _ensure_fa3_available():
17
  "FlashAttention-3 via Hugging Face `kernels` is required. "
18
  "Tried `get_kernel('kernels-community/vllm-flash-attn3')` and failed with:\n"
19
  f"{_kernels_err}"
 
20
 
21
 
22
  @torch.library.custom_op("flash::flash_attn_func", mutates_args=())
 
17
  "FlashAttention-3 via Hugging Face `kernels` is required. "
18
  "Tried `get_kernel('kernels-community/vllm-flash-attn3')` and failed with:\n"
19
  f"{_kernels_err}"
20
+ )
21
 
22
 
23
  @torch.library.custom_op("flash::flash_attn_func", mutates_args=())