From e7a4e4018115a004d2f4ba164ba9aaebb965c510 Mon Sep 17 00:00:00 2001 From: sayakpaul Date: Fri, 10 Apr 2026 20:31:56 +0530 Subject: [PATCH] fix fa4 integration --- src/diffusers/models/attention_dispatch.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/diffusers/models/attention_dispatch.py b/src/diffusers/models/attention_dispatch.py index 9bb3a6fbd0ce..837d573d8c4d 100644 --- a/src/diffusers/models/attention_dispatch.py +++ b/src/diffusers/models/attention_dispatch.py @@ -540,7 +540,7 @@ def _check_attention_backend_requirements(backend: AttentionBackendName) -> None f"Backend '{backend.value}' needs to be used with a `kernels` version of at least 0.12. Please update with `pip install -U kernels`." ) - if backend == AttentionBackendName.FLASH_4_HUB and not is_kernels_available(">=", "0.12.3"): + if backend == AttentionBackendName.FLASH_4_HUB and not is_kernels_version(">=", "0.12.3"): raise RuntimeError( f"Backend '{backend.value}' needs to be used with a `kernels` version of at least 0.12.3. Please update with `pip install -U kernels`." )