Skip to content

Commit bb496ca

Browse files
hebiao064xwu-intel
authored andcommitted
Remove unecessary is_fa3_supported check (sgl-project#6112)
1 parent 2a50198 commit bb496ca

File tree

1 file changed

+0
-4
lines changed

1 file changed

+0
-4
lines changed

sgl-kernel/python/sgl_kernel/flash_attn.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -144,10 +144,6 @@ def flash_attn_with_kvcache(
144144
logsumexp of each row of the matrix QK^T * scaling (e.g., log of the softmax
145145
normalization factor).
146146
"""
147-
if not is_fa3_supported():
148-
raise NotImplementedError(
149-
"flash_attn at sgl-kernel is only supported on sm90 and cu123 above"
150-
)
151147
assert k_cache.stride(-1) == 1, "k_cache must have contiguous last dimension"
152148
assert v_cache.stride(-1) == 1, "v_cache must have contiguous last dimension"
153149
if softmax_scale is None:

0 commit comments

Comments
 (0)