Skip to content

Add cuDNN flash attention sequence packing #758

Add cuDNN flash attention sequence packing

Add cuDNN flash attention sequence packing #758

Bazel single accelerator CUDA tests (linux-x86-n2-16, Python 3.12, x64=0)

succeeded Feb 3, 2025 in 11m 18s