Skip to content

Commit 5bd51bd

Browse files
sayakpaulstevhliu
andauthored
Update attention_backends.md to update FA3 minimum support to Ampere (#13283)
* Update attention_backends.md * Update docs/source/en/optimization/attention_backends.md Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
1 parent 7b107d3 commit 5bd51bd

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

docs/source/en/optimization/attention_backends.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ The [`~ModelMixin.set_attention_backend`] method iterates through all the module
3535
The example below demonstrates how to enable the `_flash_3_hub` implementation for FlashAttention-3 from the [`kernels`](https://github.com/huggingface/kernels) library, which allows you to instantly use optimized compute kernels from the Hub without requiring any setup.
3636

3737
> [!NOTE]
38-
> FlashAttention-3 is not supported for non-Hopper architectures, in which case, use FlashAttention with `set_attention_backend("flash")`.
38+
> FlashAttention-3 requires Ampere GPUs at a minimum.
3939
4040
```py
4141
import torch

0 commit comments

Comments
 (0)