Skip to content

Commit 1293f2a

Browse files
Merge branch 'main' into ltx2pipelinespeedup
2 parents bbd6444 + 3577280 commit 1293f2a

2 files changed

Lines changed: 5 additions & 4 deletions

File tree

.github/workflows/pr_style_bot.yml

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,14 @@ on:
55
types: [created]
66

77
permissions:
8-
contents: write
98
pull-requests: write
9+
contents: read
1010

1111
jobs:
1212
style:
13-
uses: huggingface/huggingface_hub/.github/workflows/style-bot-action.yml@e000c1c89c65aee188041723456ac3a479416d4c # main
13+
uses: huggingface/huggingface_hub/.github/workflows/style-bot-action.yml@e2867e92c07d15e1bf18994d0a945ef5ad6b8d65
1414
with:
1515
python_quality_dependencies: "[quality]"
1616
secrets:
17-
bot_token: ${{ secrets.HF_STYLE_BOT_ACTION }}
17+
app_id: ${{ secrets.HF_BOT_STYLE_APP_ID }}
18+
app_private_key: ${{ secrets.HF_BOT_STYLE_SECRET_PEM }}

docs/source/en/optimization/attention_backends.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ The [`~ModelMixin.set_attention_backend`] method iterates through all the module
3535
The example below demonstrates how to enable the `_flash_3_hub` implementation for FlashAttention-3 from the [`kernels`](https://github.com/huggingface/kernels) library, which allows you to instantly use optimized compute kernels from the Hub without requiring any setup.
3636

3737
> [!NOTE]
38-
> FlashAttention-3 is not supported for non-Hopper architectures, in which case, use FlashAttention with `set_attention_backend("flash")`.
38+
> FlashAttention-3 requires Ampere GPUs at a minimum.
3939
4040
```py
4141
import torch

0 commit comments

Comments
 (0)