diff --git a/python/tutorials/06-fused-attention.py b/python/tutorials/06-fused-attention.py index 030fe2c2b..89aadb1b4 100644 --- a/python/tutorials/06-fused-attention.py +++ b/python/tutorials/06-fused-attention.py @@ -1,6 +1,7 @@ """ Fused Attention =============== +This is a Triton implementation of the Flash Attention algorithm (Dao et al., https://arxiv.org/pdf/2205.14135v2.pdf) """ import pytest