From 971f5782b40d609a648dc35ae11c80993b5ce427 Mon Sep 17 00:00:00 2001 From: Phil Tillet Date: Mon, 11 Jul 2022 18:56:48 -0700 Subject: [PATCH] [tutorials] Added flash attention credits in tutorial --- python/tutorials/06-fused-attention.py | 1 + 1 file changed, 1 insertion(+) diff --git a/python/tutorials/06-fused-attention.py b/python/tutorials/06-fused-attention.py index 030fe2c2b..89aadb1b4 100644 --- a/python/tutorials/06-fused-attention.py +++ b/python/tutorials/06-fused-attention.py @@ -1,6 +1,7 @@ """ Fused Attention =============== +This is a Triton implementation of the Flash Attention algorithm (Dao et al., https://arxiv.org/pdf/2205.14135v2.pdf) """ import pytest