Dao-AILab/flash-attention

Fast and memory-efficient exact attention

PythonC++CudaOther
This is stars and forks stats for /Dao-AILab/flash-attention repository. As of 24 Apr, 2024 this repository has 7254 stars and 600 forks.

FlashAttention This repository provides the official implementation of FlashAttention and FlashAttention-2 from the following papers. FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, Christopher RĂ© Paper: https://arxiv.org/abs/2205.14135 IEEE Spectrum article about our submission to the MLPerf 2.0 benchmark using FlashAttention. FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning Tri Dao Paper:...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
databricks-academy/large-language-modelsPython48302270
seanpm2001/AI2001_Category-Source_Code-SC-VHDLRRMarkdownVHDL2010
UCSB-Library-Research-Data-Services/bren-meds213-spring-2023RPythonShell1000
seanpm2001/AI2001_Category-Source_Code-SC-Windows-Registry-EntryRRMarkdownWindows Registry Entries2010
mw46d/robot_armOther0000
yuankunzhang/charmingRustJavaScriptOther1.2k+331+2
code-423n4/2023-07-lensSolidityTypeScriptOther9090
sweatydev/svelte-pocketbaseSvelteJavaScriptOther13000
Sigil-Wen/Dream-with-Vision-ProSwiftPython910130
mlresearch/v195TeXOther0020