This is stars and forks stats for /lucidrains/flash-cosine-sim-attention repository. As of 06 May, 2024 this repository has 184 stars and 8 forks.
Dive into Deep Learning, redone by Quanta Magazine Flash Cosine Similarity Attention Implementation of fused cosine similarity attention in the same style as Flash Attention. The observation is that by adopting l2 normalized queries and keys, you no longer need to keep track of the row maximums for numerical stability. This greatly simplifies the flash attention algorithm, assuming cosine similarity attention comes at no generalization cost. In other words, stable, fast, memory efficient, and longer...
Dive into Deep Learning, redone by Quanta Magazine Flash Cosine Similarity Attention Implementation of fused cosine similarity attention in the same style as Flash Attention. The observation is that by adopting l2 normalized queries and keys, you no longer need to keep track of the row maximums for numerical stability. This greatly simplifies the flash attention algorithm, assuming cosine similarity attention comes at no generalization cost. In other words, stable, fast, memory efficient, and longer...
repo | techs | stars | weekly | forks | weekly |
---|---|---|---|---|---|
gskinner/flutter_animate | DartC++CMake | 665 | 0 | 48 | 0 |
vyos/vyos-documentation | PythonDockerfileShell | 155 | 0 | 285 | 0 |
Activiti/activiti-cloud-common-chart | MustacheSmartyMakefile | 2 | 0 | 1 | 0 |
AndyGlx/FullControl-GCode-Designer | VBAPythonShell | 461 | 0 | 65 | 0 |
ztachip/ztachip | VHDLVerilogSystemVerilog | 162 | 0 | 22 | 0 |
mingyuan-zhang/MotionDiffuse | Python | 674 | 0 | 66 | 0 |
VirtualHotBar/HotPEToolBox | BatchfileLuaPython | 306 | 0 | 23 | 0 |
xinntao/Real-ESRGAN-ncnn-vulkan | CC++CMake | 1k | 0 | 108 | 0 |
HavocFramework/Modules | CPythonC++ | 127 | 0 | 35 | 0 |
picotorrent/picotorrent | C++C#CMake | 2.5k | 0 | 181 | 0 |