lucidrains/flash-cosine-sim-attention

Implementation of fused cosine similarity attention in the same style as Flash Attention

CudaPythonC++Makefiledeep-learningartificial-intelligenceattention-mechanisms
This is stars and forks stats for /lucidrains/flash-cosine-sim-attention repository. As of 06 May, 2024 this repository has 184 stars and 8 forks.

Dive into Deep Learning, redone by Quanta Magazine Flash Cosine Similarity Attention Implementation of fused cosine similarity attention in the same style as Flash Attention. The observation is that by adopting l2 normalized queries and keys, you no longer need to keep track of the row maximums for numerical stability. This greatly simplifies the flash attention algorithm, assuming cosine similarity attention comes at no generalization cost. In other words, stable, fast, memory efficient, and longer...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
gskinner/flutter_animateDartC++CMake6650480
vyos/vyos-documentationPythonDockerfileShell15502850
Activiti/activiti-cloud-common-chartMustacheSmartyMakefile2010
AndyGlx/FullControl-GCode-DesignerVBAPythonShell4610650
ztachip/ztachipVHDLVerilogSystemVerilog1620220
mingyuan-zhang/MotionDiffusePython6740660
VirtualHotBar/HotPEToolBoxBatchfileLuaPython3060230
xinntao/Real-ESRGAN-ncnn-vulkanCC++CMake1k01080
HavocFramework/ModulesCPythonC++1270350
picotorrent/picotorrentC++C#CMake2.5k01810