philipturner/metal-flash-attention

Faster alternative to Metal Performance Shaders

SwiftMetalmachine-learningmetalartificial-intelligencegpgpuhigh-performance-computingattention-mechanismtransformer-modelsquantum-chemistry-simulationstable-diffusion
This is stars and forks stats for /philipturner/metal-flash-attention repository. As of 29 Apr, 2024 this repository has 169 stars and 4 forks.

Metal FlashAttention A faster alternative to Metal Performance Shaders, a reference implementation of modern GPU algorithms, and a step toward defragmenting the AI ecosystem. Algorithms: Attention Dense (90.5% ALU) Block-Sparse GEMM FP16 (93.3% ALU) FP32 (87.2% ALU) Fused Biases Usage Progamming Language MFA Supports MPSGraph Supports PyTorch Supports CPU C++ (metal-cpp) ✅ ❌ ✅ GPU C++ (Indirect Command Buffers) ✅ ❌ ❌ Swift (iPadOS, Playgrounds) ✅ ✅ ❌ Swift (macOS, Xcode) ✅ ✅ ✅ Predecessor...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
metal-by-example/metal-spatial-renderingObjective-C++MetalObjective-C122090
e2b-dev/awesome-ai-agents1.8k+63134+9
CarloLucibello/GraphNeuralNetworks.jlJuliaJupyter Notebook1670410
italomandara/CXPatcherASP.NETSwiftRich Text Format522+415-1
WeTransfer/WeScanSwiftOther2.7k05180
mosaicml/streamingPythonOther6120830
canisminor1990/sd-webui-lobe-themeTypeScriptPythonOther1.4k01250
whylabs/langkitJupyter NotebookPythonMakefile4310380
VictorKabata/NotflixKotlinSwiftRuby4450620
olucurious/Awesome-ARKitSwift7.7k09240