lucidrains/vit-pytorch

Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

Pythoncomputer-visiontransformersartificial-intelligenceimage-classificationattention-mechanism
This is stars and forks stats for /lucidrains/vit-pytorch repository. As of 24 Apr, 2024 this repository has 15476 stars and 2511 forks.

Table of Contents Vision Transformer - Pytorch Install Usage Parameters Simple ViT NaViT Distillation Deep ViT CaiT Token-to-Token ViT CCT Cross ViT PiT LeViT CvT Twins SVT CrossFormer RegionViT ScalableViT SepViT MaxViT NesT MobileViT Masked Autoencoder Simple Masked Image Modeling Masked Patch Prediction Masked Position Prediction Adaptive Token Sampling Patch Merger Vision Transformer for Small Datasets 3D Vit ViVit Parallel ViT Learnable Memory ViT Dino EsViT Accessing Attention Research Ideas Efficient...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
adap/flowerPythonJupyter NotebookSwift3.1k06300
NVlabs/eg3dPythonCudaC++3k03280
tirthajyoti/Data-science-best-resources2.6k09070
SalesforceFoundation/EDAApexJavaScriptHTML21401460
corkami/picsAssemblyHTMLTeX10k+6724+2
flipperdevices/flipperzero-firmwareCPythonHTML8.5k02k0
InfiniTimeOrg/InfiniTimeCC++Python2.4k+15816+6
openresty/lua-nginx-moduleCLuaPerl10.8k+92k0
OISF/suricataCRustM43.5k+251.3k+9
ExpressLRS/ExpressLRSC++CPython2.7k+10537+1