Guitaricet/relora

Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates

Jupyter NotebookPythonShellnlpdeep-learningtransformerllamadistributed-trainingpeft
This is stars and forks stats for /Guitaricet/relora repository. As of 03 May, 2024 this repository has 279 stars and 23 forks.

ReLoRA -- PEFT Pretraining Official code for Stack More Layers Differently: High-Rank Training Through Low-Rank Updates https://arxiv.org/abs/2307.05695 Setup All requirements are listed in requirements.txt and kept up-to-date. cd peft_pretraining pip install -r requirements.txt Usage To train a model using ReLoRA, first, perform a warmup through regular training. Train language model with PEFT torchrun --nproc-per-node <N_GPUS> torchrun_main.py \ --model_config configs/llama_250m.json...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
telekom-mms/sectpmctl-kmodsign-wrapperMakefileShell0000
nengyi1226/msfautoMakefileShellC15000
rocky/Perl-Devel-Trepan-ShellPerlShell1000
tejado/pgoapiProtocol BufferPython1.4k04970
haoheliu/AudioLDM2Python1.7k01290
charlesbel/Microsoft-Rewards-FarmerPython70501330
FlagOpen/FlagEmbeddingPython1.2k0630
huggingface/trlPythonShellMakefile5.9k06270
NomaDamas/KICE_slayer_AI_KoreanPython3440200
conan-io/conan-center-indexPythonCMakeC++81601.4k0