Guitaricet/peft_pretraining

Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates

Jupyter NotebookPythonShellnlpdeep-learningtransformerllamadistributed-trainingpeft
This is stars and forks stats for /Guitaricet/peft_pretraining repository. As of 03 May, 2024 this repository has 279 stars and 23 forks.

ReLoRA -- PEFT Pretraining Official code for Stack More Layers Differently: High-Rank Training Through Low-Rank Updates https://arxiv.org/abs/2307.05695 Setup All requirements are listed in requirements.txt and kept up-to-date. cd peft_pretraining pip install -r requirements.txt Usage To train a model using ReLoRA, first, perform a warmup through regular training. Train language model with PEFT torchrun --nproc-per-node <N_GPUS> torchrun_main.py \ --model_config configs/llama_250m.json...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
daenuprobst/molzipPythonJupyter Notebook49090
Muirey03/RemoteLogObjective-CPython35050
PoomSmart/PSHeaderObjective-CCShell11030
kohenkatz/chamoOCamlShellC0000
ingydotnet/lingyPerlClojureRaku39040
Codium-ai/pr-agentPythonOther2.3k01550
amnemonic/Quansheng_UV-K5_FirmwarePythonCBatchfile2750580
jina-ai/vectordbPythonShellDockerfile3530210
lavishsheth/codeShell1301360
Shaunwei/RealCharJavaScriptPythonSwift5.3k05940