okuvshynov/slowllama

Finetune llama2-70b and codellama on MacBook Air without quantization

PythonShellllamafine-tuningapple-siliconllama2
This is stars and forks stats for /okuvshynov/slowllama repository. As of 06 May, 2024 this repository has 291 stars and 16 forks.

slowllama Fine-tune Llama2 and CodeLLama models, including 70B/35B on Apple M1/M2 devices (for example, Macbook Air or Mac Mini) or consumer nVidia GPUs. slowllama is not using any quantization. Instead, it offloads parts of model to SSD or main memory on both forward/backward passes. In contrast with training large models from scratch (unattainable) or inference, where we are likely to care about interactivity, we can still get something finetuned if you let it run for a while. Current version is...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
DhanushNehru/Ultimate-Web-Development-ResourcesHTMLJavaScriptCSS306152
Automattic/node-canvasJavaScriptC++C9.6k1.2k
microsoft/msphpsqlPHPC++C1.7k399
chao325/QmaoTaiPythonDockerfileHTML29071
recommenders-team/recommendersPythonOther16.5k2.9k
eric-ai-lab/MiniGPT-5Python38018
AlaaLab/InstructCVPythonShell20617
FoxIO-LLC/ja4RustPython13918
jdx/rtxRustShellJust3.3k101
ponces/treble_build_aospShellMakefile345