PiotrNawrot/nanoT5

Fast & Simple repository for pre-training and fine-tuning T5-style models

Python
This is stars and forks stats for /PiotrNawrot/nanoT5 repository. As of 02 May, 2024 this repository has 822 stars and 49 forks.

nanoT5 (Encoder-Decoder / Pre-training + Fine-Tuning) [Paper] | TLDR | Motivation | Setup | Pre-training | Fine-tuning | Extras | Conclusions | References | Cite | Issues TLDR: This repository comprises the code to reproduce the pre-training of a "Large Language Model" (T5) under a limited budget (1xA100 GPU, < 24 hours) in PyTorch. We start from the randomly initialised T5-base-v1.1 (248M parameters) model, and we pre-train it on the English subset of the C4 dataset and then fine-tune it on Super-Natural...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
rogue-hack-lab/transluxOtherPython2030
zen-mod/ZENSQFPawnPython1040470
sdatkinson/NeuralAmpModelerPluginPythonC++Shell1.5k+51100
PySlurm/pyslurmCythonPythonOther39801080
catppuccin/minecraftPythonGLSL198040
navapbc/platform-test-nextjsHCLShellTypeScript2000
drshahizan/special-topic-data-engineeringPythonJavaScriptCSS810630
Ayanaminn/N46WhisperJupyter NotebookPython1.2k0990
rpinedaec83/pachaBackJavaScriptPythonHTML00260
Beomi/KoAlpacaJupyter NotebookPythonShell1.3k02190