This is stars and forks stats for /NVIDIA/Megatron-LM repository. As of 29 Apr, 2024 this repository has 6500 stars and 1432 forks.
Megatron (1, 2, and 3) is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Below are some of the projects where we have directly used Megatron: BERT and GPT Studies Using Megatron BioMegatron:...
Megatron (1, 2, and 3) is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Below are some of the projects where we have directly used Megatron: BERT and GPT Studies Using Megatron BioMegatron:...
repo | techs | stars | weekly | forks | weekly |
---|---|---|---|---|---|
mit-han-lab/bevfusion | PythonC++Cuda | 1.6k | 0 | 294 | 0 |
NobuoTsukamoto/meta-tensorflow-lite | BitBakeC++ | 28 | 0 | 14 | 0 |
Cracked5pider/KaynStrike | COther | 362 | 0 | 60 | 0 |
sumatrapdfreader/sumatrapdf | CC++Assembly | 11k | 0 | 1.6k | 0 |
dekuNukem/bob_cassette_rewinder | CHTMLMakefile | 1.3k | 0 | 38 | 0 |
opuntiaOS-Project/opuntiaOS | CC++Python | 641 | 0 | 30 | 0 |
espeak-ng/espeak-ng | CJavaShell | 2.4k | +23 | 705 | +4 |
dekuNukem/duckyPad | CHTMLMakefile | 1.1k | 0 | 164 | 0 |
donet5/SqlSugar | C#Other | 4.6k | +14 | 1.3k | +2 |
CustomEntity/crNormz | CrystalCPython | 27 | 0 | 0 | 0 |