intel/neural-compressor

Provide unified APIs for SOTA model compression techniques, such as low precision (INT8/INT4/FP4/NF4) quantization, sparsity, pruning, and knowledge distillation on mainstream AI frameworks such as TensorFlow, PyTorch, and ONNX Runtime.

PythonJavaScriptShellTypeScriptJupyter NotebookCSSOthersparsitypruningquantizationknowledge-distillationauto-tuninglow-precisionquantization-aware-trainingpost-training-quantizationlarge-language-modelssmoothquant
This is stars and forks stats for /intel/neural-compressor repository. As of 29 Apr, 2024 this repository has 1421 stars and 191 forks.

Intel® Neural Compressor An open-source Python library supporting popular model compression techniques on all mainstream deep learning frameworks (TensorFlow, PyTorch, ONNX Runtime, and MXNet) Architecture   |   Workflow   |   Results   |   Examples   |   Documentations Intel® Neural Compressor aims to provide popular model compression techniques such as quantization, pruning (sparsity), distillation, and neural architecture search on mainstream frameworks such as TensorFlow, PyTorch, ONNX Runtime,...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
facebookresearch/DiTPythonJupyter Notebook1.6k01160
home-assistant/intentsPythonShellHTML35404150
kakaobrain/karloPythonShell6550390
ThioJoe/Auto-Synced-Translated-DubsPython1.3k01210
mrh0wl/CloudmarePython1.1k+21860
microsoft/LMOpsPythonMDXJupyter Notebook2.4k01510
bzsome/idcard_generatorPythonBatchfile1k02300
mrc-ide/outpack.orderlyRShellDockerfile0000
biobricks-ai/chemharmonyRPythonShell0000
testdouble/standardRubyShell2.4k01980