This is stars and forks stats for /intel-analytics/BigDL repository. As of 24 Apr, 2024 this repository has 4426 stars and 1145 forks.
BigDL-LLM bigdl-llm is a library for running LLM (large language model) on Intel XPU (from Laptop to GPU to Cloud) using INT4 with very low latency1 (for any PyTorch model). It is built on top of the excellent work of llama.cpp, ggml, gptq, bitsandbytes, qlora, llama-cpp-python, gptq_for_llama, chatglm.cpp, redpajama.cpp, gptneox.cpp, bloomz.cpp, etc. Latest update [New] bigdl-llm now supports QLoRA fintuning on Intel GPU; see the the example here. bigdl-llm now supports Intel GPU (including Arc,...
BigDL-LLM bigdl-llm is a library for running LLM (large language model) on Intel XPU (from Laptop to GPU to Cloud) using INT4 with very low latency1 (for any PyTorch model). It is built on top of the excellent work of llama.cpp, ggml, gptq, bitsandbytes, qlora, llama-cpp-python, gptq_for_llama, chatglm.cpp, redpajama.cpp, gptneox.cpp, bloomz.cpp, etc. Latest update [New] bigdl-llm now supports QLoRA fintuning on Intel GPU; see the the example here. bigdl-llm now supports Intel GPU (including Arc,...
repo | techs | stars | weekly | forks | weekly |
---|---|---|---|---|---|
mahmoud/awesome-python-applications | Jupyter Notebook | 14.9k | 0 | 2.6k | 0 |
mrousavy/react-native-blurhash | KotlinSwiftJava | 1.7k | 0 | 57 | 0 |
schism-dev/schism | LLVMFortranC | 68 | 0 | 74 | 0 |
pakeke-constructor/PushOps | LuaClojureGLSL | 15 | 0 | 0 | 0 |
aws/eks-anywhere-build-tooling | MakefileShellGo | 41 | 0 | 79 | 0 |
GlitchyTurtle/Avatar-Addon | JavaScript | 25 | 0 | 28 | 0 |
comby-tools/comby | OCamlShellStandard ML | 2.2k | +2 | 57 | 0 |
frones/ACBr | PascalJavaC# | 129 | 0 | 151 | 0 |
mrash/fwknop | PerlCRoff | 966 | 0 | 214 | 0 |
Plutonomicon/cardano-transaction-lib | PureScriptJavaScriptNix | 78 | 0 | 49 | 0 |