cgbur/llama2.zig

Inference Llama 2 in one file of pure Zig

Zigzigsimdllamaziglangllmllm-inferencellama2
This is stars and forks stats for /cgbur/llama2.zig repository. As of 02 May, 2024 this repository has 68 stars and 3 forks.

llama2.zig This is the Zig version of llama2.c by Andrej Karpathy. It runs inference for the llama2 model architecture recently published by Meta. It currently supports: Inference of llama2 model checkpoints Temperature control Top-p (nucleus) sampling Prompt handling (bpe tokenization) Sequence length control Custom tokenizers Multiquery support Running really fast The ultimate goal is to create a fast, portable, and user-friendly implementation of the llama2 model architecture. The code prioritizes...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
chatchat-space/langchain-ChatGLMPythonShell16.7k+4542.8k+58
gorilla-llm/gorilla-cliPython8960480
axilla-io/axTypeScriptJavaScript9510370
xorbitsai/inferencePythonJavaScriptOther1k0630
ziglibs/eggzonZig17010
QwenLM/Qwen-7BPythonShell5.4k03890
schmee/habuZig27010
jmaczan/gdansk-aiTypeScriptCSSPython1510130
YiVal/YiValPythonCSS2k+8359+34
melih-unsal/DemoGPTPython1.3k+22151+3