This is stars and forks stats for /mit-han-lab/streaming-llm repository. As of 06 Dec, 2023 this repository has 4561 stars and 234 forks.
Efficient Streaming Language Models with Attention Sinks [paper] streamingllm_demo.mp4 TL;DR We deploy LLMs for infinite-length inputs without sacrificing efficiency and performance. Abstract Deploying Large Language Models (LLMs) in streaming applications such as multi-round dialogue, where long interactions are expected, is urgently needed but poses two major challenges. Firstly, during the decoding stage, caching previous tokens' Key and Value states (KV) consumes...
Efficient Streaming Language Models with Attention Sinks [paper] streamingllm_demo.mp4 TL;DR We deploy LLMs for infinite-length inputs without sacrificing efficiency and performance. Abstract Deploying Large Language Models (LLMs) in streaming applications such as multi-round dialogue, where long interactions are expected, is urgently needed but poses two major challenges. Firstly, during the decoding stage, caching previous tokens' Key and Value states (KV) consumes...
repo | techs | stars | weekly | forks | weekly |
---|---|---|---|---|---|
danielgross/localpilot | Python | 1.2k | 0 | 60 | 0 |
Mayandev/where-is-douban250 | PythonJavaScript | 509 | 0 | 40 | 0 |
ray-project/ray-llm | PythonOther | 736 | +38 | 47 | +2 |
win3zz/CVE-2023-43261 | Python | 44 | 0 | 7 | 0 |
ONLYOFFICE/DocSpace-buildtools | Rich Text FormatShellPython | 0 | 0 | 1 | 0 |
DHEERAJHARODE/Hacktoberfest2023-Open-source- | Jupyter NotebookC++C | 340 | 0 | 2.3k | 0 |
Source2ZE/CS2Fixes | C++PythonC | 53 | 0 | 33 | 0 |
NVIDIA/MatX | C++CudaCMake | 989 | 0 | 57 | 0 |
pct/TiApp | CoffeeScriptRubyPython | 9 | 0 | 0 | 0 |
naveen3011/WebD_project_hacktober2023 | CSSJavaScriptHTML | 25 | 0 | 105 | 0 |