vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
[view on github]last commit: Apr 15, 2026
stars
76,787
7d
+992
30d
-
90d
-
## star history
A high-throughput and memory-efficient inference and serving engine for LLMs