vllm-project/vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

[view on github]last commit: Apr 15, 2026
stars
76,787
7d
+992
30d
-
90d
-
## star history
## found in