vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
A high-throughput and memory-efficient inference and serving engine for LLMs
To install this package, run one of the following:
Easy, fast, and cheap LLM serving for everyone
Summary
A high-throughput and memory-efficient inference and serving engine for LLMs
Last Updated
Jan 10, 2026 at 12:13
License
Apache-2.0 AND BSD-3-Clause
Total Downloads
14.1K
Supported Platforms
Documentation
https://vllm.readthedocs.io/en/latest/