CMD + K

vllm

Community

A high-throughput and memory-efficient inference and serving engine for LLMs

Installation

To install this package, run one of the following:

Conda
$conda install conda-forge::vllm

Usage Tracking

0.10.0
0.9.2
0.8.3
3 / 8 versions selected
Downloads (Last 6 months): 0

Description

Easy, fast, and cheap LLM serving for everyone

About

Summary

A high-throughput and memory-efficient inference and serving engine for LLMs

Last Updated

Sep 26, 2025 at 08:29

License

Apache-2.0 AND BSD-3-Clause

Total Downloads

13.7K

Supported Platforms

linux-64
macOS-arm64

Unsupported Platforms

linux-aarch64 Last supported version: 0.9.2