Package profile
vllm
- Summary: A high-throughput and memory-efficient inference and serving engine for LLMs
- Author: vLLM Team
- Homepage: https://github.com/vllm-project/vllm
- Source: https://github.com/vllm-project/vllm (Repo profile)
- Number of releases: 62
- First release: 0.0.1 on 2023-06-19
- Latest release: 0.9.2 on 2025-07-08
Releases
PyPI Downloads
Dependencies
Vllm has 62 dependencies, 9 of which optional.Dependency | Optional |
---|---|
aiohttp | false |
blake3 | false |
cachetools | false |
cloudpickle | false |
compressed-tensors | false |
depyf | false |
einops | false |
fastapi | false |
filelock | false |
gguf | false |
huggingface-hub | false |
importlib_metadata | false |
lark | false |
llguidance | false |
lm-format-enforcer | false |
mistral_common | false |
msgspec | false |
ninja | false |
numba | false |
numpy | false |
openai | false |
opencv-python-headless | false |
outlines | false |
partial-json-parser | false |
pillow | false |
prometheus_client | false |
prometheus-fastapi-instrumentator | false |
protobuf | false |
psutil | false |
py-cpuinfo | false |
pybase64 | false |
pydantic | false |
python-json-logger | false |
pyyaml | false |
pyzmq | false |
ray | false |
regex | false |
requests | false |
scipy | false |
sentencepiece | false |
setuptools | false |
six | false |
tiktoken | false |
tokenizers | false |
torch | false |
torchaudio | false |
torchvision | false |
tqdm | false |
transformers | false |
typing_extensions | false |
watchfiles | false |
xformers | false |
xgrammar | false |
boto3 | true |
datasets | true |
fastsafetensors | true |
librosa | true |
pandas | true |
runai-model-streamer | true |
runai-model-streamer-s3 | true |
soundfile | true |
tensorizer | true |