PySpect

Home

Invoices

search

top

Package search

vllm

  • Summary: A high-throughput and memory-efficient inference and serving engine for LLMs
  • Author: vLLM Team
  • Homepage: https://github.com/vllm-project/vllm
  • Source: https://github.com/vllm-project/vllm
  • Repo profile
  • Number of releases: 58
  • First release: 0.0.1 on 2023-06-19T08:21:42
  • Latest release: 0.8.5.post1 on 2025-05-02T22:31:02
Dependencies of vllm (62).
DependencyOptional
aiohttpfalse
blake3false
cachetoolsfalse
cloudpicklefalse
compressed-tensorsfalse
depyffalse
einopsfalse
fastapifalse
filelockfalse
gguffalse
huggingface-hubfalse
importlib_metadatafalse
larkfalse
llguidancefalse
lm-format-enforcerfalse
mistral_commonfalse
msgspecfalse
ninjafalse
numbafalse
numpyfalse
openaifalse
opencv-python-headlessfalse
opentelemetry-apifalse
opentelemetry-exporter-otlpfalse
opentelemetry-sdkfalse
opentelemetry-semantic-conventions-aifalse
outlinesfalse
partial-json-parserfalse
pillowfalse
prometheus_clientfalse
prometheus-fastapi-instrumentatorfalse
protobuffalse
psutilfalse
py-cpuinfofalse
pydanticfalse
python-json-loggerfalse
pyyamlfalse
pyzmqfalse
rayfalse
requestsfalse
scipyfalse
sentencepiecefalse
setuptoolsfalse
sixfalse
tiktokenfalse
tokenizersfalse
torchfalse
torchaudiofalse
torchvisionfalse
tqdmfalse
transformersfalse
typing_extensionsfalse
watchfilesfalse
xformersfalse
xgrammarfalse
boto3true
fastsafetensorstrue
librosatrue
runai-model-streamertrue
runai-model-streamer-s3true
soundfiletrue
tensorizertrue