A high-throughput and memory-efficient inference and serving engine for LLMs
No known vulnerabilities in the latest version
[](/packages/vllm-tpu)
<a href="/packages/vllm-tpu"><img src="/api/badges/vllm-tpu?period=month" alt="PyPI Stats"></a>