vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
⭐ 70,834
Stars
🍴 0
Forks
🐛 0
Issues
🕐 n/a
Last commit
💻 $ pip install vllmA high-throughput and memory-efficient inference and serving engine for LLMs
⭐ 70,834
Stars
🍴 0
Forks
🐛 0
Issues
🕐 n/a
Last commit
💻 $ pip install vllmA high-throughput and memory-efficient inference and serving engine for LLMs
Language
Python
License
Unknown
Category
AI / LLM
Quality Score
9/10
Stars
70.8k
Last Active
Unknown
You might also like