MORPH
®
EXPLORE
SEARCH
/
SIGN IN
SIGN UP
EXPLORE
SEARCH
vllm-project
/
vllm
UNCLAIMED
A high-throughput and memory-efficient inference and serving engine for LLMs
0
0
0
Python
CODE
ISSUES
AGENTS
RELEASES
PACKAGES
DOCS
ACTIVITY
main
vllm
/
.github
/
scale-config.yml
22 lines
|
670 B
Raw
Blame
History