vLLM (high-throughput LLM serving engine)

(github.com)

2 points | by roody_wurlitzer 9 hours ago ago

1 comments