flagos-ai/vllm-FL

Python 2 stars

A high-throughput and memory-efficient inference and serving engine for LLMs

⟳ Syncing… Share on X →
README badge: [![ngmi](https://ngmi.review/badge/flagos-ai/vllm-FL.svg)](https://ngmi.review/repo/flagos-ai/vllm-FL)
0 Merged PRs
Avg Merge Time
Fastest PR
Slowest PR