Global Rank#5532Apr 14, 2026
Global Rank#5532Apr 14, 2026
since Feb 2023
FMInference/FlexLLMGen
Running large language models on a single GPU for throughput-oriented scenarios.
PythonApache-2.0
9.4k
593
19
Weekly
0stars
·0pushes
Stars
Contributors
New Stars
New Pushes
Forks
GitHub README
<p align="center"> <a href="https://www.star-history.com/fminference/flexllmgen"> <picture> <source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/badge?repo=FMInference/FlexLLMGen&theme=dark" /> <source media="(prefers-color-scheme: light)" srcset="https://api.star-history.com/badge?repo=FMInference/FlexLLMGen" /> <img alt="Star History Rank" src="https://api.star-history.com/badge?repo=FMInference/FlexLLMGen" /> </picture> </a> </p>
Markdown (Light)
[](https://www.star-history.com/fminference/flexllmgen)
Markdown (Dark)
[](https://www.star-history.com/fminference/flexllmgen)
HTML (Light)
<p align="center"> <a href="https://www.star-history.com/fminference/flexllmgen"><img src="https://api.star-history.com/badge?repo=FMInference/FlexLLMGen" alt="Star History Rank" /></a> </p>
HTML (Dark)
<p align="center"> <a href="https://www.star-history.com/fminference/flexllmgen"><img src="https://api.star-history.com/badge?repo=FMInference/FlexLLMGen&theme=dark" alt="Star History Rank" /></a> </p>