vLLM
一种高吞吐量且内存高效的推理与服务引擎,专为大型语言模型(LLMs)设计。「A high-throughput and memory-efficient inference and serving engine for LLMs」
Main metrics
Overview
| Name With Owner | vllm-project/vllm |
|---|---|
| Primary Language | Python |
| Program language | (Language Count: 8) |
| Platform | |
| License: | Apache License 2.0 |
| Release Count | 97 |
| Last Release Name | v0.11.2 (Posted on ) |
| First Release Name | submission (Posted on ) |
| Created At | 2023-02-09 19:23:20 |
| Pushed At | 2025-11-21 17:41:20 |
| Last Commit At | |
| Stargazers Count | 63616 |
| Watchers Count | 459 |
| Fork Count | 11449 |
| Commits Count | 11516 |
| Has Issues Enabled | |
| Issues Count | 12059 |
| Issue Open Count | 1909 |
| Pull Requests Count | 11440 |
| Pull Requests Open Count | 1249 |
| Pull Requests Close Count | 3779 |
| Has Wiki Enabled | |
| Is Archived | |
| Is Fork | |
| Is Locked | |
| Is Mirror | |
| Is Private |
