vLLM

一种高吞吐量且内存高效的推理与服务引擎,专为大型语言模型(LLMs)设计。「A high-throughput and memory-efficient inference and serving engine for LLMs」

Main metrics

Overview

Name With Ownervllm-project/vllm
Primary LanguagePython
Program language (Language Count: 8)
Platform
License:Apache License 2.0
Release Count87
Last Release Namev0.11.0rc6 (Posted on )
First Release Namesubmission (Posted on )
Created At2023-02-09 11:23:20
Pushed At2025-10-03 11:29:45
Last Commit At
Stargazers Count59377
Watchers Count440
Fork Count10502
Commits Count10099
Has Issues Enabled
Issues Count11131
Issue Open Count1868
Pull Requests Count10022
Pull Requests Open Count1170
Pull Requests Close Count3195
Has Wiki Enabled
Is Archived
Is Fork
Is Locked
Is Mirror
Is Private
To the top