vLLM

一种高吞吐量且内存高效的推理与服务引擎,专为大型语言模型(LLMs)设计。「A high-throughput and memory-efficient inference and serving engine for LLMs」

Main metrics

Overview

Name With Ownervllm-project/vllm
Primary LanguagePython
Program language (Language Count: 8)
Platform
License:Apache License 2.0
Release Count75
Last Release Namev0.10.1.1 (Posted on )
First Release Namesubmission (Posted on )
Created At2023-02-09 11:23:20
Pushed At2025-08-22 10:07:24
Last Commit At
Stargazers Count56020
Watchers Count430
Fork Count9572
Commits Count8803
Has Issues Enabled
Issues Count10336
Issue Open Count1804
Pull Requests Count8723
Pull Requests Open Count1011
Pull Requests Close Count2717
Has Wiki Enabled
Is Archived
Is Fork
Is Locked
Is Mirror
Is Private
To the top