vLLM

一种高吞吐量且内存高效的推理与服务引擎,专为大型语言模型(LLMs)设计。「A high-throughput and memory-efficient inference and serving engine for LLMs」

Main metrics

Overview

Name With Ownervllm-project/vllm
Primary LanguagePython
Program language (Language Count: 8)
Platform
License:Apache License 2.0
Release Count97
Last Release Namev0.11.2 (Posted on )
First Release Namesubmission (Posted on )
Created At2023-02-09 19:23:20
Pushed At2025-11-21 17:41:20
Last Commit At
Stargazers Count63616
Watchers Count459
Fork Count11449
Commits Count11516
Has Issues Enabled
Issues Count12059
Issue Open Count1909
Pull Requests Count11440
Pull Requests Open Count1249
Pull Requests Close Count3779
Has Wiki Enabled
Is Archived
Is Fork
Is Locked
Is Mirror
Is Private
To the top