8000 GitHub - zhajiahe/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

zhajiahe/vllm

 
 

About

A high-throughput and memory-efficient inference and serving engine for LLMs

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 85.8%
  • Cuda 13.1%
  • Other 1.1%
0