Logo
Explore Help
Sign In
xinyun/vllm
1
0
Fork 0
You've already forked vllm
mirror of https://git.datalinker.icu/vllm-project/vllm.git synced 2025-12-10 05:25:00 +08:00
Code Issues Packages Projects Releases Wiki Activity
vllm/docker
History
Woosuk Kwon 4172235ab7
[V0 deprecation] Deprecate V0 Neuron backend (#21159)
Signed-off-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
2025-09-06 16:15:18 -07:00
..
Dockerfile
Upgrade FlashInfer to v0.3.0 (#24086)
2025-09-04 09:49:20 -07:00
Dockerfile.cpu
[Bugfix] Fix environment variable setting in CPU Dockerfile (#21730)
2025-07-28 11:02:39 +00:00
Dockerfile.nightly_torch
[CI/Build] get rid of unused VLLM_FA_CMAKE_GPU_ARCHES (#21599)
2025-07-31 15:00:08 +08:00
Dockerfile.ppc64le
Fixed ppc build when it runs on non-RHEL based linux distros (#18422)
2025-06-06 11:54:26 -07:00
Dockerfile.rocm
[Deprecation] Remove prompt_token_ids arg fallback in LLM.generate and LLM.embed (#18800)
2025-08-22 10:56:57 +08:00
Dockerfile.rocm_base
[V1] [ROCm] [AITER] Upgrade AITER to commit 916bf3c and bugfix APIs (#20880)
2025-07-13 15:19:32 +00:00
Dockerfile.s390x
[Hardware][IBM Z]Enable v1 for s390x and s390x dockerfile fixes (#22725)
2025-08-19 04:40:37 +00:00
Dockerfile.tpu
Always use cache mounts when installing vllm to avoid populating pip cache in the image. Also remove apt cache. (#23270)
2025-08-21 18:01:03 -04:00
Dockerfile.xpu
[XPU] upgrade torch 2.8 on for XPU (#22300)
2025-08-08 17:03:45 -07:00
Powered by Gitea Version: 1.23.1 Page: 540ms Template: 3ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API