This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-20 16:04:32 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
platforms
History
Kunshang Ji
16ded21eeb
[XPU] support Triton Attention backend on Intel GPU (
#24149
)
...
Signed-off-by: Kunshang Ji <kunshang.ji@intel.com>
2025-09-04 20:41:08 +08:00
..
__init__.py
[TPU] Support Pathways in vLLM (
#21417
)
2025-07-30 10:02:12 -07:00
cpu.py
[CPU] Enable data parallel for CPU backend (
#23903
)
2025-08-29 02:19:58 -07:00
cuda.py
[Attention] FlashAttn MLA (
#14258
)
2025-09-04 02:47:59 -07:00
interface.py
[Attention] FlashAttn MLA (
#14258
)
2025-09-04 02:47:59 -07:00
neuron.py
[Refactor]Abstract Platform Interface for Distributed Backend and Add xccl Support for Intel XPU (
#19410
)
2025-07-07 04:32:32 +00:00
rocm.py
[XPU] Add xpu torch.compile support (
#22609
)
2025-08-27 05:33:27 +00:00
tpu.py
[Kernel] Add FP8 support with FlashMLA backend (
#22668
)
2025-08-22 02:26:32 +00:00
xpu.py
[XPU] support Triton Attention backend on Intel GPU (
#24149
)
2025-09-04 20:41:08 +08:00