This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-03-23 18:35:53 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
tests
/
compile
History
Richard Zou
b90b0852e9
[easy] Print number of needed GPUs in skip message (
#17594
)
...
Signed-off-by: rzou <zou3519@gmail.com>
2025-05-02 15:27:43 -07:00
..
piecewise
…
__init__.py
…
backend.py
…
conftest.py
…
test_basic_correctness.py
[easy] Print number of needed GPUs in skip message (
#17594
)
2025-05-02 15:27:43 -07:00
test_full_graph.py
Make name of
compressed-tensors
quant method consistent across vLLM (
#17255
)
2025-04-28 16:28:13 +00:00
test_functionalization.py
[torch.compile] Add torch inductor pass for fusing silu_and_mul with subsequent scaled_fp8_quant operations (
#10867
)
2025-05-01 07:59:28 -07:00
test_fusion.py
[Feature] support sequence parallelism using compilation pass (
#16155
)
2025-04-27 06:29:35 -07:00
test_pass_manager.py
[Feature] support sequence parallelism using compilation pass (
#16155
)
2025-04-27 06:29:35 -07:00
test_sequence_parallelism.py
[Feature] support sequence parallelism using compilation pass (
#16155
)
2025-04-27 06:29:35 -07:00
test_silu_mul_quant_fusion.py
[torch.compile] Add torch inductor pass for fusing silu_and_mul with subsequent scaled_fp8_quant operations (
#10867
)
2025-05-01 07:59:28 -07:00
test_wrapper.py
…