Logo
Explore Help
Sign In
xinyun/vllm
1
0
Fork 0
You've already forked vllm
mirror of https://git.datalinker.icu/vllm-project/vllm.git synced 2026-03-23 18:35:53 +08:00
Code Issues Packages Projects Releases Wiki Activity
vllm/tests/compile
History
Richard Zou b90b0852e9
[easy] Print number of needed GPUs in skip message (#17594)
Signed-off-by: rzou <zou3519@gmail.com>
2025-05-02 15:27:43 -07:00
..
piecewise
…
__init__.py
…
backend.py
…
conftest.py
…
test_basic_correctness.py
[easy] Print number of needed GPUs in skip message (#17594)
2025-05-02 15:27:43 -07:00
test_full_graph.py
Make name of compressed-tensors quant method consistent across vLLM (#17255)
2025-04-28 16:28:13 +00:00
test_functionalization.py
[torch.compile] Add torch inductor pass for fusing silu_and_mul with subsequent scaled_fp8_quant operations (#10867)
2025-05-01 07:59:28 -07:00
test_fusion.py
[Feature] support sequence parallelism using compilation pass (#16155)
2025-04-27 06:29:35 -07:00
test_pass_manager.py
[Feature] support sequence parallelism using compilation pass (#16155)
2025-04-27 06:29:35 -07:00
test_sequence_parallelism.py
[Feature] support sequence parallelism using compilation pass (#16155)
2025-04-27 06:29:35 -07:00
test_silu_mul_quant_fusion.py
[torch.compile] Add torch inductor pass for fusing silu_and_mul with subsequent scaled_fp8_quant operations (#10867)
2025-05-01 07:59:28 -07:00
test_wrapper.py
…
Powered by Gitea Version: 1.23.1 Page: 5251ms Template: 13ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API