This website requires JavaScript.
Explore
Help
Sign In
xinyun
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced
2026-01-16 15:54:29 +08:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
/
layers
/
quantization
/
compressed_tensors
History
Tyler Michael Smith
1260180c67
Revert "[Performance] Move apply_w8a8_block_fp8_linear to an op class… (
#25607
)
...
Signed-off-by: Tyler Michael Smith <tlrmchlsmth@gmail.com>
2025-09-25 08:05:21 +00:00
..
schemes
Revert "[Performance] Move apply_w8a8_block_fp8_linear to an op class… (
#25607
)
2025-09-25 08:05:21 +00:00
transform
[Docs] add __init__.py to vllm/model_executor/layers/quantization/compressed_tensors/transform (
#24974
)
2025-09-19 18:32:27 +00:00
__init__.py
[Kernel] Initial Activation Quantization Support (
#4525
)
2024-05-23 21:29:18 +00:00
compressed_tensors_moe.py
[Model] Add LongCat-Flash (
#23991
)
2025-09-24 21:53:40 -07:00
compressed_tensors.py
Revert "[Performance] Move apply_w8a8_block_fp8_linear to an op class… (
#25607
)
2025-09-25 08:05:21 +00:00
triton_scaled_mm.py
[AMD][Kernel][BugFix] fix test_rocm_compressed_tensors_w8a8 for rocm (
#19509
)
2025-06-12 07:14:24 +00:00
utils.py
[Doc]: fix typos in Python comments (
#24093
)
2025-09-02 21:05:45 -07:00