mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-24 00:45:02 +08:00
[CI/Build][AMD] Skip test_cutlass_w4a8_moe tests on ROCm sine they require cutlass_pack_scale_fp8 (#30508)
Signed-off-by: Randall Smith <ransmith@amd.com> Signed-off-by: Michael Goin <mgoin64@gmail.com> Signed-off-by: mgoin <mgoin64@gmail.com> Co-authored-by: Randall Smith <ransmith@amd.com> Co-authored-by: Michael Goin <mgoin64@gmail.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
This commit is contained in:
parent
0ab23c2b2b
commit
ba80926681
@ -18,7 +18,9 @@ from vllm.model_executor.layers.quantization.utils.quant_utils import (
|
|||||||
from vllm.platforms import current_platform
|
from vllm.platforms import current_platform
|
||||||
from vllm.scalar_type import ScalarType, scalar_types
|
from vllm.scalar_type import ScalarType, scalar_types
|
||||||
|
|
||||||
IS_SUPPORTED_BY_GPU = current_platform.get_device_capability()[0] >= 9
|
IS_SUPPORTED_BY_GPU = (
|
||||||
|
current_platform.is_cuda() and current_platform.get_device_capability()[0] >= 9
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def to_fp8(tensor: torch.Tensor) -> torch.Tensor:
|
def to_fp8(tensor: torch.Tensor) -> torch.Tensor:
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user