mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 02:25:01 +08:00
[Doc] Fix quantization link titles (#19478)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
parent
68b4a26149
commit
a5115f4ff5
@ -7,16 +7,16 @@ Quantization trades off model precision for smaller memory footprint, allowing l
|
||||
|
||||
Contents:
|
||||
|
||||
- [Supported_Hardware](supported_hardware.md)
|
||||
- [Auto_Awq](auto_awq.md)
|
||||
- [Bnb](bnb.md)
|
||||
- [Bitblas](bitblas.md)
|
||||
- [Gguf](gguf.md)
|
||||
- [Gptqmodel](gptqmodel.md)
|
||||
- [Int4](int4.md)
|
||||
- [Int8](int8.md)
|
||||
- [Fp8](fp8.md)
|
||||
- [Modelopt](modelopt.md)
|
||||
- [Quark](quark.md)
|
||||
- [Quantized_Kvcache](quantized_kvcache.md)
|
||||
- [Torchao](torchao.md)
|
||||
- [Supported Hardware](supported_hardware.md)
|
||||
- [AutoAWQ](auto_awq.md)
|
||||
- [BitsAndBytes](bnb.md)
|
||||
- [BitBLAS](bitblas.md)
|
||||
- [GGUF](gguf.md)
|
||||
- [GPTQModel](gptqmodel.md)
|
||||
- [INT4 W4A16](int4.md)
|
||||
- [INT8 W8A8](int8.md)
|
||||
- [FP8 W8A8](fp8.md)
|
||||
- [NVIDIA TensorRT Model Optimizer](modelopt.md)
|
||||
- [AMD Quark](quark.md)
|
||||
- [Quantized KV Cache](quantized_kvcache.md)
|
||||
- [TorchAO](torchao.md)
|
||||
|
||||
@ -1,5 +1,5 @@
|
||||
---
|
||||
title: AMD QUARK
|
||||
title: AMD Quark
|
||||
---
|
||||
[](){ #quark }
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user