From a5115f4ff567e1311e0a5016519e8aa4ba6e0c98 Mon Sep 17 00:00:00 2001 From: Cyrus Leung Date: Wed, 11 Jun 2025 16:27:22 +0800 Subject: [PATCH] [Doc] Fix quantization link titles (#19478) Signed-off-by: DarkLight1337 --- docs/features/quantization/README.md | 26 +++++++++++++------------- docs/features/quantization/quark.md | 2 +- 2 files changed, 14 insertions(+), 14 deletions(-) diff --git a/docs/features/quantization/README.md b/docs/features/quantization/README.md index 71f62065f63d2..614b43dd00444 100644 --- a/docs/features/quantization/README.md +++ b/docs/features/quantization/README.md @@ -7,16 +7,16 @@ Quantization trades off model precision for smaller memory footprint, allowing l Contents: -- [Supported_Hardware](supported_hardware.md) -- [Auto_Awq](auto_awq.md) -- [Bnb](bnb.md) -- [Bitblas](bitblas.md) -- [Gguf](gguf.md) -- [Gptqmodel](gptqmodel.md) -- [Int4](int4.md) -- [Int8](int8.md) -- [Fp8](fp8.md) -- [Modelopt](modelopt.md) -- [Quark](quark.md) -- [Quantized_Kvcache](quantized_kvcache.md) -- [Torchao](torchao.md) +- [Supported Hardware](supported_hardware.md) +- [AutoAWQ](auto_awq.md) +- [BitsAndBytes](bnb.md) +- [BitBLAS](bitblas.md) +- [GGUF](gguf.md) +- [GPTQModel](gptqmodel.md) +- [INT4 W4A16](int4.md) +- [INT8 W8A8](int8.md) +- [FP8 W8A8](fp8.md) +- [NVIDIA TensorRT Model Optimizer](modelopt.md) +- [AMD Quark](quark.md) +- [Quantized KV Cache](quantized_kvcache.md) +- [TorchAO](torchao.md) diff --git a/docs/features/quantization/quark.md b/docs/features/quantization/quark.md index 51da98cc09d3f..35e9dbe2609be 100644 --- a/docs/features/quantization/quark.md +++ b/docs/features/quantization/quark.md @@ -1,5 +1,5 @@ --- -title: AMD QUARK +title: AMD Quark --- [](){ #quark }