From 6a25ea5f0ea193e35b5a83cb0285c48964bc9eb1 Mon Sep 17 00:00:00 2001 From: Uranus <109661872+UranusSeven@users.noreply.github.com> Date: Wed, 19 Nov 2025 13:30:08 +0800 Subject: [PATCH] [Docs] Update oneshot imports (#28188) Signed-off-by: UranusSeven <109661872+UranusSeven@users.noreply.github.com> --- docs/features/quantization/fp8.md | 2 +- docs/features/quantization/int4.md | 2 +- docs/features/quantization/int8.md | 2 +- docs/features/quantization/quantized_kvcache.md | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/features/quantization/fp8.md b/docs/features/quantization/fp8.md index 0c5111fb8af0d..d4a6176b236f1 100644 --- a/docs/features/quantization/fp8.md +++ b/docs/features/quantization/fp8.md @@ -60,7 +60,7 @@ Since simple RTN does not require data for weight quantization and the activatio ??? code ```python - from llmcompressor.transformers import oneshot + from llmcompressor import oneshot from llmcompressor.modifiers.quantization import QuantizationModifier # Configure the simple PTQ quantization diff --git a/docs/features/quantization/int4.md b/docs/features/quantization/int4.md index 035e7ea291f9e..9752039097d63 100644 --- a/docs/features/quantization/int4.md +++ b/docs/features/quantization/int4.md @@ -80,7 +80,7 @@ Now, apply the quantization algorithms: ??? code ```python - from llmcompressor.transformers import oneshot + from llmcompressor import oneshot from llmcompressor.modifiers.quantization import GPTQModifier from llmcompressor.modifiers.smoothquant import SmoothQuantModifier diff --git a/docs/features/quantization/int8.md b/docs/features/quantization/int8.md index ec8a77f74ffef..701ca6378cb16 100644 --- a/docs/features/quantization/int8.md +++ b/docs/features/quantization/int8.md @@ -87,7 +87,7 @@ Now, apply the quantization algorithms: ??? code ```python - from llmcompressor.transformers import oneshot + from llmcompressor import oneshot from llmcompressor.modifiers.quantization import GPTQModifier from llmcompressor.modifiers.smoothquant import SmoothQuantModifier diff --git a/docs/features/quantization/quantized_kvcache.md b/docs/features/quantization/quantized_kvcache.md index 56cf057678be6..d26a5e217f314 100644 --- a/docs/features/quantization/quantized_kvcache.md +++ b/docs/features/quantization/quantized_kvcache.md @@ -78,7 +78,7 @@ Here's a complete example using `meta-llama/Llama-3.1-8B-Instruct` (most models ```python from datasets import load_dataset from transformers import AutoModelForCausalLM, AutoTokenizer - from llmcompressor.transformers import oneshot + from llmcompressor import oneshot # Select model and load it MODEL_ID = "meta-llama/Llama-3.1-8B-Instruct"