diff --git a/README.md b/README.md index 9021c6193..ffc651da8 100644 --- a/README.md +++ b/README.md @@ -56,10 +56,9 @@ Note that the model can be swapped for a local or remote HF-compatible checkpoin Quantization is applied by selecting an algorithm and calling the `oneshot` API. ```python -from llmcompressor.modifiers.quantization import GPTQModifier from llmcompressor.modifiers.smoothquant import SmoothQuantModifier +from llmcompressor.modifiers.quantization import GPTQModifier from llmcompressor.transformers import oneshot -from transformers import AutoModelForCausalLM # Select quantization algorithm. In this case, we: # * apply SmoothQuant to make the activations easier to quantize