mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-29 03:44:43 +08:00
[Doc] Installed version of llmcompressor for int8/fp8 quantization (#11103)
Signed-off-by: Guangda Liu <bingps@users.noreply.github.com> Co-authored-by: Guangda Liu <bingps@users.noreply.github.com>
This commit is contained in:
parent
b2f775456e
commit
fd22220687
@ -45,7 +45,7 @@ To produce performant FP8 quantized models with vLLM, you'll need to install the
|
||||
|
||||
.. code-block:: console
|
||||
|
||||
$ pip install llmcompressor==0.1.0
|
||||
$ pip install llmcompressor
|
||||
|
||||
Quantization Process
|
||||
--------------------
|
||||
|
||||
@ -19,7 +19,7 @@ To use INT8 quantization with vLLM, you'll need to install the `llm-compressor <
|
||||
|
||||
.. code-block:: console
|
||||
|
||||
$ pip install llmcompressor==0.1.0
|
||||
$ pip install llmcompressor
|
||||
|
||||
Quantization Process
|
||||
--------------------
|
||||
@ -142,4 +142,4 @@ Best Practices
|
||||
Troubleshooting and Support
|
||||
---------------------------
|
||||
|
||||
If you encounter any issues or have feature requests, please open an issue on the ``vllm-project/llm-compressor`` GitHub repository.
|
||||
If you encounter any issues or have feature requests, please open an issue on the ``vllm-project/llm-compressor`` GitHub repository.
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user