diff --git a/docs/models/supported_models.md b/docs/models/supported_models.md index d726b5350f807..9039c0480547e 100644 --- a/docs/models/supported_models.md +++ b/docs/models/supported_models.md @@ -736,37 +736,6 @@ Some models are supported only via the [Transformers backend](#transformers). Th !!! note To use `TIGER-Lab/Mantis-8B-siglip-llama3`, you have to pass `--hf_overrides '{"architectures": ["MantisForConditionalGeneration"]}'` when running vLLM. -!!! warning - The output quality of `AllenAI/Molmo-7B-D-0924` (especially in object localization tasks) has deteriorated in recent updates. - - For the best results, we recommend using the following dependency versions (tested on A10 and L40): - - ??? code "Dependency versions" - - ```text - # Core vLLM-compatible dependencies with Molmo accuracy setup (tested on L40) - torch==2.5.1 - torchvision==0.20.1 - transformers==4.48.1 - tokenizers==0.21.0 - tiktoken==0.7.0 - vllm==0.7.0 - - # Optional but recommended for improved performance and stability - triton==3.1.0 - xformers==0.0.28.post3 - uvloop==0.21.0 - protobuf==5.29.3 - openai==1.60.2 - opencv-python-headless==4.11.0.86 - pillow==10.4.0 - - # Installed FlashAttention (for float16 only) - flash-attn>=2.5.6 # Not used in float32, but should be documented - ``` - - **Note:** Make sure you understand the security implications of using outdated packages. - !!! note The official `openbmb/MiniCPM-V-2` doesn't work yet, so we need to use a fork (`HwwwH/MiniCPM-V-2`) for now. For more details, please see: