diff --git a/docs/design/custom_op.md b/docs/design/custom_op.md index 0345e42d07e79..41df8fec084cd 100644 --- a/docs/design/custom_op.md +++ b/docs/design/custom_op.md @@ -61,6 +61,7 @@ For example: ```python --8<-- "vllm/attention/layers/mm_encoder_attention.py:mm_encoder_attn" + --8<-- "vllm/model_executor/layers/mla.py:multi_head_latent_attention" ``` @@ -68,15 +69,25 @@ For example: ```python --8<-- "vllm/model_executor/layers/activation.py:silu_and_mul" + --8<-- "vllm/model_executor/layers/activation.py:mul_and_silu" + --8<-- "vllm/model_executor/layers/activation.py:gelu_new" + --8<-- "vllm/model_executor/layers/activation.py:gelu_fast" + --8<-- "vllm/model_executor/layers/activation.py:quick_gelu" + --8<-- "vllm/model_executor/layers/activation.py:gelu_and_mul" + --8<-- "vllm/model_executor/layers/activation.py:gelu_and_mul_sparse" + --8<-- "vllm/model_executor/layers/activation.py:relu2" + --8<-- "vllm/model_executor/layers/activation.py:xielu" + --8<-- "vllm/model_executor/layers/activation.py:swigluoai_and_mul" + --8<-- "vllm/model_executor/layers/activation.py:fatrelu_and_mul" ``` @@ -84,6 +95,7 @@ For example: ```python --8<-- "vllm/model_executor/layers/conv.py:conv2d" + --8<-- "vllm/model_executor/layers/conv.py:conv3d" ``` @@ -91,6 +103,7 @@ For example: ```python --8<-- "vllm/model_executor/layers/vocab_parallel_embedding.py:vocab_parallel_embedding" + --8<-- "vllm/model_executor/layers/vocab_parallel_embedding.py:parallel_lm_head" ``` @@ -98,8 +111,10 @@ For example: ```python --8<-- "vllm/model_executor/layers/linear.py:row_parallel_linear" ---8<-- "vllm/model_executor/layers/linear.py:row_parallel_linear:column_parallel_linear" ---8<-- "vllm/model_executor/layers/linear.py:row_parallel_linear:replicated_linear" + +--8<-- "vllm/model_executor/layers/linear.py:column_parallel_linear" + +--8<-- "vllm/model_executor/layers/linear.py:replicated_linear" ``` **6. Logits Processor:** @@ -112,9 +127,13 @@ For example: ```python --8<-- "vllm/model_executor/layers/mamba/mamba_mixer.py:mamba_mixer" + --8<-- "vllm/model_executor/layers/mamba/mamba_mixer2.py:mamba_mixer2" + --8<-- "vllm/model_executor/layers/mamba/mamba_mixer2.py:mixer2_gated_rms_norm" + --8<-- "vllm/model_executor/models/plamo2.py:plamo2_mamba_mixer" + --8<-- "vllm/model_executor/layers/mamba/short_conv.py:short_conv" ``` @@ -122,9 +141,13 @@ For example: ```python --8<-- "vllm/model_executor/layers/fused_moe/layer.py:fused_moe" + --8<-- "vllm/model_executor/layers/fused_moe/fused_moe_modular_method.py:modular_fused_moe" + --8<-- "vllm/model_executor/layers/fused_moe/unquantized_fused_moe_method.py:unquantized_fused_moe" + --8<-- "vllm/model_executor/models/transformers/moe.py:transformers_fused_moe" + --8<-- "vllm/model_executor/layers/fused_moe/fused_moe.py:grouped_topk" ``` @@ -132,7 +155,9 @@ For example: ```python --8<-- "vllm/model_executor/layers/layernorm.py:rms_norm" + --8<-- "vllm/model_executor/layers/layernorm.py:rms_norm_gated" + --8<-- "vllm/model_executor/layers/layernorm.py:gemma_rms_norm" ``` @@ -146,7 +171,9 @@ For example: ```python --8<-- "vllm/model_executor/layers/rotary_embedding/base.py:rotary_embedding" + --8<-- "vllm/model_executor/layers/rotary_embedding/dual_chunk_rope.py:dual_chunk_rotary_embedding" + --8<-- "vllm/model_executor/layers/rotary_embedding/common.py:apply_rotary_emb" ```