mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-01-03 08:01:50 +08:00
[Misc] Fix spelling typos in model comments (#31117)
Signed-off-by: c0de128 <kevin.mckay@outlook.com>
This commit is contained in:
parent
42b42824ae
commit
14c3e6ade3
@ -401,7 +401,7 @@ class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig):
|
||||
# of attention tokens that would fit mamba_page_size:
|
||||
# e.g. for mamba page size = 788kB
|
||||
# attn_1_token = 2kB -> fits ~394 tokens
|
||||
# then round up to a mulitple of 256 -> 512 tokens
|
||||
# then round up to a multiple of 256 -> 512 tokens
|
||||
# End result:
|
||||
# attn_block_size = 512
|
||||
# mamba_block_size = 512 (aligned to a multiple of chunk_size)
|
||||
|
||||
@ -323,7 +323,7 @@ class Qwen3Omni_VisionTransformer(nn.Module):
|
||||
hidden_size=self.hidden_size,
|
||||
)
|
||||
|
||||
# vit pos embeding, TODO: spatial_patch_size vs patch_size
|
||||
# vit pos embedding, TODO: spatial_patch_size vs patch_size
|
||||
if self.apply_vit_abs_pos_embed:
|
||||
self.pos_embed = nn.Embedding(self.num_grid_per_side**2, self.hidden_size)
|
||||
else:
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user