mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-13 18:05:49 +08:00
[Doc]: fixing typos in multiple files. (#29685)
Signed-off-by: Didier Durand <durand.didier@gmail.com>
This commit is contained in:
parent
3bcbb30cbf
commit
fae6943068
@ -304,7 +304,7 @@ class SequenceParallelismPass(VllmPatternMatcherPass):
|
|||||||
def __init__(self, config: VllmConfig):
|
def __init__(self, config: VllmConfig):
|
||||||
super().__init__(config)
|
super().__init__(config)
|
||||||
|
|
||||||
# Used to cleanup redundant views created temporarily
|
# Used to clean up redundant views created temporarily
|
||||||
# to circumvent residual shape change issues
|
# to circumvent residual shape change issues
|
||||||
self.noop_cleanup = NoOpEliminationPass(config)
|
self.noop_cleanup = NoOpEliminationPass(config)
|
||||||
self.noop_cleanup.pass_name = f"{self.pass_name}.{self.noop_cleanup.pass_name}"
|
self.noop_cleanup.pass_name = f"{self.pass_name}.{self.noop_cleanup.pass_name}"
|
||||||
|
|||||||
@ -1472,7 +1472,7 @@ class NemotronH_Nano_VL_V2(
|
|||||||
return []
|
return []
|
||||||
|
|
||||||
# # The result multimodal_embeddings is tuple of tensors, with each
|
# # The result multimodal_embeddings is tuple of tensors, with each
|
||||||
# tensor correspoending to a multimodal data item (image or video).
|
# tensor corresponding to a multimodal data item (image or video).
|
||||||
multimodal_embeddings: tuple[torch.Tensor, ...] = ()
|
multimodal_embeddings: tuple[torch.Tensor, ...] = ()
|
||||||
|
|
||||||
# NOTE: It is important to iterate over the keys in this dictionary
|
# NOTE: It is important to iterate over the keys in this dictionary
|
||||||
|
|||||||
@ -19,7 +19,7 @@
|
|||||||
# This code is based off the following work:
|
# This code is based off the following work:
|
||||||
# https://huggingface.co/stabilityai/stablelm-3b-4e1t/blob/main/modeling_stablelm_epoch.py
|
# https://huggingface.co/stabilityai/stablelm-3b-4e1t/blob/main/modeling_stablelm_epoch.py
|
||||||
# https://huggingface.co/stabilityai/stablelm-3b-4e1t/blob/main/config.json
|
# https://huggingface.co/stabilityai/stablelm-3b-4e1t/blob/main/config.json
|
||||||
"""Inference-only StabeLM (https://github.com/Stability-AI/StableLM)
|
"""Inference-only StableLM (https://github.com/Stability-AI/StableLM)
|
||||||
model compatible with HuggingFace weights."""
|
model compatible with HuggingFace weights."""
|
||||||
|
|
||||||
from collections.abc import Iterable
|
from collections.abc import Iterable
|
||||||
|
|||||||
@ -17,7 +17,7 @@ class GCDebugConfig:
|
|||||||
"""
|
"""
|
||||||
Config for GC Debugger.
|
Config for GC Debugger.
|
||||||
- 0: disable GC debugger
|
- 0: disable GC debugger
|
||||||
- 1: enable GC debugger with gc.collect elpased times
|
- 1: enable GC debugger with gc.collect elapsed times
|
||||||
- '{"top_objects":5}': enable GC debugger with top 5 collected objects
|
- '{"top_objects":5}': enable GC debugger with top 5 collected objects
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|||||||
@ -269,7 +269,7 @@ def _find_longest_matched_ngram_and_propose_tokens(
|
|||||||
prev_lps = lps[max_ngram - 1]
|
prev_lps = lps[max_ngram - 1]
|
||||||
i += 1
|
i += 1
|
||||||
elif prev_lps != 0:
|
elif prev_lps != 0:
|
||||||
# Token mismatch: try the second longest prefix
|
# Token mismatch: try the second-longest prefix
|
||||||
# among all suffix of tokens[:i],
|
# among all suffix of tokens[:i],
|
||||||
# which is the longest prefix of tokens[:prev_lps]
|
# which is the longest prefix of tokens[:prev_lps]
|
||||||
prev_lps = lps[prev_lps - 1]
|
prev_lps = lps[prev_lps - 1]
|
||||||
|
|||||||
@ -59,7 +59,7 @@ class ECConnectorModelRunnerMixin:
|
|||||||
)
|
)
|
||||||
|
|
||||||
# This context manager must be used within an active forward context.
|
# This context manager must be used within an active forward context.
|
||||||
# It encapsulates the entire EC conector lifecycle within execute_model
|
# It encapsulates the entire EC connector lifecycle within execute_model
|
||||||
@staticmethod
|
@staticmethod
|
||||||
@contextmanager
|
@contextmanager
|
||||||
def _get_ec_connector_output(
|
def _get_ec_connector_output(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user