mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2025-12-10 05:25:00 +08:00
[V1][PP] Fix intermediate tensor values (#13417)
Signed-off-by: Cody Yu <hao.yu.cody@gmail.com>
This commit is contained in:
parent
4c21ce9eba
commit
6ac485a953
@ -1137,6 +1137,9 @@ class IntermediateTensors:
|
|||||||
def __setitem__(self, key: str, value: torch.Tensor):
|
def __setitem__(self, key: str, value: torch.Tensor):
|
||||||
self.tensors[key] = value
|
self.tensors[key] = value
|
||||||
|
|
||||||
|
def items(self):
|
||||||
|
return self.tensors.items()
|
||||||
|
|
||||||
def __len__(self):
|
def __len__(self):
|
||||||
return len(self.tensors)
|
return len(self.tensors)
|
||||||
|
|
||||||
|
|||||||
@ -151,7 +151,8 @@ class GPUModelRunner(LoRAModelRunnerMixin):
|
|||||||
self.positions = torch.zeros(self.max_num_tokens,
|
self.positions = torch.zeros(self.max_num_tokens,
|
||||||
dtype=torch.int64,
|
dtype=torch.int64,
|
||||||
device=self.device)
|
device=self.device)
|
||||||
# self.intermediate_tensors # Set after load_model
|
# None in the first PP rank. The rest are set after load_model.
|
||||||
|
self.intermediate_tensors: Optional[IntermediateTensors] = None
|
||||||
|
|
||||||
# Only relevant for models using M-RoPE (e.g, Qwen2-VL)
|
# Only relevant for models using M-RoPE (e.g, Qwen2-VL)
|
||||||
if self.uses_mrope:
|
if self.uses_mrope:
|
||||||
@ -922,6 +923,11 @@ class GPUModelRunner(LoRAModelRunnerMixin):
|
|||||||
if get_pp_group().is_first_rank:
|
if get_pp_group().is_first_rank:
|
||||||
intermediate_tensors = None
|
intermediate_tensors = None
|
||||||
else:
|
else:
|
||||||
|
assert intermediate_tensors is not None
|
||||||
|
assert self.intermediate_tensors is not None
|
||||||
|
for k, v in intermediate_tensors.items():
|
||||||
|
self.intermediate_tensors[k][:num_input_tokens].copy_(
|
||||||
|
v[:num_input_tokens], non_blocking=True)
|
||||||
intermediate_tensors = IntermediateTensors({
|
intermediate_tensors = IntermediateTensors({
|
||||||
k: v[:num_input_tokens]
|
k: v[:num_input_tokens]
|
||||||
for k, v in self.intermediate_tensors.items()
|
for k, v in self.intermediate_tensors.items()
|
||||||
@ -1120,7 +1126,7 @@ class GPUModelRunner(LoRAModelRunnerMixin):
|
|||||||
if get_pp_group().is_first_rank:
|
if get_pp_group().is_first_rank:
|
||||||
intermediate_tensors = None
|
intermediate_tensors = None
|
||||||
else:
|
else:
|
||||||
if not hasattr(self, "intermediate_tensors"):
|
if self.intermediate_tensors is None:
|
||||||
self.intermediate_tensors = (
|
self.intermediate_tensors = (
|
||||||
self.model.make_empty_intermediate_tensors(
|
self.model.make_empty_intermediate_tensors(
|
||||||
batch_size=self.max_num_tokens,
|
batch_size=self.max_num_tokens,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user