mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-05-14 11:34:30 +08:00
Complete fix for duplicate torch operations - all ops use unique library names
Co-authored-by: ProExpertProg <11367180+ProExpertProg@users.noreply.github.com>
This commit is contained in:
parent
9d6f0372e5
commit
47dcf0940f
@ -90,7 +90,7 @@ class Attention(nn.Module):
|
|||||||
x = self.pre_attn(x)
|
x = self.pre_attn(x)
|
||||||
x = self.rms_norm_ref(x)
|
x = self.rms_norm_ref(x)
|
||||||
attn_output = torch.empty_like(x)
|
attn_output = torch.empty_like(x)
|
||||||
torch.ops.silly.attention(x, x, x, attn_output)
|
torch.ops.silly_multiple.attention(x, x, x, attn_output)
|
||||||
x = attn_output
|
x = attn_output
|
||||||
x = self.rms_norm_ref(x)
|
x = self.rms_norm_ref(x)
|
||||||
x = self.post_attn(x)
|
x = self.post_attn(x)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user