mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-03-27 05:59:06 +08:00
[Kernel] Revert the API change of Attention.forward (#12038)
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
This commit is contained in:
parent
bb354e6b2d
commit
1f18adb245
@ -134,8 +134,8 @@ class Attention(nn.Module):
|
||||
query: torch.Tensor,
|
||||
key: torch.Tensor,
|
||||
value: torch.Tensor,
|
||||
_kv_cache: torch.Tensor,
|
||||
_attn_metadata: AttentionMetadata,
|
||||
kv_cache: torch.Tensor,
|
||||
attn_metadata: AttentionMetadata,
|
||||
) -> torch.Tensor:
|
||||
if self.use_output:
|
||||
output = torch.empty_like(query)
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user