make compatible with comfy save/load latents
This commit is contained in:
parent
9bb3a79275
commit
f714748ad4
2
nodes.py
2
nodes.py
@ -266,6 +266,8 @@ class MochiDecode:
|
||||
device = mm.get_torch_device()
|
||||
offload_device = mm.unet_offload_device()
|
||||
samples = samples["samples"]
|
||||
samples = samples.to(torch.bfloat16).to(device)
|
||||
|
||||
|
||||
def blend_v(a: torch.Tensor, b: torch.Tensor, blend_extent: int) -> torch.Tensor:
|
||||
blend_extent = min(a.shape[3], b.shape[3], blend_extent)
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user