-
Notifications
You must be signed in to change notification settings - Fork 23
The confusion about the usage of MMGP #9
Copy link
Copy link
Open
Description
from mmgp import offload
# model = Inference.load_state_dict(args, model, pretrained_model_path)
offload.load_model_data(model, pretrained_model_path, pinToMemory = pinToMemory, partialPinning = partialPinning)
from mmgp import offload, profile_type
pipe = hunyuan_video_sampler.pipeline
offload.profile(pipe, profile_no= profile_type.HighRAM_LowVRAM_Fast)
If I want to leverage MMGP to reduce the load on CPU and RAM with multiple cards, which approach should I use?
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels