Skip to content
This repository has been archived by the owner on Dec 14, 2023. It is now read-only.

First GPU occupies more VRAM in distributed training #66

Open
suzhenghang opened this issue May 22, 2023 · 0 comments
Open

First GPU occupies more VRAM in distributed training #66

suzhenghang opened this issue May 22, 2023 · 0 comments
Labels
bug Something isn't working

Comments

@suzhenghang
Copy link

suzhenghang commented May 22, 2023

link
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
cached_latent = torch.load(self.cached_data_list[index], map_location=device)
Otherwise, in multi-GPU distributed training, the first GPU may occupy excessive VRAM compared to the other GPUs.

@ExponentialML ExponentialML added the bug Something isn't working label Jun 25, 2023
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants