Closed
Description
I am trying to use multiple GPUs to generate a single image. Not batch images in parallel like #2977
I want my images to generate fast and not be bottlenecked by memory constants when I generate larger images or attempt in/out painting. I tried to use deepspeed; however, wsl2 is insanely slow and deepspeed just doesn't install the inference module I wanted to use to achieve the multigpu.
(Note: I can reinstall whatever is needed, I've just uninstalled and reinstalled so many versions to attempt to get deepspeed working I just gave up at this point.)
Current Environment:
- Windows 10
- Python 3.10
- Cuda Toolkit 12.1
- Torch 2.1.0
- 2 Nvidia RTX 4090
- AMD 5950x
- 128 GB RAM
- 2 4tb nvme sn850x drives