Should this work, even if slowly, on just a 2080Ti?
See original GitHub issueHi, just tried to run it via cog
using r8.im/kuprel/min-dalle
and even though it setups correctly I’m getting a PyTorch OOM error :
RuntimeError: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 10.73 GiB total capacity; 9.87 GiB already allocated; 20.06 MiB free; 9.88 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
This is “just” a gaming card but I’m wondering still if that’s a bug, a hard limit on minimum hardware, or rather a default parameter based on T4 starting at 16GB of VRAM being the bare minimum replicate
expects that could eventually be changed. Or even if starting from this repository instead could solve it.
PS: nvidia-smi
shows no other process using the GPU, 11264MiB to use and CUDA 11.7
Issue Analytics
- State:
- Created a year ago
- Comments:7 (4 by maintainers)
Top GitHub Comments
Super, thanks in indeed it works now relying on this Dockerfile :
Yeah float16 from the repo might work. You should be able to just run the notebook.