Mannequins/Blogpost Trained with 16K VQGAN
See original GitHub issuehttps://www.dropbox.com/s/x9gbxiartrqaewf/afiaka_544I_96TSL_12D_16H_64DH_sparse.zip
@robvanvolt @rom1504 sorry i dont have time to make a PR to dalle-models yet but I will asap.
If you’re here from the discord - sorry I didn’t leave this better packaged. The zip is 2 GiB but the main dalle.pt
is just 445 megabytes. It’s not a huge model or anything but you don’t really need one on this dataset. If you wanna finetune from it, i’ve included in the optimizer states from the deep speed training session as well. I used deep speeds WarmupDecayScheduler.
Anyway I finally hit the loss floor with it which I believe to be around 2.2. If anyone thinks they can get lower than that - good luck.
Sorry there are no generations posted - I’m sure they look great; but I’m only posting because there are newer…better VQGANs out now. And I must of course switch to those.
Issue Analytics
- State:
- Created 2 years ago
- Reactions:3
- Comments:6 (3 by maintainers)
Top GitHub Comments
Made a colab specifically for inferencing this model: https://colab.research.google.com/drive/11V2xw1eLPfZvzW8UQyTUhqCEU71w6Pr4?usp=sharing
Awesome stuff! I’ll try to be testing / playing with this model on my part and possibly add it to the colab if your comfortable with that.