Training on Google's Image Captioning dataset (~3.3 million images) #125
TheoCoombes
started this conversation in
Show and tell
Replies: 2 comments 5 replies
-
Hello TheCoombes! I was about to do exactly the same thing. If you could share your parameters and how you managed to integrate this I would be very glad |
Beta Was this translation helpful? Give feedback.
0 replies
-
Any update on this? Excited to see how things went. |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi guys,
I've just finished playing around with
train_dalle.py
and have finally integrated Google's Conceptual Captioning dataset into DALLE-pytorch. It uses the requests module to stream the data from the URLs instead of having to download all 3 million images (I don't have the disk space for that lmao). Whilst the dataset was not explicitly designed for text to image generation, (instead for automatic image captioning), I believe it still has a major use case here.I will be sharing the results as well as my code once it has finished its initial training session.
(p.s. I am very new to ML so if my code is a mess then that's probably why haha)
Beta Was this translation helpful? Give feedback.
All reactions