-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add VAE to txt-to-speech Inference #32
Comments
Hey, @digiphd! Thanks for getting this on my radar. I'll have a chance to take a look during this coming week. As a preliminary comment, I like the idea of being able to switch the VAE at runtime, although there will be a lot of work involved to adapt how we currently cache models. P.S. If you're impatient, in the meantime, I think you could probably:
Alternatively, with your current setup, it's possible that if you set Anyways, have a great weekend and we'll be in touch next week 😀 |
Hey @gadicc great, thanks for your suggestions I will give them ago! You're a legend! Another thing I was wondering, was if docker-diffusers-api text-to-image supports negative keywords? I did put it as an argument and it seemed to negatively affect the output images. |
Yup! The There's also a note there now about using the Thanks for all the kind words! 🙌 |
Hey @digiphd, I had a quick moment to try These worked for me: $ python test.py txt2img --call-arg MODEL_ID="dreamlike-art/dreamlike-photoreal-2.0" --call-arg MODEL_PRECISION=""
$ python test.py txt2img --call-arg MODEL_ID="dreamlike-art/dreamlike-photoreal-2.0" --call-arg MODEL_PRECISION="fp16" I just tried in the default "runtime" config. If you have this issue specifically in the |
Related: #26 |
Hey hey!
So I am using some models that either have VAE baked in or require a separate VAE to be defined during inference like this:
when I either manually added the vae or used a model with a vae baked in for the
MODEL_ID
, I received the following error, for example with the modeldreamlike-art/dreamlike-photoreal-2.0
Line 382 in the inference function which looks like this:
images = pipeline(**model_inputs).images
Perhaps we need to add a .half() to the input somewhere, not sure where. though.
Any help would be greatly appreciated!
It's the last hurdle I am facing to be generating images.
IDEA:
It would be awesome if we could define an optional VAE when making API call like this:
The text was updated successfully, but these errors were encountered: