-
Notifications
You must be signed in to change notification settings - Fork 71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
python api code for text to video generation? #633
Comments
Hi @SutirthaChakraborty with opensora_hpcai, we support generating 720P videos of 16 seconds (408 frames). For model combination, do you mean combining text-to-image model with image-to-video model? If so, I would suggest using SD3 or Flux.1 (in PR) for T2I generation followed by DynamiCrafter for I2V generation for the best visual quality. If you prefer long videos, you may use opensora_hpcai to do I2V. Thanks for your attention to our aigc kit. |
Hi @SamitHuang Thanks for your detailed reply. |
Sure.
>>> import mindspore
>>> from mindone.diffusers import StableDiffusion3Pipeline
>>> pipe = StableDiffusion3Pipeline.from_pretrained(
... "stabilityai/stable-diffusion-3-medium-diffusers",
... mindspore_dtype=mindspore.float16,
... )
>>> prompt = "A cat holding a sign that says hello world"
>>> image = pipe(prompt)[0][0]
>>> image.save("sd3.png")
|
There is no direct way to run the dynamicrafter? |
Sorry that it can only run with |
1 similar comment
Sorry that it can only run with |
How long videos can we generate? what would be the best model combinations ?
The text was updated successfully, but these errors were encountered: