Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[AMD GPU] Slower speed and 40G VRAM for app #44

Open
xalteropsx opened this issue Aug 31, 2024 · 8 comments
Open

[AMD GPU] Slower speed and 40G VRAM for app #44

xalteropsx opened this issue Aug 31, 2024 · 8 comments

Comments

@xalteropsx
Copy link

i using bf16 and it occupy more then 40gb vram and very slow

@Zheng-Chong
Copy link
Owner

Running Gardio app takes only about 8G VRAM in our test. Your description about the issue is not sufficient enough. Can you provide more details for me to further check your problem?

@xalteropsx
Copy link
Author

xalteropsx commented Sep 2, 2024

Running Gardio app takes only about 8G VRAM in our test. Your description about the issue is not sufficient enough. Can you provide more details for me to further check your problem?

sorry for late reply last day was little fever i will do it now will provide u sample video and some changes what i made brb it take 10-30min

@xalteropsx
Copy link
Author

xalteropsx commented Sep 2, 2024

image
as u can see from screenshot i clone ur current repo and change this line

default="runwayml/stable-diffusion-inpainting" - > default="benjamin-paine/stable-diffusion-v1-5-inpainting"

it hard to make video at such large gpu consuming even freezing my display i will reduce the resolution so it can reduce resolution since

1024 * 768 - 42gb nearly
let me check what number should be good for 20 vram resolution so i can show u proof

@Zheng-Chong
Copy link
Owner

According to the information you provided, you are using the AMD GPU. Machine learning tasks usually require Nvidia GPU to accelerate with CUDA, otherwise it will be performed at a very slow speed.
In addition, your 46.5G VRAM is likely to be occupied by other applications. Maybe you can view it through process manager.

@xalteropsx
Copy link
Author

xalteropsx commented Sep 2, 2024

u can download the video and check it its impossible vram consume by it if u want proof i can google remote video host or google meet or provide me something i can share my screen with u? dont bother about slow speed bro all i want is less vram if possible

@Zheng-Chong Zheng-Chong changed the title metion only 8gb vram consume 40gb vram [AMD GPU] consume 40gb vram Sep 2, 2024
@Zheng-Chong Zheng-Chong changed the title [AMD GPU] consume 40gb vram [AMD GPU] Slower speed and 40G VRAM for app Sep 2, 2024
@Zheng-Chong
Copy link
Owner

I have checked the video you sent, and indeed there is a 40G memory usage. I suspect the reason is that AMD GPUs do not support low-precision or mixed-precision inference, or the current code cannot effectively utilize the AMD GPUs, leading to increased memory usage and slow speed.
However, I have too little experience with machine learning on AMD GPU to provide suggestions for reducing high memory usage. Perhaps you can seek help from relevant communities to see if there are methods to reduce memory consumption and accelerate the process.
I have modified the title of this issue so that others can view it and offer assistance.

@xalteropsx
Copy link
Author

xalteropsx commented Sep 2, 2024

i will try to investigate if i know anything or may be ask someone who using amd for this question as well but was using zluda torch in window which check in linux on pure torch and will let u know the feedback again

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants