-
Notifications
You must be signed in to change notification settings - Fork 115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[AMD GPU] Slower speed and 40G VRAM for app #44
Comments
Running Gardio app takes only about 8G VRAM in our test. Your description about the issue is not sufficient enough. Can you provide more details for me to further check your problem? |
sorry for late reply last day was little fever i will do it now will provide u sample video and some changes what i made brb it take 10-30min |
According to the information you provided, you are using the AMD GPU. Machine learning tasks usually require Nvidia GPU to accelerate with CUDA, otherwise it will be performed at a very slow speed. |
u can download the video and check it its impossible vram consume by it if u want proof i can google remote video host or google meet or provide me something i can share my screen with u? dont bother about slow speed bro all i want is less vram if possible |
I have checked the video you sent, and indeed there is a 40G memory usage. I suspect the reason is that AMD GPUs do not support low-precision or mixed-precision inference, or the current code cannot effectively utilize the AMD GPUs, leading to increased memory usage and slow speed. |
i will try to investigate if i know anything or may be ask someone who using amd for this question as well but was using zluda torch in window which check in linux on pure torch and will let u know the feedback again |
i using bf16 and it occupy more then 40gb vram and very slow
The text was updated successfully, but these errors were encountered: