-
-
Notifications
You must be signed in to change notification settings - Fork 215
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Should the module be unloaded from VRAM after its use? #325
Comments
Hi. If you're able to run Because the expected behavior when changing the Whisper model is to update the current model to it, not load it additionally. But if you tried to run Music removal model together while transcribing, you might get CUDA errors if you have <12GB VRAM. |
VRAM wise I should be OK, I have 12GB (3060), other AI stuff is running on another GPU. I noticed that other model managers such ollama offload the models after certain time of not being used, or even they unload them when the user want to select a different model. I.e. If i try a transcription with large-v2 and I don't like the result and I want to try large-v3, I need to shutdown the webui to offload the large-v2 model, as it's always in memory. |
This is weird and not expected behavior. If you're able to run If each different model runs fully on a different GPU, this should not happen. Probably something is wrong with the setup, but I don't have multiple GPUs so I can't reproduce/test about it. |
Not exactly sure what is it, after the transcription finish, using large-v3, or any other model, there's a remaining processes in the gpu: Wed Oct 9 11:30:47 2024 +-----------------------------------------------------------------------------------------+ This is the python3.10 using to run the webui it's like it never offload the model competely from vram. |
Which OS are you using?
I've noticed that after running a transcription the model remains on VRAM making impossible to do another transcription with a different model as there's not enough vram. Is there any way to offload the model after certain period of incatvity?
Thanks.
The text was updated successfully, but these errors were encountered: