-
-
Notifications
You must be signed in to change notification settings - Fork 136
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU memory overflow in ROCm environment #131
Comments
Hi! We've attempted to address this in #15. If anyone has idea or PR for a better solution, it would be very appreciated! |
@jhj0517 Upon checking the document at https://pytorch.org/docs/stable/notes/cuda.html#memory-management, there is a section stating, Below are the details of my testing:This is the baseline state. The Graphics pipe shown in the screenshot represents the energy in use / VRAM as GPU memory. Other applications are running, occupying 4607M of graphics memory. This is during the execution of large-v3.1 11569M, 5466M was observed. Some time after the execution of large-v3 ended, 11116M was observed. This is during the execution with the medium model. It shows 8238M of memory in use. This is just after the execution of the medium model has ended. 7838M was observed. The whisper-webui process has been terminated. The memory has returned to its initial state. |
Thanks for sharing your experience! Here's someone's experience running this web UI on an AMD GPU: According to this, faster-whisper does not work with ROCm.
|
I'm looking for an option to release GPU memory after Whisper tasks. Sometimes my PC shuts down due to overflowing GPU memory.
Running on Debian(Bookworm) + Docker(rocm/pytorch:rocm6.0.2_ubuntu22.04_py3.10_pytorch_2.1.2) + RX6800
I am aware that there is no official support for Linux or Docker environments.
However, I believe that supporting these options would definitely have a positive impact in the future.
GENERATE SUBTITLE
is finishedGENERATE SUBTITLE
is finishedI have no experience with Python, pytorch, etc., so my ability to interpret the project is limited.
If could determine whether the feature implementation is possible, I will try to contribute in any way I can.
The text was updated successfully, but these errors were encountered: