RAM memory leak #355
Labels
No Label
bug
duplicate
enhancement
help wanted
insufficient info
invalid
news
not a bug
question
wontfix
No Milestone
No project
No Assignees
3 Participants
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: mrq/ai-voice-cloning#355
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Hello, I've been using this project for a couple of weeks, and been having issues after prolonged usage.
I am using the project to try and make audio books so my inputs are generally long. The inputs are usually fine but over time, the ram usage continues to climb higher until the program crashes or the system crashes. I'm not sure if there is a better way, but at the moment after generating for a period, I will have to close and reopen the app to reclaim the memory in ram.
I tried doing various things, it is similar to #61 sort of, but instead of gpu vram it is my ram.
I am using linux mint so linux flavor of install. It works fine until I run out of ram and either the application crashes or my system crashes because of no more ram.
Ah yeah. I'm not too sure where the issue lies, as even using the repo to transcribe large datasets will eventually hang after a while, from presumably a memory leak somewhere. I'm just not too sure where if it's independent from continuous transcriptions. For my problem, I just modified the
cli.py
, and the only problem was having to constantly reload models.For normal people, the existing
cli.py
should be fine enough, a shell script like this should be fine:The downside is that, each call to
cli.py
will require the models to reload again. Using the websocket server added in might be a better approach to work around constantly reloading the model, but you'll most likely still run into a memory leak problem (assuming the issue is not gradio).I see, thanks for the insight, yeah I'm not sure either. I tried doing a single long generation with a delimiter and multiple smaller ones but the memory never clears unless I just kill the python script.
Loading the models with massive amount of input sounds like it will take significantly longer so I'm unsure if the cli.py method is desirable sadly.
I may just try monitoring memory usage and then kill and restart it programmatically since I was sort of doing a process where I generate text, send it somewhere to clean up the output, and then send it somewhere else to match the voice more closely. I was going to automate it down the line anyways so another cog might not be too bad...
I will leave up the issue just in case
I finally got around to messing with this again, and it seems that the issue does not occur if using the api, I will have to test if providing longer inputs changes anything, but currently using roughly 100 lines of input does not seem to have the memory leak compared to using the webui. I made roughly 5 different api calls and the memory stayed roughly the same.
Nevermind, after trying to do it overnight, it died from the memory leak.
Hey, after prolonged usage, I’m having the same problem I think. I cannot generate anything else through Mrq without having the OOM Cuda allocation error. What’s even weirder, I just shut down my computer yesterday thinking that it would clear the GPU ram, but this morning after launching a fresh instance of gradio, I immediately got the OOM message
EDIT : after experimenting with this, I found out that you can shut down the Gradio interface, and then while still in the dedicated environment (conda in my case), you write
And then when I restart the UI I don't have any problem anymore. Will it remain so? Time will tell