-
Notifications
You must be signed in to change notification settings - Fork 388
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of GPU memory when creating multiple sessions #94
Comments
The problem is similar to my problem. Can we optimize it? |
It sounds like this may be an issue in the upstream llama.cpp library. Do you have this same problem if you use that directly? |
It also get the same error when I repeat session creation and closing on winform app. |
Is this still an issue with newer versions of LLamaSharp/llama.cpp? |
Since there's been no response for a while I'll close this, it should be fixed in newer versions. If it's still a problem please feel free to re-open the issue! |
I have tested the ability to recover memory when we create a new session and close them again.
Looks like the GPU memory is not freed.
With my GPU 3080 -8G VRAM -->I created and closed the session 4 times then the memory was thrown.
I see in the source code: llama.cpp/ggml-cuda.cu
Can we release the GPU memory?
Please refer to the error information below.
Hope you have a new update.Thank!
The text was updated successfully, but these errors were encountered: