-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[How to release cpu memory after session Run] #20640
Labels
platform:windows
issues related to the Windows platform
Comments
github-actions
bot
added
ep:CUDA
issues related to the CUDA execution provider
platform:windows
issues related to the Windows platform
labels
May 10, 2024
When using the C++ API, you probably do not want to call onnxruntime/include/onnxruntime/core/session/onnxruntime_cxx_api.h Lines 575 to 581 in e81c867
The underlying C API release function should get called automatically when the C++ API object goes out of scope. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Describe the issue
hi,Here is a very common situation: after using ONNXruntime for inference, the system has nearly 2GB of memory(not gpu memory) that cannot be released. I have tried many ways to release it, but none have solved the problem....Does ONNXruntime not provide a mechanism to release CPU memory after inference?
To reproduce
This is a common and recurring issue in many version.
Urgency
No response
Platform
Windows
OS Version
10
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
15.1
ONNX Runtime API
C++
Architecture
X64
Execution Provider
CUDA
Execution Provider Library Version
No response
Model File
No response
Is this a quantized model?
No
The text was updated successfully, but these errors were encountered: