You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hi,Here is a very common situation: after using ONNXruntime for inference, the system has nearly 2GB of memory(not gpu memory) that cannot be released. I have tried many ways to release it, but none have solved the problem....Does ONNXruntime not provide a mechanism to release CPU memory after inference?
Describe the issue
hi,Here is a very common situation: after using ONNXruntime for inference, the system has nearly 2GB of memory(not gpu memory) that cannot be released. I have tried many ways to release it, but none have solved the problem....Does ONNXruntime not provide a mechanism to release CPU memory after inference?
To reproduce
This is a common and recurring issue in many version.
Urgency
No response
Platform
Windows
OS Version
10
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
15.1
ONNX Runtime API
C++
Architecture
X64
Execution Provider
CUDA
Execution Provider Library Version
No response
Model File
No response
Is this a quantized model?
No
The text was updated successfully, but these errors were encountered: