-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Converting engine file from onnx file with ReduceMax failure of TensorRT 8.5.10 when running trtexec on GPU Orin #3866
Comments
Please add |
Hi, I replaced the original nn.Layernorm block by nn.BatchNormalization block. Now my new network onnx file is :
Please check and have a nice day |
And if I remove the LayerNorm or BatchNormlization block, can success to generate the engine file. |
You can try to convert these two modules(LayerNorm or BatchNormlization block as a subgraph onnx) separately. |
no disk space? |
Hi, thanks for your reply. I tried again with new .pt, and success to create the engine file. |
Please check our release note, I think you need at least TRT 8.6 or 9.0, can't remember exactly which one. |
Description
I tried to generate engine file from onnx file on Orin GPU, but it failed:
[05/15/2024-11:45:16] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +0, GPU +4, now: CPU 0, GPU 4 (MiB)
[05/15/2024-11:45:16] [E] Saving engine to file failed.
[05/15/2024-11:45:16] [E] Engine set up failed
Environment
TensorRT Version:
NVIDIA GPU:
NVIDIA Driver Version:
CUDA Version:
CUDNN Version:
Operating System:
Python Version (if applicable):
Tensorflow Version (if applicable):
PyTorch Version (if applicable):
Baremetal or Container (if so, version):
Relevant Files
Model link:
Steps To Reproduce
Commands or scripts:
Have you tried the latest release?:
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
):The text was updated successfully, but these errors were encountered: