TensorRT produces wrong results when running valid onnx model on GPU 3080 #4473
Labels
Module:Accuracy
Output mismatch between TensorRT and other frameworks
triaged
Issue has been triaged by maintainers
Description
For the following valid onnx model,
it can be executed by onnxruntime. The results are as follows:
However, when I run this model using TensorRT, it produces wrong results as follows:
Environment
TensorRT Version: 10.11.0.33
NVIDIA GPU: GeForce RTX 3080
NVIDIA Driver Version: 535.183.01
CUDA Version: 12.2
CUDNN Version: none
Operating System: ubuntu 20.04
Python Version (if applicable): 3.12.9
Tensorflow Version (if applicable): none
PyTorch Version (if applicable): none
Baremetal or Container (if so, version): none
Steps To Reproduce
This bug can be reproduced by the following code with the model in the attachment. As shown in the code, the model can be executed by onnxruntime.
testcase.zip
Commands or scripts:
Have you tried the latest release?: yes
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
): the mode can be executed by onnxruntime.The text was updated successfully, but these errors were encountered: