-
Notifications
You must be signed in to change notification settings - Fork 369
Open
Labels
bugSomething isn't workingSomething isn't working
Description
Bug Description
A user reported Torch-TRT has slower inference time on RAFT compared to the original Pytorch model:
Backend Time (ms) Speedup
----------------------------------------------------------------------
Original 7.12 -
Torch-TensorRT 20.35 0.35x
ONNX-TRT 2.96 2.41x
I printed out the engine profiles of Torch-TensorRT and ONNX-TRT where the num of layers of Torch-TRT are twice more than ONNX-TRT.
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working