-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dynamic batchsize onnx model Strangely results in fixed batchsize output model with trtexec #996
Comments
I found that after Pytorch's interpolate with bilinear mode and align_corner=true,the resulted trt engine becomes a fixed batchsize model. It make it impossible to create a trt plan file which support dynamic batching. A very simple example could be :
then create a plan file with trtexec :
when load the trt model with trtserver, it outputs:
|
Probably, it's a bug in tensorrt's Resize op or the builder module. I hava test on tensorrt 7.2.1.4 with python sdk, code snippet:
output:
|
Hello @handoku , thanks for reporting. |
@ttyio |
Hello @handoku , broadcast and dynamic shapes are orthogonal features, though sometimes they break each other. BTW, where did you see the broadcast limitation on resize op? I have just tried 7.2.1.6, generate onnx using your instructions and generate trt engine using command line:
then using the script you provide, I see output:
Could you take a try? thanks! |
@ttyio Anyway, thanks for your test, I may have another try with Docs link : support matrix |
closing since no activity for more than 3 weeks, please reopen if you still have question, thanks! |
Description
I have exported a onnx model from pytorch,some code snippet like:
The onnx file looks ok in Netron, then I produced a trt engine with
trtexec
.However, when load the plan file with triton inference server, it outputs:
I have no idea how the
dims 5
came out.Could you help me out for this
Environment
TensorRT Version: 7.0.0
GPU Type: Tesla-T4
Nvidia Driver Version: 418.xx
CUDA Version: 10.1
Operating System + Version: ubuntu 18.04
Relevant Files
onnx file : https://drive.google.com/file/d/1hzFuOPQtN0ivOsP6IOi6h4gw3-0NPH0d/view?usp=sharing
Steps To Reproduce
then load the model with triton inference server
The text was updated successfully, but these errors were encountered: