-
Notifications
You must be signed in to change notification settings - Fork 3.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tacotron get slower using pytorch TRT #545
Comments
one more question, I want to know what cause this difference, for the fp16 result. |
You can see that results in main readme wasn't updated for a lot longer then results in trt repo. |
Thanks for your reply. I test trt_cpp and get better inference speed now. |
Nope I never used benchmark scripts, also I used my own trained models. |
Hi, I follow your guide, get the right docker env.
I run those script successfully, but the result I got is weird.
My Gpu is V100, I test three type of inference:
No amp, amp-run (fp16), and pytorch trt.
the total latency of this three are: 1.54, 1.3627, 1.34 (decrease as expect)
the waveglow latency decrease as expect: 0.46065, 0.3097, 0.1374
but the tacotron latency of trt get even longer! 1.08, 1.053, 1.208
the tacotron2 model I use is:
https://ngc.nvidia.com/catalog/models/nvidia:tacotron2pyt_fp16/files?version=3
Because you only provide the result of trt accerlerate on T4 GPU. I want to know is there anything wrong with my result? Why the trt make tacotron2 inference slower??
Thanks for your patience and Looking forward your reply.
The text was updated successfully, but these errors were encountered: