-
Notifications
You must be signed in to change notification settings - Fork 11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Much lower frame rate has been observed on the 1080ti #16
Comments
According to the below test, the speed was fine with the normal
I will be glad to assist you in finding the problem or help you update the codes as follows(these are simple for you but maybe you don't have the time for that)
|
I tried to use 3090 to test image, and the average inference speed is 160ms per image(1918x1280, batch size1), and maybe some information in CMakeList (sm version )influence the inference performance. I suggest to comment some info in CMakeList.txt : #option(CUDA_USE_STATIC_CUDA_RUNTIME OFF) |
Please note that I have changed the sm version to the 1080ti corresponding as follows:
It got worse after I commented on these lines. Time for inference has been changed from 600ms to 1400ms. |
I can not test 1080ti performance but I searched and find a similar issue: NVIDIA/TensorRT#1221 maybe use different cudnn can help |
Hi, could I know you how to get the test result here? I tried the |
https://github.com/wang-xinyu/tensorrtx/tree/master/unet please use newest repo for testing, this old repo is no longer maintained |
Hello, I've tested the repo on the 1080ti and on the below configuration I see that the elapsed time is about 600ms, which is extremely high, and I'm thinking maybe I'm doing something wrong.
I have downloaded the weights from this link and converted them. I checked the MD5sum output with you and it was fine.I have attached the inference log.
Test in pytorch
The same configuration was tested in Pytorch and the logs are attached
Could you please let me know why it takes so long? Thank you in advance
Saeed
The text was updated successfully, but these errors were encountered: