-
Notifications
You must be signed in to change notification settings - Fork 165
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the size of engine file #66
Comments
Are they same calibrations data feeding? And are they the same tensorRT version? My int8 engine by 1060 is also the 60M. Maybe your TiTAN X is created as mode fp32? I am very interesting in your results. Please inform me.Thanks |
They are same calibrations data feeding and tensorRT version, my TITAN X is create as model int8. The detection results is good on your 1060 as model int8? The lager defferences of detetcion results between int8 and float16? |
I created the int8 engine in 1060 1070, 2060 and T4, The speed become more and more slower, the test results become more and more better, and the files become more and more bigger. |
when I create Int8 engine on TiTan X, the warings "Int8 supprot resquested on hardware without native Int8 support, preformance will be negatively affected ...", but the engine file is created and I test the images and the detetction results is good, the speed become faster. The defference of sizes of engine file between int8 and float16 is small. You meet the problem? |
I have solved it, the accuracy is poor because I only used one image as the calibration data . The ImageNet need 500 images to calibrate the engine, the result become good when I used 100 images as the calibration data. |
I create the int8 engine file on GPU 1060, the size of engine file is only 60M, the detection result is very poor. But I create the int8 engine file on GPU TiTAN X, the size of engine file is only 500-600M, the detection result is good.
The text was updated successfully, but these errors were encountered: