You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I tested the same model on the same machine with TensorRT 8.6.1.6 and 8.4.0.6 as following steps
use bin/trtexec to trans onnx model to engine file with 8.6 and 8.4 respectively by the same command bin/trtexec --onnx=xxxx.onnx --saveEngine=xxx.engine(this model has fixed input shapes)
load this engine in the same c++ code and run.
check gpu memory usage by nvidia-smi
The following are the results
8.6
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 510.47.03 Driver Version: 510.47.03 CUDA Version: 11.6 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 Tesla T4 Off | 00000000:3B:00.0 Off | 0 |
| N/A 66C P0 63W / 70W | 283MiB / 15360MiB | 61% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
8.4
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 510.47.03 Driver Version: 510.47.03 CUDA Version: 11.6 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 Tesla T4 Off | 00000000:3B:00.0 Off | 0 |
| N/A 69C P0 73W / 70W | 853MiB / 15360MiB | 59% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
It can be seen that the gpu memory consumption is significantly reduced from the above.
So
is there any gpu memory optimization in TensorRT 8.6.1.6 compare to 8.4.0.6 ?
If not, is there anything special parameter should be set when using bin/trtexec in 8.4 but default in 8.6 ?
Thanks !!!
The text was updated successfully, but these errors were encountered:
Hi, I have a further question about difference between createExecutionContext and createExecutionContextWithoutDeviceMemory, Is there have any gpu memory difference when using the corresponding context ?
Hi, I tested the same model on the same machine with TensorRT 8.6.1.6 and 8.4.0.6 as following steps
bin/trtexec
to trans onnx model to engine file with 8.6 and 8.4 respectively by the same commandbin/trtexec --onnx=xxxx.onnx --saveEngine=xxx.engine
(this model has fixed input shapes)nvidia-smi
The following are the results
8.6
8.4
It can be seen that the gpu memory consumption is significantly reduced from the above.
So
bin/trtexec
in 8.4 but default in 8.6 ?Thanks !!!
The text was updated successfully, but these errors were encountered: