You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It just be too small to consume the power of GPU, e.g. consider the difference between mobilenet and resnet. You can increase the throughput(not inference time) by increase batch or use multi-thread inference.
Description
When I run yolov8 model on 3060, only 7% of GPU is used. I want to run more the one model in parallel in order to reduce the inference duration.
Environment
TensorRT Version:
8.6
NVIDIA GPU:
3060
NVIDIA Driver Version:
CUDA Version:
12.1
CUDNN Version:
8.6
Operating System:
Python Version (if applicable):
Tensorflow Version (if applicable):
PyTorch Version (if applicable):
Baremetal or Container (if so, version):
Relevant Files
Model link:
Steps To Reproduce
Commands or scripts:
Have you tried the latest release?:
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
):The text was updated successfully, but these errors were encountered: