You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One of the ways to accelerate running the NNTrainer in inference mode is to fuse operations.
We are currently using this fusion when exporting to TensorFlow Lite.
By applying it to the current NNTrainer, we can improve the speed during inference.
Many Deeplearning Model using Batchnorm after Conv layer
lhs8928
changed the title
Support Convolution&BAtchnorm Fusing for Optimized Inference Mode
Support Convolution&Batchnorm Fusing for Optimized Inference Mode
Jul 4, 2024
Fusing the operation on infererence means that this graph is little bit different from the graph used to train the model.
To support fusing operation we might need to make 2 graphs for 1 model. (One for training and the other one for inference)
Fusing the operation on infererence means that this graph is little bit different from the graph used to train the model. To support fusing operation we might need to make 2 graphs for 1 model. (One for training and the other one for inference)
That's right !! when we save()&load() model we check [INFERENCE|TRAIN] Mode and make graph
One of the ways to accelerate running the NNTrainer in inference mode is to fuse operations.
We are currently using this fusion when exporting to TensorFlow Lite.
By applying it to the current NNTrainer, we can improve the speed during inference.
Many Deeplearning Model using Batchnorm after Conv layer
and when inference we can fusing ops to below
image ref
The text was updated successfully, but these errors were encountered: