Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inference time using TF-TRT is the same as Native Tensorflow for Object Detection Models (SSD Resnet640x640 and EfficientDetD0) #287

Open
Abdellah-Laassairi opened this issue Feb 25, 2022 · 3 comments

Comments

@Abdellah-Laassairi
Copy link

Description

I obtained the same inference time with my optimized model (sometimes slower the baseline model) using the Tensorflow TensorRT API.
I’ve included a set of two tests on both SSD Resnet640x640 and EfficientDetD0.

Environment

TensorRT Version: 8.2.2
GPU Type: Tesla T4
Nvidia Driver Version: 450.51.05
CUDA Version: 11.6
CUDNN Version: 7.0.0
Python Version: 3.8
TensorFlow Version: 2.7.0
Container : nvcr.io/nvidia/tensorflow:22.01-tf2-py3(build 31081301)

Relevant Files

Models obtained from Tensorflow Object Detection API Models Zoo

Steps To Reproduce

Github Repository containing all the notebooks with results and steps to reproduce

@Abdellah-Laassairi
Copy link
Author

Any update?

@ncomly-nvidia
Copy link

Hi @Abdellah-Laassairi. We are aware of some minor issues in object detection networks causing the TensorRT engines to fallback to tensorflow, giving 1:1 performance in TF-TRT and native TF.

We have resolved many of these issues in our 22.04 container which will be available later this month!

@Abdellah-Laassairi
Copy link
Author

Thanks! looking forward to it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants