2021 IEEE International Symposium on Workload Characterization (IISWC) 2021
DOI: 10.1109/iiswc53511.2021.00030
|View full text |Cite
|
Sign up to set email alerts
|

Demystifying TensorRT: Characterizing Neural Network Inference Engine on Nvidia Edge Devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 32 publications
(12 citation statements)
references
References 31 publications
0
3
0
Order By: Relevance
“…The proposed architecture was deployed, in real‐time, on the Nvidia tensor‐RT framework to check its potential applications for a real‐time EEG‐BCI. It is a C++ library that helps to boost the inference time of the NVIDIA GPUs (Shafi et al, 2021). Figure 6 shows a flow chart of the architecture inferencing in the Tensor RT framework.…”
Section: Resultsmentioning
confidence: 99%
“…The proposed architecture was deployed, in real‐time, on the Nvidia tensor‐RT framework to check its potential applications for a real‐time EEG‐BCI. It is a C++ library that helps to boost the inference time of the NVIDIA GPUs (Shafi et al, 2021). Figure 6 shows a flow chart of the architecture inferencing in the Tensor RT framework.…”
Section: Resultsmentioning
confidence: 99%
“…However, not all objects need to be detected, so we focused on the 17 most common categories. The process involves inputting color images to YOLOV5, accelerated with TensorRT [29], accelerating to obtain semantic labels of the COCO categories. Bounding boxes help approximate the region of dynamic objects in the images, and feature points within these boxes are assigned semantic labels.…”
Section: Semantic Label Incremental Updating With Bayes' Rulementioning
confidence: 99%
“…This requires the design of accelerators to speed up neural network inference in edge scenarios. The mainstream choice for accelerating neural network inference in edge scenarios is through the design of specific hardware accelerators such as NVDLA and other NPUs [29]. However, hardware accelerators with general architectures implemented through ASICs not only have a high design difficulty and a long development cycle but also may not be sufficient to meet the real-time requirements in terms of acceleration ratio.…”
Section: Introductionmentioning
confidence: 99%