Hanjun Kim Professor School of Electrical and Electronic Engineering, Yonsei University Ph.D. 2013, Department of Computer Science, Princeton University Office: Engineering Hall #3-C415 Phone: +82-2-2123-2770 Email: first_name at yonsei.ac.kr |
[Home] [Curriculum Vitae] [Publications] [CoreLab] [Korean] |
Refereed International Conference PublicationsReal-Time Object Detection System with Multi-Path Neural Networks [abstract] (IEEE Xplore, PyTorch, Detectron, PDF)
Thanks to the recent advances in Deep Neural
Networks (DNNs), DNN-based object detection systems becomes
highly accurate and widely used in real-time environments such
as autonomous vehicles, drones and security robots. Although
the systems should detect objects within a certain time limit that
can vary depending on their execution environments such as
vehicle speeds, existing systems blindly execute the entire long-
latency DNNs without reflecting the time-varying time limits,
and thus they cannot guarantee real-time constraints. This work
proposes a novel real-time object detection system that employs
multi-path neural networks based on a new worst-case execution
time (WCET) model for DNNs on a GPU. This work designs
the WCET model for a single-layer of DNNs analyzing processor
and memory contention on GPUs, and extends the WCET model
to the end-to-end networks. This work also designs the multi-
path networks with three new operators such as skip, switch,
and dynamic generate proposals that dynamically change their
execution paths and the number of target objects. Finally, this
work proposes a path decision model that chooses the optimal
execution path at run-time reflecting dynamically changing en-
vironments and time constraints. Our detailed evaluation using
widely-used driving datasets shows that the proposed real-time
object detection system performs as good as a baseline object
detection system without violating the time-varying time limits.
Moreover, the WCET model predicts the worst-case execution
latency of convolutional and group normalization layers with
only 28% and 64% errors on average, respectively.
|