Fusion of Deep Sort and Yolov5 for Effective Vehicle Detection and Tracking Scheme in Real-Time Traffic Management Sustainable Systemopen access
- Authors
- Kumar, Sunil; Singh, Sushil Kumar; Varshney, Sudeep; Singh, Saurabh; Kumar, Prashant; Kim, Bong-Gyu; Ra, In-Ho
- Issue Date
- Dec-2023
- Publisher
- MDPI
- Keywords
- convolutional neural networks; deep SORT; Kalman filter; vehicle detection and tracking; YOLOv5
- Citation
- Sustainability, v.15, no.24, pp 1 - 24
- Pages
- 24
- Indexed
- SCIE
SSCI
SCOPUS
- Journal Title
- Sustainability
- Volume
- 15
- Number
- 24
- Start Page
- 1
- End Page
- 24
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/22796
- DOI
- 10.3390/su152416869
- ISSN
- 2071-1050
2071-1050
- Abstract
- In recent years, advancements in sustainable intelligent transportation have emphasized the significance of vehicle detection and tracking for real-time traffic flow management on the highways. However, the performance of existing methods based on deep learning is still a big challenge due to the different sizes of vehicles, occlusions, and other real-time traffic scenarios. To address the vehicle detection and tracking issues, an intelligent and effective scheme is proposed which detects vehicles by You Only Look Once (YOLOv5) with a speed of 140 FPS, and then, the Deep Simple Online and Real-time Tracking (Deep SORT) is integrated into the detection result to track and predict the position of the vehicles. In the first phase, YOLOv5 extracts the bounding box of the target vehicles, and in second phase, it is fed with the output of YOLOv5 to perform the tracking. Additionally, the Kalman filter and the Hungarian algorithm are employed to anticipate and track the final trajectory of the vehicles. To evaluate the effectiveness and performance of the proposed algorithm, simulations were carried out on the BDD100K and PASCAL datasets. The proposed algorithm surpasses the performance of existing deep learning-based methods, yielding superior results. Finally, the multi-vehicle detection and tracking process illustrated that the precision, recall, and mAP are 91.25%, 93.52%, and 92.18% in videos, respectively.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Department of Mechanical, Robotics and Energy Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.