Improved vanishing point reference detection to early detect and track distant oncoming vehicles for adaptive traffic light signaling
Main Article Content
Abstract
Real-time traffic monitoring is essential for the operation of an adaptive traffic lighting system and plays a significant role in decision-making, particularly signaling in roadworks. When only one lane is accessible due to temporary road blockage, early detection of oncoming vehicles is crucial to minimize bottlenecks near the traffic light that could result in congestion and accidents. This research aimed to enhance the detection and tracking of traffic at a distance from the traffic light. We utilized the vanishing point as a reference for detection and calculated the region of interest. We implemented the proposed method on twelve traffic surveillance videos and evaluated the system performance based on how quickly it could detect incoming traffic compared with the R-CNN method. The proposed method detected target vehicles in an average of 17.75 frames, while the R-CNN method required an average of 63.36 frames. Moreover, the proposed method’s precision depends on the number of pixel orientations used to estimate the vanishing point and the definition of the region of interest. Therefore, the proposed method for enhancing the safety and reliability of an adaptive traffic light system is reliable.
Downloads
Article Details
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work
References
2. Samoail, N. A. & Yousif, S. A study into urban roadworks with shuttle-lane operation. Proc. Conf. Mathematics in Transport Planning and Control. 1998, 79-88, doi:10.1108/9780585474182-008.
3. Schwietering, C. & Feldges, M. Improving Traffic Flow at Long-term Roadworks. Transportation Research Procedia. 2016, 15, 267-282, doi: 10.1016/j.trpro.2016.06.23.
4. Yousif, S; Nassrullah, Z; Norgate, S. H. Narrow lanes and their effect on drivers’ behaviour at motorway roadworks. Transportation Research Part F. 2017, 47, 86-100, doi: 10.1016/j.trf.2017.04.016
5. Walker, G. & Calvert, M. Driver behaviour at roadworks. Applied Ergonomics. 2015, 51, 18-29, doi: 10.106/j.apergo.2015.03.019.
6. Araghi, S.; Khosravi, A.; Creighton, D. A review on computational intelligence methods for controlling traffic signal timing. Expert Syst. Appl. 2015, 42, 1538–1550, doi:10.1016/j.eswa.2014.09.003.
7. Lian, F.; Chen, B.; Zhang, K.; Miao, L.; Wu, J.; Luan, S. Adaptive traffic signal control algorithms based on probe vehicle data. J. Intell. Transp. Syst. Technol. Planning, Oper. 2021, 25, 41–57, doi:10.1080/15472450.2020.1750384.
8. Koziarski, M.; Cyganek, B. Impact of low resolution on image recognition with deep neural networks: An experimental study. Int. J. Appl. Math. Comput. Sci. 2018, 28, 735–744, doi:10.2478/amcs-2018-0056.
9. Xue, L.; Zhong, X.; Wang, R.; Yang, J.; Hu, M. Low - resolution vehicle recognition based on deep feature fusion. Multimed. Tools Appl. 2018, 77, 27617–27639, doi:10.1007/s11042-018-5940-6.
10. Zou, W.; Yuen, P.C. Very low resolution face recognition problem. IEEE 4th Int. Conf. Biometrics Theory, Appl. Syst. BTAS 2010 2010, doi:10.1109/BTAS.2010.5634490.
11. Yang, W.; Luo, X.; Fang, B.; Zhang, D.; Tang, Y.Y. Fast and accurate vanishing point detection in complex scenes. 2014 17th IEEE Int. Conf. Intell. Transp. Syst. ITSC 2014 2014, 93–98, doi:10.1109/ITSC.2014.6957672.
12. Welch, G.; Bishop, G. An Introduction to the Kalman Filter. In Pract. 2006, 7, 1–16, doi:10.1.1.117.6808.
13. Matlab Motion-Based Multiple Object Tracking Available online: https://de.mathworks.com/help/vision/ug/motion-based-multiple-object-tracking.html.
14. Moghadam, P.; Starzyk, J.A.; Wijesoma, W.S. Fast vanishing-point detection in unstructured environments. IEEE Trans. Image Process. 2012, 21, 425–430, doi:10.1109/TIP.2011.2162422.
15. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2014, 580–587, doi:10.1109/CVPR.2014.81.
16. Matlab Train Object Detector Using R-CNN Deep Learning Available online: https://de.mathworks.com/help/vision/ug/object-detection-using-deep-learning.html.