e-ISSN 2231-8542
ISSN 1511-3701

Home / Regular Issue / JTAS Vol. 30 (4) Oct. 2022 / JST-3189-2021


Data Safety Prediction Using Bird’s Eye View and Social Distancing Monitoring for Penang Roads

Lek Ming Lim, Majid Khan Majahar Ali, Mohd. Tahir Ismail and Ahmad Sufril Azlan Mohamed

Pertanika Journal of Tropical Agricultural Science, Volume 30, Issue 4, October 2022


Keywords: Bird’s eye view, near miss, social distancing monitoring vehicle detection

Published on: 28 September 2022

In terms of fatalities, Malaysia ranks third among ASEAN countries. Every year, there is an increase in accidents and fatalities. The state of the road is one factor contributing to near misses. A near miss is an almost-caused accident, an unplanned situation that could result in injury or accidents. The Majlis Bandar Pulau Pinang (MBPP) has installed 1841 closed-circuit television (CCTV) cameras around Penang to monitor traffic and track near miss incidents. When installing CCTVs, the utilisation of video allows resources to be used and optimised in situations when maintaining video memories is difficult and costly. Highways, industrial regions, and city roads are the most typical places where accidents occur. Accidents occurred at 200 per year on average in Penang from 2015 to 2017. Near misses are what create accidents. One of the essential factors in vehicle detection is the “near miss.” In this study, You Only Look Once version 3 (YOLOv3) and Faster Region-based Convolutional Neural Network (Faster RCNN) are used to solve transportation issues. In vehicle detection, a faster RCNN was used. Bird’s Eye View and Social Distancing Monitoring are used to detect the only vehicle in image processing and observe how near misses occur. This experiment tests different video quality and lengths to compare test time and error detection percentage. In conclusion, YOLOv3 outperforms Faster RCNN. In high-resolution videos, Faster RCNN outperforms YOLOv3, while in low-resolution videos, YOLOv3 outperforms Faster RCNN.

  • Albelwi, S., & Mahmood, A. (2017). A framework for designing the architectures of deep convolutional neural networks. Entropy, 19 (242), 1-20.

  • Aldred, R. (2016). Cycling near misses: Their frequency, impact and prevention. Transport Research Part A, 90(1), 69-83.

  • Aldred, R., & Crosweller, S. (2015). Investigating the rates and impacts of near misses and related incidents among UK cyclists. Journal of Transport & Health, 2(3), 379-393.

  • Alganci, U., Soydas, M., & Sertel, E. (2020). Comparative research on deep learning approaches for airplane detection from very high-resolution satellite images. Remote sensing, 12(3), Article No. 458.

  • Aqqa, M., Mantini, P., & Shah, S. K. (2019, February). Understanding how video quality affects object detection algorithms. In VISIGRAPP (5: VISAPP) (pp. 96-104). Science and Technology Publications.

  • Arinaldi, A., Pradana, J., A., & Gurusniaga, A., A. (2018). Detection and classification of vehicles for traffic video analytics. Procedia Computer Science, 144, 259-268.

  • Behl, A., Bhatia, A., & Puri, A. (2014). Convolution and applications of convolution. International Journal of Innovative Research in Technology (IJIRT), 1(6), 2123-2126.

  • Bull, C., B., Hagen, L., A., V., Lubin, A., Shivaraman, G., & Chibbaro, D. (2017). Predictable is preventable: Tracking pedestrian near-miss incidents. New Jersey Safe Routes to School Resource Center & Alan M. Voorhees Transportation Center.

  • Calles, M., B., Nelson, T., & Winters, M. (2017). Comparing crowdsourced near-miss and collision cycling data and official bike safety reporting. Transportation Research Record: Journal of the Transportation Research Board, 2662(1), 1-11.

  • Cao, C., Wang, B., Zhang, W., Zeng, X., Yan, X., Feng, Z., Liu, Y., & Wu, Z. (2019). An improved faster R-CNN for small object detection. IEEE Access, 7, 106838-106846.

  • Cepni, S., Atik, M. E., & Druan, Z. (2020). Vehicle detection using different deep learning algorithms from image sequence. Baltic Journal of Modern Computing, 8(2), 347-358.

  • Ciberlin, J., Grbic, R., Teslic, N., & Pilipovic, M. (2019). Object detection and object tracking in front of the vehicle using front view camera. In 2019 Zooming Innovation in Consumer Technologies Conference (ZINC) (pp. 27-32). IEEE Publishing.

  • Ding, X., & Yang, R. (2019). Vehicle and parking space detection based on improved YOLO network model. Journal of Physics: Conference Series, 1325, Article 012084.

  • Dixit, K. S., Chadaga, M. G., Savalgimath, S. S., Rakshith, G. R., & Kumar, M. N. (2019). Evaluation and evolution of object detection techniques YOLO and R-CNN. International Journal of Recent Technology and Engineering (IJRTE), 8(2S3), 824-829. 154 07 0782S319

  • Gad, A. F. (2020). Faster R-CNN explained for object detection tasks. PaperspaceBlog.

  • Girotto, E., Andrade, S., M. D., & Gonzalez, A. D. (2016). Professional experience and traffic accidents/ near-miss accidents among truck drivers. ELSEVIER: Accidents Analysis and Prevention, 95(Pt A), 299-304.

  • Huang, Y. Q., Zheng, J. C., Sun, S. D., Yang, C. F., & Liu, J. (2020). Optimized YOLOv3 algorithm and its application in traffic flow detections. Applied Sciences, 10(9), Article 3079.

  • Johnson, K. D., Patel, S. R., Baur, D. M., Edens, E. Sherry, P. Malhotra, A., & Kales, S., N. (2014). Association of sleep habits with accidents and near misses in United States transportation operators. Journal of Occupational & Environmental Medicine, 56(5), 510-515.

  • Kataoka, H., Suzuki, T., Oikawa, S., Matsui, Y., & Satoh, Y. (2018). Drive video analysis for the detection of traffic near-miss incidents. In IEEE International Conference on Robotics and Automation (ICRA) (pp. 3421-3428). IEEE Publishing.

  • Ke, R., Lutin, J., Spears, J., & Wang, Y. (2017). A cost-effective framework for automated vehicle-pedestrian near-miss detection through onboard monocular vision. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (pp. 25-32). IEEE Publishing.

  • Mahdi, N. N. R. N., Bachok, N., Mohamed, N., & Shafei, M. N. (2014). Risk factors for near miss incident among long distance bus drivers in Malaysia. Iranian Journal of Public Health, 43(3), 117-124.

  • Makizako, H., Shimada, H., Hotta, R., Doi, T., Tsutsumimoto, K., Nakakubo, S., & Makino, K. (2018). Associations of Near-Miss traffic incidents with attention and executive function among older Japanese drivers. Gerontology, 64, 495-502.

  • Matsui, Y., Hitosugi, M., Doi, T., Oikawa, S., Takahashi, K., & Ando, K. (2013). Features of pedestrian behavior in car-to-pedestrian contact situations in Near-Miss incidents in Japan. Traffic Injury Prevention, 14(1), 58-63.

  • Matsui, Y., Takahashi, K., Imaizumi, R., & Ando, K. (2011). Car-to-pedestrian contact situations in near-miss incidents and real-world accidents in Japan. In 22nd International Technical Conference on the Enhanced Safety of Vehicles (No. 110164). National Traffic Safety and Environment Laboratory.

  • Nadai, S. D., Parodi, F., & Pizzorni, D. (2012). A system of systems to Near Miss accidents in dangerous goods road transportation. In IEEE 2012 7th International Conference on System of Systems Engineering (SoSE) (pp. 219-222). IEEE Publishing.

  • Nostikasari, D., & Shelton, K. (2017). Learning from close calls: A glimpse into near-miss experiences. Rice University Kinder Institute of Urban Research.

  • Ong, Y. (2020). Near miss vehicle collisions estimation using YOLO (Master dissertation). Universiti Sains Malaysia, Malaysia.

  • Poulos, R., G., Hatfield, J., Rissel, C., Flack, L., K., Shaw, L., Grzebieta, R., & McIntosh, A., S. (2017). Near miss experiences of transport and recreational cyclists in New South Wales, Australia. Findings from a prospective cohort study. Accidents Analysis and Prevention, 101, 143-153.

  • Qin, Z., Yu, F., Liu, C., & Chen, X. (2018). How convolutional neural networks see the world - A survey of convolutional neural network visualization methods. Mathematical Foundations Computing, 1(2), 149-180.

  • Rahman, A., Salam, A., Islam, M., & Sarker, P. (2012). An image based approach to compute object distance. International Journal of Computational Intelligence Systems, 1(4), 304-312.

  • Rawat, W., & Wang, Z. (2017). Deep convolutional neural networks for image classification: A comprehensive review. Neural Computation, 29(9), 2352-2449.

  • Rome, L. D., Brown, J., Baldock, M., & Fitzharris, M. (2018). Near-miss crashes and other predictors of motorcycle crashes: Findings from a population-based survey. Traffic Injury Prevention, 19(2), S20-S26.

  • Sanders, R., L. (2015). Perceived traffic risk for cyclists: The impact of near miss and collision experiences. Accidents Analysis and Prevention, 75, 26-34.

  • Silva Consultants. (2016). Using near miss reporting in security. Silva Consultants.

  • Siregar, M. L., Agah, H. R., & Hidayatullah, F. (2018). Near-miss accident analysis for traffic safety improvement at a “channelized” junction with U-turn. International Journal of Safety and Security Engineering, 8(1), 31-38.

  • Sonnleitner, E., Barth, O., Palmanshofer, A., & Kurz, M. (2020). Traffic measurement and congestion detection based on real-time highway video data. Applied Sciences, 10(18), Article 6270.

  • Srivastava, S., Divekar, A., V., Anilkumar, C., Naik, I., Kulkarni, V., & Pattabiraman, V. (2021). Comparative analysis of deep learning image detection algorithms. Journal of Big Data volume, 8(66), 1-31.

  • Storgard, J., Erdogan, I., Lappalainen, J., & Tapanien, U. (2012). Developing incident and near miss reporting in the maritime industry - A case study on the Baltic Sea. Procedia Social and Behavioral Sciences, 48, 1010-1021.

  • Uchida, N., Kawakoshi, M., Tagawa, T., & Mochida, T. (2010). An investigation of factors contributing to major crash types in Japan based on naturalistic driving data. International Association of Traffic and Safety Sciences (IATSS), 34(1), 22-30.

  • Vinitha, V., & Velantina, V. (2020). Social distancing detection system with artificial intelligence using computer vision and deep learning. International Research Journal of Engineering and Technology (IRJET), 7(8), 4049-4053.

  • Wang, C., Dai, Y., Zhou, W., & Geng, Y. (2020). A vision-based video crash detection framework for mixed traffic flow environment considering low-visibility condition. Hindawi, Journal of Advanced Transportation, 2020, Article 9194028.

  • WHO. (2015). World Health Statistics 2015. World Health Organization

  • Zhang, Z., Trivedi, C., & Liu, X. (2018). Automated detection of grade-crossing-trespassing near misses based on computer vision analysis of surveillance video data. Safety Science, 110(Part B), 276-285.

  • Zohra, A., F., Kamilia, S., & Souad, S. (2018). Detection and classification of vehicles using deep learning. International Journal of Computer Science Trends and Technology (IJCST), 6(3), 23-29.

ISSN 1511-3701

e-ISSN 2231-8542

Article ID


Download Full Article PDF

Share this article

Recent Articles