Recognizing road surface traffic signs based on Yolo models considering image flips

Christine Dewi*, Rung Ching Chen, Yong Cun Zhuang, Xiaoyi Jiang, Hui Yu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

61 Downloads (Pure)

Abstract

In recent years, there have been significant advances in deep learning and road marking recognition due to machine learning and artificial intelligence. Despite significant progress, it often relies heavily on unrepresentative datasets and limited situations. Drivers and advanced driver assistance systems rely on road markings to help them better understand their environment on the street. Road markings are signs and texts painted on the road surface, including directional arrows, pedestrian crossings, speed limit signs, zebra crossings, and other equivalent signs and texts. Pavement markings are also known as road markings. Our experiments briefly discuss convolutional neural network (CNN)-based object detection algorithms, specifically for Yolo V2, Yolo V3, Yolo V4, and Yolo V4-tiny. In our experiments, we built the Taiwan Road Marking Sign Dataset (TRMSD) and made it a public dataset so other researchers could use it. Further, we train the model to distinguish left and right objects into separate classes. Furthermore, Yolo V4 and Yolo V4-tiny results can benefit from the “No Flip” setting. In our case, we want the model to distinguish left and right objects into separate classes. The best model in the experiment is Yolo V4 (No Flip), with a test accuracy of 95.43% and an IoU of 66.12%. In this study, Yolo V4 (without flipping) outperforms state-of-the-art schemes, achieving 81.22% training accuracy and 95.34% testing accuracy on the TRMSD dataset.

Original languageEnglish
Article number54
Number of pages19
JournalBig Data and Cognitive Computing
Volume7
Issue number1
DOIs
Publication statusPublished - 22 Mar 2023

Keywords

  • CNN
  • machine learning
  • object recognition
  • road sign
  • Yolo

Cite this