Advanced Driver Assistance Systems (ADAS) Based on Machine Learning Techniques for the Detection and Transcription of Variable Message Signs on Roads
Abstract
:1. Introduction
1.1. Motivation
1.2. Vehicle Safety Systems
- Vision systems. These have cameras (monocular, stereo and infrared) placed at strategic points of the vehicle that provide images of the environment from which knowledge of the scene is extracted. These kinds of systems have problems with depth and lens obstructions; however, they are affordable [27];
- LiDAR (light detection and ranging). This is a technology that generates a 3D environment by projecting rays and measuring the distance to different objects. This allows the vehicle to know the elements around it in high resolution. It is a cutting-edge technology, but at the same time expensive. There is currently a debate between LiDAR and conventional cameras. Companies on a par with Tesla bet on the extraction of knowledge through multiple cameras plus other devices, such as radars. Others, for instance Waymo, believe that LiDAR is the solution of the future [28];
- Radars. These systems measure the speed and distance of objects in the environment (thanks to the Doppler effect). They emit a series of microwaves and measure the change in wave frequency. One case of use is adaptive cruise control [27];
- Ultrasound. Using a series of sound waves, these systems measure the distance to nearby objects. An example is the parking collision warning device [27];
- All these ADASs are complemented with other functionalities to improve their accuracy. For example, IMUs (inertial measurement units) or GPSs (global positioning systems) are auxiliary systems for distance measurement [27].
1.3. Recognition Systems
1.3.1. Object Recognition
- DPM (deformable part-based model) [37]. This method is an extension of the HOG detector, which applies the divide and conquer strategy. For example, the problem of recognizing a car can be decomposed into locating parts such as wheels or windows. It consists of a main filter and several secondary filters configured by supervised learning as if they were latent variables [33].
- Faster R-CNN [43]. This model eliminates the bottleneck that fast R-CNN had when selecting the region of interest (RoI) [33] by using a CNN called a region proposal network (RPN) to predict it. Faster R-CNN merges the RPN and fast R-CNN into a single network, so that the first one tells the second one where to focus. This is achieved by sharing their convolutional characteristics. This way, the RoI selection is practically zero cost, and the system is very close to real time [43].
- YOLO (You Only Look Once v1 [44], v2/9000 [45], v3 [46], v4 [47]). This is a real-time object recognition system thanks to the fact that the entire detection process is done by a single network. The process consists of a phase in which the system resizes the image to 488 x 488 and then executes a single CNN that returns the confidence of the detected object [44]. There are several enhancements to this model that are focused on increasing the accuracy but keeping the fast execution. The most recent version is v4 [45,46,47];
- RetinaNet [49]. Thanks to the authors of [49], it was found that the extreme imbalance of the foreground class is the main cause of their lower accuracy. To solve it, they introduced a new loss function called "focal loss" to make the classifier focus on the most difficult examples of the misclassified ones. This brings this model up to the accuracy of the two-stage models.
1.3.2. Text Recognition
- Sliding window (SW). This model works by sliding a small multi-scale window through all possible locations on the image, classifying whether text is present or not [51].
- Detection. Text detection can be defined as a subset of the problem of object detection, in which there are three tendencies [52]:
- Reduction of pipelines to simplify the training process and reduce error. Decomposition into subtexts and then joining them into a complete instance. Specific recognition in cases such as curved text, irregularly shaped text or text with complex backgrounds;
- End-to-end systems. Instead of dividing the main problem into detection and recognition subproblems, these systems integrate the entire process for reading directly from the image [52];
- Auxiliary techniques. An important aspect is techniques that improve training quality, such as creating synthetic examples, reducing noise in the image or incorporating information from the environment [52].
2. Methodology
2.1. VMS Recognizer
2.2. Content Extractor and Speaker
3. Variable Message Sign Recognition
3.1. Dataset
Labeled Image Collection
- Collection. By searching Google Images, YouTube, several websites and manual clippings combined with scraping scripts.
- Labeling. Each image is manually annotated using the software in [63], which generates an XML (Extensible Markup Language) file in PASCAL VOC (Visual Object Classes) format.
- Data augmentation. Data augmentation is a widespread method that consists of applying modifications to the image (rotations, cropping, translations, etc.) in order to create apparently new instances. For this project, since the VMS will always be in the top position of the image, we have chosen to flip the image on the y-axis. That way, the signs on one side will be placed on the opposite side, generating a new instance.
- Manual. As in the first acquisition, the VMS images are manually selected. The difference is that the labeling is performed by the basic model;
- Semiautomatic. In this case, we select videos to be analyzed by the basic model in order to extract a set of labeled candidate images from hours of footage, which would otherwise be much more tedious.
3.2. Final Dataset
3.3. VMS Recognizer
4. Text Extraction
4.1. Preprocessing
4.1.1. Image Straightening
- Edge detection. This task is carried out by the Canny algorithm on a grayscale image, on which a 5 × 5 Gaussian filter has been previously applied to reduce noise (although the Canny algorithm already applies one by default). The parameterization used is inspired by [76]. Thresholds are automatically calculated as follows:
- Obtain the average pixel intensity, ;
- Apply the following formulas with to find the lower and upper thresholds:
- ○
- ○
- Straight line recognition within the image. The Hough transform is applied on the output image of the Canny algorithm, obtaining a list of pairs. The parameters established are:
- ○ Accumulator distance on the axis ;
- ○ Accumulator distance on the axis ;
- ○ Threshold .
- Calculation of the rotation angle, . For each pair , Equation (1) is applied to find the equation of the line in the plane. From it, the slope, , required to transform it into degrees using Equation (2) is obtained and entered into a list. The rotation angle, , is estimated by the arithmetic mean of all the slopes of the detected lines.
- Calculation of the rotation matrix, R. Finally, by applying a rotation matrix, (3), to the original image, the straightened image is obtained. For this, it is necessary to calculate and by means of Equations (4) and (5), knowing that , and is the value obtained in step three.
4.1.2. Image Cropping
- Find the equations of the lines on the image.
- 2.
- Calculate the intersection point with the image limits.
- 3.
- Identify the cutting points and extract the subsection.
- Horizontal cut. Identify the upper, , and lower, , cut-off points of that satisfy:
- Vertical cut. Identify the left, , and right, , cut-off points of that satisfy:
4.1.3. Color Adjustment for OCR
- Binarize the image.
- Convert to grayscale. By applying the formula presented in [77], the gray value is obtained ( being the values of the red, green and blue channels, respectively).
- Reverse the image color. The output of Otsu’s method is an image with white text on a black background. Therefore, it is necessary to apply the NOT logic gate on each value.
- Join discontinuous strokes.
- 3.
- Histogram equalization. Finally, it is necessary to increase the contrast so that the subsequent OCR model will be able to recognize the text. For this purpose, the histogram [69,83] of the image, , has been equalized by mapping it to the normalized cumulative distribution, , q, which is more uniform.
4.2. Recognition and Speech
5. Results and Discussion
5.1. VMS Detector
5.2. Image Preprocessing and Text Extraction
6. Conclusions
Author Contributions
Funding
Conflicts of Interest
Appendix A. Dataset Sources
URL | Duration | URL | Duration |
---|---|---|---|
youtu.be/MFzuxq4V0XI | 00:06:59 | youtu.be/dJcH8YFuvY4 | 00:25:12 |
youtu.be/GREMRp7rvoY | 00:08:16 | youtu.be/M2rvG-e04HE | 00:16:22 |
youtu.be/lNhy2mT94Ao | 00:05:22 | youtu.be/8ifk3BHz1_c | 00:13:35 |
youtu.be/37YgdfidwkA | 00:19:58 | youtu.be/6bkOZcBECsk | 00:12:37 |
youtu.be/JctnDDdoy0A | 00:22:56 | youtu.be/afCyj52txC0 | 00:06:40 |
youtu.be/H1gxWeWsa_E | 01:20:25 | youtu.be/vU82-jnUi_E | 00:07:57 |
youtu.be/UZFLDp_LLj4 | 00:27:23 | youtu.be/D5RHKJNhw7I | 00:08:01 |
youtu.be/tz8bEIirIx4 | 00:10:50 | youtu.be/S1DE3pvnG8s | 00:12:33 |
youtu.be/QJ_XSlOeCBw | 00:10:48 | youtu.be/XLQbclKjNrw | 00:03:36 |
youtu.be/4s-WfvYUbPM | 00:19:57 |
References
- Dargay, J.; Gately, D.; Sommer, M. Vehicle ownership and income growth, Worldwide: 1960–2030. Energy J. 2007, 28, 143–170. [Google Scholar] [CrossRef] [Green Version]
- Dirección General de Tráfico—Ministerio del Interior. Anuario Estadístico General. 2018. Available online: http://www.dgt.es/es/seguridad-vial/estadisticas-e-indicadores/publicaciones/anuario-estadistico-general/ (accessed on 7 May 2021).
- Instituto Nacional de Estadística. Encuesta Continua de Hogares. 2018. Available online: https://www.ine.es/prensa/ech_2018.pdf (accessed on 7 May 2021).
- Organización Mundial de la Salud. Decade of Action for Road Safety 2011–2020. 2011. Available online: https://www.who.int/publications/i/item/decade-of-action-for-road-safety-2011-2020 (accessed on 7 May 2021).
- Dirección General de Tráfico—Ministerio del Interior. Las Principales Cifras de la Siniestralidad Vial. 2018. Available online: http://www.dgt.es/es/seguridad-vial/estadisticas-e-indicadores/publicaciones/principales-cifras-siniestralidad/ (accessed on 7 May 2021).
- Dirección General de Tráfico—Ministerio del Interior. Anuario Estadístico de Accidentes. 2018. Available online: http://www.dgt.es/es/seguridad-vial/estadisticas-e-indicadores/publicaciones/anuario-estadistico-accidentes/ (accessed on 7 May 2021).
- Dirección General de Tráfico—Ministerio del Interior. Las distracciones Causan Uno de Cada Tres Accidentes Mortales. 2018. Available online: http://www.dgt.es/es/prensa/notas-de-prensa/2018/20180917_campana_distracciones.shtml (accessed on 7 May 2021).
- Smiley, A.; Brookhuis, K.A. Alcohol, drugs and traffic safety. In Road Users and Traffic Safety; Transportation Research Board: Washington, DC, USA, 1987; pp. 83–104. [Google Scholar]
- Dirección General de Tráfico—Ministerio del Interior. Distracciones al Volante. Available online: http://www.dgt.es/PEVI/documentos/catalogo_recursos/didacticos/did_adultas/Distracciones_al_volante.pdf (accessed on 7 May 2021).
- Dirección General de Tráfico—Ministerio del Interior. Las Distracciones Son la Causa de Uno de Cada Cuatro Accidentes. 2019. Available online: http://www.dgt.es/es/prensa/notas-de-prensa/2019/Las-distracciones-son-la-causa-de-uno-de-cada-cuatro-accidentes.shtml (accessed on 7 May 2021).
- Billington, J. The Prometheus Project: The Story behind One of AV’s Greatest Developments. 2018. Available online: https://www.autonomousvehicleinternational.com/features/the-prometheus-project.html (accessed on 8 May 2021).
- Brookhuis, K.A.; de Waard, D.; Janssen, W.H. Behavioural impacts of advanced driver assistance systems—An overview. Eur. J. Transp. Infrastruct. Res. 2001, 1, 246–253. [Google Scholar]
- Michon, J.A. Generic Intelligent Driver Support: A Comprehensive Report on GIDS; CRC Press: Boca Raton, FL, USA, 1993; pp. 3–18. [Google Scholar]
- BCG. A Roadmap to Safer Driving through Advanced Driver Assistance Systems. 2015. Available online: https://image-src.bcg.com/Images/MEMA-BCG-A-Roadmap-to-Safer-Driving-Sep-2015_tcm9-63787.pdf (accessed on 8 May 2021).
- Nygårdhs, S.; Helmers, G. VMS—Variable Message Signs: A Literature Review; Transportation Research Board: Washington, DC, USA, 2007. [Google Scholar]
- Autopistas.com. Paneles de Mensajería Variable. Available online: https://www.autopistas.com/blog/paneles-de-mensajeria-variable/ (accessed on 8 May 2021).
- Kolisetty, V.G.B.; Iryo, T.; Asakura, Y.; Kuroda, K. Effect of variable message signs on driver speed behavior on a section of expressway under adverse fog conditions—A driving simulator approach. J. Adv. Transp. 2006, 40, 47–74. [Google Scholar] [CrossRef]
- Peeta, S.; Ramos, J.L. Driver response to variable message signs-based traffic information. IEEE Proc. Intell. Transp. Syst. 2006, 153, 2–10. [Google Scholar] [CrossRef] [Green Version]
- Guattari, C.; De Blasiis, M.R.; Calvi, A. The effectiveness of variable message signs information: A driving simulation study. Procedia Soc. Behav. Sci. 2012, 53, 692–702. [Google Scholar] [CrossRef]
- Gopher, G. Attentional Allocation in Dual Task Environments, Attention and Performance III; Elsevier: Amsterdam, The Netherlands, 1990. [Google Scholar]
- Roca, J.; Insa, B.; Tejero, P. Legibility of text and pictograms in variable message signs: Can single word messages outperform pictograms? J. Hum. Factors Ergon. Soc. 2018, 60, 384–396. [Google Scholar] [CrossRef]
- Simlinger, P.; Egger, S.; Galinski, C. Proposal on Unified Pictograms, Keywords, Bilingual Verbal Messages and Typefaces for VMS in the TERN, International Institute for Information Design. 2008. Available online: https://ec.europa.eu/transport/road_safety/sites/roadsafety/files/pdf/projects_sources/in-safety_d2_3.pdf (accessed on 8 May 2021).
- Universitat de València. READit VMS. 2018. Available online: https://www.uv.es/uvweb/estructura-investigacion-interdisciplinar-lectura/es/productos-tecnologicos/productos-tecnologicos/readit-vms-1286067296453.html (accessed on 8 May 2021).
- Erke, A.; Sagberg, F.; Hagman, R. Effects of route guidance variable message signs (VMS) on driver behaviour. Transp. Res. Part F Traffic Psychol. Behav. 2007, 10, 447–457. [Google Scholar] [CrossRef]
- IBM. Watson Speech to Text. Available online: https://www.ibm.com/es-es/cloud/watson-speech-to-text (accessed on 11 July 2021).
- Dirección General de Tráfico—Ministerio del Interior. Cuestiones De Seguridad Vial. 2018. Available online: http://www.dgt.es/Galerias/seguridad-vial/formacion-vial/cursos-para-profesores-y-directores-de-autoescuelas/XXI-Cuso-Profesores/Manual-II-Cuestiones-de-Seguridad-Vial-2018.pdf (accessed on 7 May 2021).
- Kukkala, V.K.; Tunnell, J.; Pasricha, S.; Bradley, T. Advanced driver-assistance systems: A path toward autonomous vehicles. IEEE Consum. Electron. Mag. 2018, 7, 18–25. [Google Scholar] [CrossRef]
- O’Kane, S. How Tesla and Waymo Are Tackling a Major Problem for Self-Driving Cars: Data, The Verge. 2018. Available online: https://www.theverge.com/transportation/2018/4/19/17204044/tesla-waymo-self-driving-car-data-simulation (accessed on 8 May 2021).
- Bay, O. ABI Research Forecasts 8 Million Vehicles to Ship with SAE Level 3, 4 and 5 Autonomous Technology in 2025. 2018. Available online: https://www.abiresearch.com/press/abi-research-forecasts-8-million-vehicles-ship-sae-level-3-4-and-5-autonomous-technology-2025/ (accessed on 8 May 2021).
- Stoma, M.; Dudziak, A.; Caban, J.; Droździel, P. The future of autonomous vehicles in the opinion of automotive market users. Energies 2021, 14, 4777. [Google Scholar] [CrossRef]
- Silva, D.; Csiszar, C.; Földes, D. Autonomous vehicles and urban space management. Sci. J. Sil. Univ. Technol. Ser. Transp. 2020, 110, 13. [Google Scholar]
- Synopsys. The 6 Levels of Vehicle Autonomy Explained. Available online: https://www.synopsys.com/automotive/autonomous-driving-levels.html (accessed on 8 May 2021).
- Zou, Z.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. arXiv 2019, arXiv:1905.05055v2. [Google Scholar]
- Viola, P.; Jones, M.J. Robust real-time face detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
- Viola, P.; Jones, M.J. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference, Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
- Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
- Felzenszwalb, P.; McAllester, D.; Ramanan, D. A discriminatively trained, multiscale, deformable part mode, computer vision and pattern recognition. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
- Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; Lecun, Y. Over Feat: Integrated Recognition, Localization and Detection using Convolutional Networks. arXiv 2013, arXiv:1312.6229. [Google Scholar]
- Soviany, P.; Ionescu, R.T. Optimizing the trade-off between single-stage and two-stage deep object detectors using image difficulty prediction. In Proceedings of the 2018 20th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing, Timisoara, Romania, 20–23 September 2018. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. arXiv 2017, arXiv:1612.08242v1. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
- Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 318–327. [Google Scholar] [CrossRef] [Green Version]
- Hui, J. Object Detection: Speed and Accuracy Comparison (Faster R-CNN, R-FCN, SSD, FPN, RetinaNet and YOLOv3). Available online: https://jonathan-hui.medium.com/object-detection-speed-and-accuracy-comparison-faster-r-cnn-r-fcn-ssd-and-yolo-5425656ae359 (accessed on 25 May 2021).
- Lin, H.; Yang, P.; Zhang, F. Review of scene text detection and recognition. Arch. Comput. Methods Eng. 2020, 27, 433–454. [Google Scholar] [CrossRef]
- Long, S.; He, X.; Yao, C. Scene text detection and recognition: The deep learning era. Int. J. Comput. Vis. 2020, 129, 161–184. [Google Scholar] [CrossRef]
- Zhu, Y.; Yao, C.; Bai, X. Scene text detection and recognition: Recent advances and future trends. Front. Comput. Sci. 2015, 10, 19–36. [Google Scholar] [CrossRef]
- Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, New York, NY, USA, 25–29 June 2006; Volume 2006, pp. 369–376. [Google Scholar]
- Charette, R.; Nashashibi, F. Real time visual trafric lights recognition based on spot light detection and adaptive traffic lights templates. In Proceedings of the 2009 IEEE Intelligent Vehicles Symposium, Xi’an, China, 3–5 June 2009; pp. 358–363. [Google Scholar]
- Fairfield, N.; Urmson, C. Traffic light mapping and detection. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 5421–5426. [Google Scholar]
- Chung, Y.-C.; Wang, J.-M.; Chen, S.-W. A vision-based traffic light detection system at intersections. J. Taiwan Norm. Univ. Math. Sci. Technol. 2002, 47, 67–86. [Google Scholar]
- Lu, Y.; Lu, J.; Zhang, S.; Hall, P. Traffic signal detection and classification in street views using an attention model. Comput. Vis. Media 2018, 4, 253–266. [Google Scholar] [CrossRef] [Green Version]
- Zhu, Z.; Liang, D.; Zhang, S.; Huang, X.; Li, B.; Hu, S. Traffic-sign detection and classification in the wild. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2110–2118. [Google Scholar]
- Toyota Motor Sales. Toyota Safety Sense™ Comes Standard on Many New Toyotas. Available online: https://www.toyota.com/safety-sense/animation/drcc (accessed on 8 May 2021).
- González, A.; Bergasa, L.M.; Yebes, J.J. Text detection and recognition on traffic panels from street-level imagery using visual appearance. IEEE Trans. Intell. Transp. Syst. 2014, 15, 228–238. [Google Scholar] [CrossRef]
- Vazquez-Reina, A.; Sastre, R.; Arroyo, S.; Gil-Jiménez, P. Adaptive traffic road sign panels text extraction. In Proceedings of the 5th WSEAS International Conference on Signal Processing, Robotics and Automation, Madrid, Spain, 15–17 February 2006; pp. 295–300. [Google Scholar]
- Tzutalin, LabelImg, Github. Available online: https://github.com/tzutalin/labelImg (accessed on 14 May 2021).
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C. Microsoft coco: Common objects in context. arXiv 2014, arXiv:1405.0312v3. [Google Scholar]
- Fizyr, Keras-Retinanet, Github. Available online: https://github.com/fizyr (accessed on 14 May 2021).
- Google. TensorFlow. Available online: https://www.tensorflow.org/tutorials/quickstart/beginner (accessed on 14 May 2021).
- Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef]
- Gonzalez, R.C.; Woods, R.E. Digital Image Processing; Prentice Hall: Hoboken, NJ, USA, 2008. [Google Scholar]
- Rebaza, J.V. Detección de Bordes Mediante el Algoritmo de Canny. Master’s Thesis, Universidad Nacional de Trujillo, Trujillo, Peru, 2007. [Google Scholar]
- OpenCV. Open Source Computer Vision, Canny Edge Detector. Available online: https://docs.opencv.org/master/da/d5c/tutorial_canny_detector.html (accessed on 30 May 2021).
- OpenCV. Open Source Computer Vision, Hough Line Transform. Available online: https://docs.opencv.org/master/d9/db0/tutorial_hough_lines.html (accessed on 30 May 2021).
- Shehata, A.; Mohammad, S.; Abdallah, M.; Ragab, M. A survey on Hough transform, theory, techniques and applications. arXiv 2015, arXiv:1502.02160. [Google Scholar]
- Hough, P.V.C. Methods and Means for Recognizing Complex Patterns. U.S. Patent US3069654A, 18 December 1962. [Google Scholar]
- Ballard, D.H.; Brown, C.M. Computer Vision; Prentice Hal: Englewood Cliffs, NJ, USA, 1982. [Google Scholar]
- Rosebrock, A. Zero-Parameter, Automatic Canny Edge Detection with Python and OpenCV. 2015. Available online: https://www.pyimagesearch.com/2015/04/06/zero-parameter-automatic-canny-edge-detection-with-python-and-opencv/ (accessed on 30 May 2021).
- OpenCV. Open Source Computer Vision, Color Conversions. Available online: https://docs.opencv.org/3.4/de/d25/imgproc_color_conversions.html (accessed on 30 May 2021).
- Otsu, N. Fronts propagating with curvature-dependent speed: Algorithms based on Hamilton-Jacobi formulations. J. Comput. Phys. 1988, 79, 12–49. [Google Scholar]
- OpenCV. Open Source Computer Vision, Image Thresholding. Available online: https://docs.opencv.org/master/d7/d4d/tutorial_py_thresholding.html (accessed on 30 May 2021).
- Morse, S. Lecture 4: Thresholding (Brigham Young University). Available online: http://homepages.inf.ed.ac.uk/rbf/CVonline/LOCAL_COPIES/MORSE/threshold.pdf (accessed on 30 May 2021).
- OpenCV. Open Source Computer Vision, Morphological Transformations. Available online: https://docs.opencv.org/trunk/d9/d61/tutorial_py_morphological_ops.html (accessed on 30 May 2021).
- Sreedhar, K.; Panlal, B. Enhancement of images using morphological transformations. Int. J. Comput. Sci. Inf. Technol. 2012, 4, 33–50. [Google Scholar] [CrossRef]
- OpenCV. Open Source Computer Vision, Histogram Equalization. Available online: https://docs.opencv.org/3.4/d4/d1b/tutorial_histogram_equalization.html (accessed on 30 May 2021).
- Smith, R. An overview of the tesseract OCR engine. In Proceedings of the Ninth International Conference on Document Analysis and Recognition (ICDAR 2007), Curitiba, Brazil, 23–26 September 2007. [Google Scholar]
- Tesseract-Ocr, Tesseract, GitHub. Available online: http://code.google.com/p/tesseract-ocr (accessed on 3 June 2021).
- Kumar, A. Performing OCR by Running Parallel Instances of Tesseract 4.0: Python. 2018. Available online: https://appliedmachinelearning.blog/2018/06/30/performing-ocr-by-running-parallel-instances-of-tesseract-4-0-python/ (accessed on 3 June 2021).
Epochs | 25 |
No. of training images | 134 |
Time | ≈01:30:00 |
Learning rate | 10−5 |
Loss | 0.174 |
Processor | Intel i7 9800K 3.6 GHz |
RAM | 32 GBs |
Graphics card | Nvidia RTX 2080 Ti |
Hard disk | 1 Tb SSD M2 |
Epochs | 16 | Best epoch | 7 |
Loss | 0.008 | Loss (epoch 7) | 0.024 |
lr | 10−5 | AP (epoch 7) | 0.703 |
IoU | 0.5 | Time | 01:20:00 |
Epochs | 14 | Best epoch | 7 |
Loss | 0.009 | Loss (epoch 7) | 0.024 |
lr | 10−7 | AP (epoch 7) | 0.703 |
IoU | 0.5 | Time | 01:15:00 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
De-Las-Heras, G.; Sánchez-Soriano, J.; Puertas, E. Advanced Driver Assistance Systems (ADAS) Based on Machine Learning Techniques for the Detection and Transcription of Variable Message Signs on Roads. Sensors 2021, 21, 5866. https://doi.org/10.3390/s21175866
De-Las-Heras G, Sánchez-Soriano J, Puertas E. Advanced Driver Assistance Systems (ADAS) Based on Machine Learning Techniques for the Detection and Transcription of Variable Message Signs on Roads. Sensors. 2021; 21(17):5866. https://doi.org/10.3390/s21175866
Chicago/Turabian StyleDe-Las-Heras, Gonzalo, Javier Sánchez-Soriano, and Enrique Puertas. 2021. "Advanced Driver Assistance Systems (ADAS) Based on Machine Learning Techniques for the Detection and Transcription of Variable Message Signs on Roads" Sensors 21, no. 17: 5866. https://doi.org/10.3390/s21175866
APA StyleDe-Las-Heras, G., Sánchez-Soriano, J., & Puertas, E. (2021). Advanced Driver Assistance Systems (ADAS) Based on Machine Learning Techniques for the Detection and Transcription of Variable Message Signs on Roads. Sensors, 21(17), 5866. https://doi.org/10.3390/s21175866