Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network
<p>Overall flowchart of proposed method.</p> "> Figure 2
<p>Convolutional neural network (CNN) architecture.</p> "> Figure 3
<p>Example images from Road marking dataset.</p> "> Figure 4
<p>Example images from Karlsruhe institute of technology and Toyota technological institute at Chicago (KITTI) dataset.</p> "> Figure 5
<p>Example images from Málaga dataset 2009.</p> "> Figure 6
<p>Example images from Málaga urban dataset.</p> "> Figure 7
<p>Example images from Naver street view dataset.</p> "> Figure 8
<p>Example images from Road/Lane detection evaluation 2013 dataset.</p> "> Figure 9
<p>Obtaining the inverse perspective mapping (IPM) image and arrow markings: (<b>a</b>) original image; (<b>b</b>) IPM transformation; and (<b>c</b>) the obtained arrow markings.</p> "> Figure 10
<p>Examples of arrow markings after size normalization and bi-linear interpolation.</p> "> Figure 11
<p>Classification accuracies of training data over 13-fold cross validation; “classifications” means “classification accuracy of training data”: (<b>a</b>) Training 1–3; (<b>b</b>) Training 4–6; (<b>c</b>) Training 7–9; (<b>d</b>) Training 10–13.</p> "> Figure 11 Cont.
<p>Classification accuracies of training data over 13-fold cross validation; “classifications” means “classification accuracy of training data”: (<b>a</b>) Training 1–3; (<b>b</b>) Training 4–6; (<b>c</b>) Training 7–9; (<b>d</b>) Training 10–13.</p> "> Figure 12
<p>The obtained filters from the 1st convolution layer through training: (<b>a</b>–<b>m</b>) the filters from the 1st–13th trainings among 13-fold cross validation are presented, respectively.</p> "> Figure 13
<p>Examples of correct recognition cases: (<b>a</b>) forward arrow (FA); (<b>b</b>–<b>d</b>) forward-left arrow (FLA); (<b>e</b>) forward-right arrow (FRA); (<b>f</b>) left arrow (LA); (<b>g</b>) right arrow (RA).</p> "> Figure 14
<p>Examples of incorrect recognition cases: (<b>a</b>) FA is incorrectly recognized into FLA; and (<b>b</b>,<b>c</b>) LA is incorrectly recognized into RA.</p> ">
Abstract
:1. Introduction
2. Related Works
- -
- We propose a CNN-based method to recognize painted arrow-road markings. This method is new as it is not reported in the state of the art. Our method results in high accuracy of recognition and it is robust to the image quality of arrow-road marking.
- -
- Our method is capable of recognizing severely damaged arrow-road markings. It also demonstrates good recognition accuracy in a variety of lighting conditions, such as shadowed, dark and dim arrow-road marking images that are not easily recognized.
- -
- We used six datasets (Road marking dataset, KITTI dataset, Málaga dataset 2009, Málaga urban dataset, Naver street view dataset, and Road/Lane detection evaluation 2013 dataset) for CNN training and testing. These datasets were obtained from different countries, each with a diverse environment. The arrow-road markings of each dataset have different sizes and different image qualities. Through the intensive training of a CNN using these datasets, our method demonstrates robust performance that is independent of the nature of the datasets.
3. Proposed Method for the Recognition of Arrow-Road Markings
3.1. Overall Flowchart of Proposed Method
3.2. Architecture of the CNN
3.3. Feature Extraction by Three Convolutional Layers
3.4. Classification by Four Fully Connected Layers
4. Experimental Results
4.1. Experimental Data and Environment
4.2. Training
4.3. Testing: Measuring the Accuracies of Arrow-Road Marking Recognition
5. Conclusions
Acknowledgments
Author Contributions
Conflicts of Interest
References
- Schreiber, M.; Poggenhans, F.; Stiller, C. Detecting Symbols on Road Surface for Mapping and Localization Using OCR. In Proceedings of the IEEE International Conference on Intelligent Transportation Systems, Qingdao, China, 8–11 October 2014; pp. 597–602.
- Suchitra, S.; Satzoda, R.K.; Srikanthan, T. Detection & Classification of Arrow Markings on Roads Using Signed Edge Signatures. In Proceedings of the Intelligent Vehicles Symposium, Alcalá de Henares, Spain, 3–7 June 2012; pp. 796–801.
- Qin, B.; Liu, W.; Shen, X.; Chong, Z.J.; Bandyopadhyay, T.; Ang, M.H., Jr.; Frazzoli, E.; Rus, D. A General Framework for Road Marking Detection and Analysis. In Proceedings of the 16th International IEEE Annual Conference on Intelligent Transportation Systems, The Hague, The Netherlands, 6–9 October 2013; pp. 619–625.
- Danescu, R.; Nedevschi, S. Detection and Classification of Painted Road Objects for Intersection Assistance Applications. In Proceedings of the 13th International IEEE Conference on Intelligent Transportation Systems, Madeira Island, Portugal, 19–22 September 2010; pp. 433–438.
- Foucher, P.; Sebsadji, Y.; Tarel, J.-P.; Charbonnier, P.; Nicolle, P. Detection and Recognition of Urban Road Markings Using Images. In Proceedings of the 4th International IEEE Conference on Intelligent Transportation Systems, Washington, DC, USA, 5–7 October 2011; pp. 1747–1752.
- Wu, T.; Ranganathan, A. A Practical System for Road Marking Detection and Recognition. In Proceedings of the Intelligent Vehicles Symposium, Alcalá de Henares, Spain, 3–7 June 2012; pp. 25–30.
- Liu, Z.; Wang, S.; Ding, X. ROI Perspective Transform Based Road Marking Detection and Recognition. In Proceedings of the International Conference on Audio, Language and Image Processing, Shanghai, China, 16–18 July 2012; pp. 841–846.
- Ding, D.; Yoo, J.; Jung, J.; Jin, S.; Kwon, S. Efficient Road-Sign Detection Based on Machine Learning. Bull. Netw. Comput. Syst. Softw. 2015, 4, 15–17. [Google Scholar]
- Greenhalgh, J.; Mirmehdi, M. Detection and Recognition of Painted Road Surface Markings. In Proceedings of the 4th International Conference on Pattern Recognition Applications and Methods, Lisbon, Portugal, 10–12 January 2015; pp. 1–9.
- Smith, R. An Overview of the Tesseract OCR Engine. In Proceedings of the 9th International Conference on Document Analysis and Recognition, Curitiba, Brazil, 23–26 September 2007; pp. 629–633.
- Wu, W.; Chen, X.; Yang, J. Detection of Text on Road Signs from Video. IEEE Trans. Intell. Transp. Syst. 2005, 6, 378–390. [Google Scholar] [CrossRef]
- Rebut, J.; Bensrhair, A.; Toulminet, G. Image Segmentation and Pattern Recognition for Road Marking. In Proceedings of the IEEE International Symposium on Industrial Electronics, Palais des Congrès & Expositions Ajaccio, France, 4–7 May 2004; pp. 727–732.
- Yamamoto, J.; Karungaru, S.; Terada, K. Japanese Road Signs Recognition Using Neural Networks. In Proceedings of the SICE Annual Conference, Nagoya, Japan, 14–17 September 2013; pp. 1144–1150.
- Rahman, M.O.; Mousumi, F.A.; Scavino, E.; Hussain, A.; Basri, H. Real Time Road Sign Recognition System Using Artificial Neural Networks for Bengali Textual Information Box. In Proceedings of the International Symposium on Information Technology, Kuala Lumpur, Malaysia, 26–29 August 2008; pp. 1–8.
- Baghdassarian, C.; Lange, H.; Sahli, H.; Laurgeau, C. Recognition of Arrows in the Environment of Road Markings. In Proceedings of the IEEE Intelligent Vehicles Symposium, Paris, France, 24–26 October 1994; pp. 219–224.
- Yamamoto, J.; Karungaru, S.; Terada, K. Road Surface Marking Recognition Using Neural Network. In Proceedings of the IEEE/SICE International Symposium on System Integration, Tokyo, Japan, 13–15 December 2014; pp. 484–489.
- Kheyrollahi, A.; Breckon, T.P. Automatic Real-time Road Marking Recognition Using a Feature Driven Approach. Mach. Vis. Appl. 2012, 23, 123–133. [Google Scholar] [CrossRef]
- Chen, T.; Chen, Z.; Shi, Q.; Huang, X. Road Marking Detection and Classification Using Machine Learning Algorithms. In Proceedings of the IEEE Intelligent Vehicles Symposium, Seoul, Korea, 28 June–1 July 2015; pp. 617–621.
- Suhr, J.K.; Jung, H.G. Fast Symbolic Road Marking and Stop-line Detection for Vehicle Localization. In Proceedings of the IEEE Intelligent Vehicles Symposium, Seoul, Korea, 28 June–1 July 2015; pp. 186–191.
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- Zhu, Z.; Liang, D.; Zhang, S.; Huang, X.; Li, B.; Hu, S. Traffic-Sign Detection and Classification in the Wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2110–2118.
- Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep Neural Network for Structural Prediction and Lane Detection in Traffic Scene. IEEE Trans. Neural Netw. Learn. Syst. 2016, 1–14. [Google Scholar] [CrossRef]
- Gurghian, A.; Koduri, T.; Bailur, S.V.; Carey, K.J.; Murali, V.N. DeepLanes: End-To-End Lane Position Estimation Using Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 38–45.
- Sukhwani, M.; Singh, S.; Goyal, A.; Belh, A.; Mohapatra, P.; Bharti, B.K.; Jawahar, C.V. Monocular Vision Based Road Marking Recognition for Driver Assistance and Safety. In Proceedings of the IEEE International Conference on Vehicular Electronics and Safety, Hyderabad, India, 16–17 December 2014; pp. 11–16.
- Convolutional Neural Network. Available online: https://en.wikipedia.org/wiki/Convolutional_neural_network (accessed on 7 October 2016).
- CS231n Convolutional Neural Networks for Visual Recognition. Available online: http://cs231n.github.io/convolutional-networks/#overview (accessed on 17 October 2016).
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25: 26th Annual Conference on Neural Information Processing Systems 2012; Curran Associates, Inc.: Red Hook, NY, USA, 2012. [Google Scholar]
- Road Marking Dataset. Available online: http://www.ananth.in/RoadMarkingDetection.html (accessed on 7 October 2016).
- The KITTI Dataset. Available online: http://www.cvlibs.net/datasets/kitti/raw_data.php (accessed on 7 October 2016).
- Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets Robotics: The KITTI Dataset. Int. J. Robot. Res. 2013, 32, 1229–1235. [Google Scholar] [CrossRef]
- Málaga Dataset 2009. Available online: http://www.mrpt.org/robotics_datasets (accessed on 7 October 2016).
- Málaga Urban Dataset. Available online: http://www.mrpt.org/MalagaUrbanDataset (accessed on 7 October 2016).
- Naver Street View. Available online: http://maps.naver.com/ (accessed on 7 October 2016).
- Road/Lane Detection Evaluation 2013 Dataset (Base Kit with: Left Color Images, Calibration and Training Labels). Available online: http://www.cvlibs.net/datasets/kitti/eval_road.php (accessed on 20 October 2016).
- Muad, A.M.; Hussain, A.; Samad, S.A.; Mustaffa, M.M.; Majlis, B.Y. Implementation of Inverse Perspective Mapping Algorithm for the Development of an Automatic Lane Tracking System. In Proceedings of the TENCON IEEE Region 10 Conference, Chiang Mai, Thailand, 21–24 November 2004; pp. 207–210.
- Intel® Core™ i7-6700 Processor. Available online: http://ark.intel.com/products/88196/Intel-Core-i7-6700-Processor-8M-Cache-up-to-4_00-GHz (accessed on 14 October 2016).
- GeForce GTX TITAN X. Available online: http://www.geforce.com/hardware/desktop-gpus/geforce-gtx-titan-x/specifications (accessed on 14 October 2016).
- Matlab Toolbox for Convolutional Neural Network. Available online: http://www.mathworks.com/help/nnet/convolutional-neural-networks.html (accessed on 14 October 2016).
- Precision and Recall. Available online: https://en.wikipedia.org/wiki/Precision_and_recall (accessed on 14 October 2016).
- He, Y.; Chen, S.; Pan, Y.; Ni, K. Using Edit Distance and Junction Feature to Detect and Recognize Arrow Road Marking. In Proceedings of the IEEE International Conference on Intelligent Transportation Systems, Qingdao, China, 8–11 October 2014; pp. 2317–2323.
Category | Methods | Advantages | Disadvantages | Performance | Year | Ref. |
---|---|---|---|---|---|---|
Non-learning-based | Geometric parameter optimization | Fast processing speed | The number of thresholds to be set is large | True positive rate (TPR) of 90% and 78% for crosswalks and arrows, respectively | 2011 | [5] |
HOG features and template matching | A good ability to cope with the cases of limited occlusions and the variations of lighting condition |
| TPR of 90.1% and false positive rate (FPR) 0.9% | 2012 | [6] | |
Template matching | Fast processing speed | Damaged road marking can cause misclassification, and the classification accuracy is affected by the illumination variation | Detection rate of 95.8% and 84% on the highway and city roads, respectively | 2012 | [7] | |
Learning-based | HOG features and total error rate (TER)-based classifier | Fast computing time compared to SVM-based method | Damaged or shadowed markings increase FPR | Overall classification accuracy of 99.2% | 2015 | [19] |
HOG features and SVM | Showing high accuracy with the trained datasets | Recognition accuracy can be affected by damaged or shadowed markings | Quantitative accuracies were not reported | 2015 | [8] | |
F-measure of 0.91 | 2015 | [9] | ||||
Average accuracy of 91.7% | 2014 | [24] | ||||
Fourier descriptor and KNN classifier | Robust to noises on road marking | Sensitive to occlusion, dirty markings or poor visibility | Average error of 6% | 2004 | [12] | |
Artificial Neural Network | Higher accuracy with trained datasets compared to non-learning-based method | Performance of testing data can be affected by trained dataset | Quantitative accuracies were not reported | 1994 | [15] | |
The average accuracy of white markings is about 71.5%, and for orange markings it was about 46% | 2014 | [16] | ||||
Accuracy of 85% for arrows | 2012 | [17] | ||||
BING, PCA network, and SVM classifier | The area of road marking can be detected by BING method without lane detection | Performance of testing data can be affected by trained dataset | Accuracy of 96.8% | 2015 | [18] | |
Proposed method (CNN) | Arrow-road markings in various environments including damaged ones can be correctly recognized independent of the kinds of datasets by intensive training of CNN | Time consuming procedure for training is required for CNN | Average accuracy and F_score are 99.88% and 99.94%, respectively |
Layer Type | Number of Filters | Size of Feature Map | Size of Kernel | Number of Stride |
---|---|---|---|---|
Image input layer | 265 (height) × 137 (width) × 1 (channel) | |||
1st convolutional layer | 180 | 131 × 67 × 180 | [5 5] | [2 2] |
ReLU layer | ||||
CCN layer | ||||
Max pooling layer | 180 | 65 × 33 × 180 | [3 3] | [2 2] |
2nd convolutional layer | 250 | 31 × 15 × 250 | [5 5] | [2 2] |
ReLU layer | ||||
CCN layer | ||||
Max pooling layer | 250 | 15 × 7 × 250 | [3 3] | [2 2] |
3rd convolutional layer | 250 | 7 × 3 × 250 | [3 3] | [2 2] |
ReLU layer | ||||
CCN layer | ||||
Max pooling layer | 250 | 3 × 1 × 250 | [3 3] | [2 2] |
1st fully connected layer | 1920 | |||
ReLu layer | ||||
2nd fully connected layer | 1024 | |||
ReLu layer | ||||
3rd fully connected layer | 512 | |||
ReLu layer | ||||
Dropout layer | ||||
4th fully connected layer | 6 | |||
Softmax layer | ||||
Classification layer (output layer) |
FA | FLA | FLRA | FRA | LA | RA | Total | |
---|---|---|---|---|---|---|---|
Number of data | 32,686 | 17,885 | 22,344 | 17,885 | 36,766 | 36,243 | 163,809 |
Total of Testing 1–13 | Recognized Arrows | ||||||
---|---|---|---|---|---|---|---|
FA | FLA | FLRA | FRA | LA | RA | ||
Actual arrows | FA | 31,447 | 1 | 0 | 1 | 0 | 2 |
FLA | 0 | 18,101 | 0 | 1 | 0 | 0 | |
FLRA | 21 | 0 | 22,254 | 0 | 0 | 0 | |
FRA | 0 | 0 | 0 | 18,824 | 24 | 0 | |
LA | 1 | 0 | 0 | 0 | 33,334 | 210 | |
RA | 0 | 0 | 0 | 0 | 0 | 31,779 |
# of Testing | FA | FLA | FLRA | FRA | LA | RA | |
---|---|---|---|---|---|---|---|
Testing 1 | ACC | 100 | 100 | 100 | 100 | 99.45 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
Testing 2 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
Testing 3 | ACC | 99.91 | 100 | 100 | 100 | 99.05 | 100 |
F_score | 99.96 | 100 | 100 | 100 | 99.52 | 100 | |
Testing 4 | ACC | 100 | 100 | 100 | 100 | 99.02 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.51 | 100 | |
Testing 5 | ACC | 100 | 100 | 100 | 100 | 99.12 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.56 | 100 | |
Testing 6 | ACC | 100 | 100 | 100 | 100 | 99.41 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.70 | 100 | |
Testing 7 | ACC | 99.96 | 100 | 100 | 100 | 99.34 | 100 |
F_score | 99.98 | 100 | 100 | 100 | 99.67 | 100 | |
Testing 8 | ACC | 100 | 100 | 100 | 99.11 | 100 | 100 |
F_score | 100 | 100 | 100 | 99.55 | 100 | 100 | |
Testing 9 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
Testing 10 | ACC | 99.96 | 100 | 100 | 100 | 99.48 | 100 |
F_score | 99.98 | 100 | 100 | 100 | 99.74 | 100 | |
Testing 11 | ACC | 100 | 100 | 99.22 | 100 | 100 | 100 |
F_score | 100 | 100 | 99.61 | 100 | 100 | 100 | |
Testing 12 | ACC | 100 | 100 | 100 | 100 | 99.44 | 100 |
F_score | 100 | 100 | 100 | 100 | 99.72 | 100 | |
Testing 13 | ACC | 100 | 99.93 | 100 | 100 | 98.99 | 100 |
F_score | 100 | 99.96 | 100 | 100 | 99.49 | 100 | |
Average ACC | 99.99 | 99.99 | 99.94 | 99.93 | 99.40 | 100 | |
99.88 | |||||||
Average F_score | 99.99 | 99.997 | 99.97 | 99.97 | 99.70 | 100 | |
99.94 |
Our Method | Previous Method [40] | |
---|---|---|
Average ACC | 99.88 | 92.8 |
Average F_score | 99.94 | 93.9 |
Total of Testing 1–13 | Recognized Arrows | ||||||
---|---|---|---|---|---|---|---|
FA | FLA | FLRA | FRA | LA | RA | ||
Actual arrows | FA | 31,451 | 0 | 0 | 0 | 0 | 0 |
FLA | 0 | 18,102 | 0 | 0 | 0 | 0 | |
FLRA | 0 | 0 | 22,275 | 0 | 0 | 0 | |
FRA | 0 | 0 | 0 | 18,848 | 0 | 0 | |
LA | 0 | 0 | 0 | 0 | 33,532 | 13 | |
RA | 0 | 0 | 0 | 0 | 0 | 31,779 |
Our Method | |
---|---|
Average ACC | 99.99 |
Average F_score | 99.99 |
© 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Vokhidov, H.; Hong, H.G.; Kang, J.K.; Hoang, T.M.; Park, K.R. Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network. Sensors 2016, 16, 2160. https://doi.org/10.3390/s16122160
Vokhidov H, Hong HG, Kang JK, Hoang TM, Park KR. Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network. Sensors. 2016; 16(12):2160. https://doi.org/10.3390/s16122160
Chicago/Turabian StyleVokhidov, Husan, Hyung Gil Hong, Jin Kyu Kang, Toan Minh Hoang, and Kang Ryoung Park. 2016. "Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network" Sensors 16, no. 12: 2160. https://doi.org/10.3390/s16122160
APA StyleVokhidov, H., Hong, H. G., Kang, J. K., Hoang, T. M., & Park, K. R. (2016). Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network. Sensors, 16(12), 2160. https://doi.org/10.3390/s16122160