Real-Time Evaluation of Perception Uncertainty and Validity Verification of Autonomous Driving
<p>Comparison graph between ground truth and perception results under the KITTI dataset based on DNN algorithms. There is uncertainty in perception results, such as missed detection, false detection, position errors, and orientation errors. Green boxes represent the ground truth of data labels and blue boxes donate the perception results of DNN algorithms based on lidar and camera. The red numbers represent the order of the dataset frames.</p> "> Figure 2
<p>The logic flow between different algorithms.</p> "> Figure 3
<p>The Schematic sequence of the perception effectiveness judgment and uncertainty evaluation.</p> "> Figure 4
<p>Scheme flow of objects matching algorithm.</p> "> Figure 5
<p>Objects matching algorithm:Diagram of triangle matching.</p> "> Figure 6
<p>Scheme flow of the perception effectiveness judgment.</p> "> Figure 7
<p>Scheme flow of PointPillars [<a href="#B27-sensors-23-02867" class="html-bibr">27</a>].</p> "> Figure 8
<p>Scheme flow of SMOKE [<a href="#B28-sensors-23-02867" class="html-bibr">28</a>].</p> "> Figure 9
<p>Judgment results of perception effectiveness: the judgment is correct. The judgment and verification are valid after matching and verifying with the ground truth.</p> "> Figure 10
<p>Judgment results of perception effectiveness: the judgment is correct. The judgment and verification are invalid after matching and verifying with the ground truth.</p> "> Figure 11
<p>Judgment results of perception effectiveness: the judgment is correct. The judgment and verification are invalid after matching and verifying with the ground truth.</p> "> Figure 12
<p>Spatial uncertainty based on deep ensemble.</p> "> Figure 13
<p>Correlation research between uncertainty and error of Pointpillars (3769 frames) In order: the horizontal direction, longitudinal direction, vertical direction, and orientation of the car.</p> "> Figure 14
<p>Correlation research between uncertainty and error of SMOKE (3769 frames) In order: the horizontal direction, longitudinal direction, vertical direction, and orientation of the car.</p> "> Figure 15
<p>The relationship between object distance and perception uncertainty.</p> "> Figure 16
<p>The relationship between object occlusion and perception uncertainty.</p> ">
Abstract
:1. Introduction
2. Related Works
3. Methodology
3.1. Deep Ensemble
Algorithm 1 Deep Ensemble |
Input: Neural networks and number of networks . |
Output: Associated classification probability , location of detected objects , rotation , and dimension . |
|
3.2. Clustering Algorithm
Algorithm 2 Basic sequential algorithmic scheme with intra-sample exclusivity |
Input: A set of predictions . |
Output: A set of clusters . |
|
3.3. Object-Matching Algorithm
- (1)
- Set object sets and object sets . Calculate the number of object sets and the number of object sets .
- (2)
- Calculate the minimum value of elements in the object sets .
- (3)
- If , use the K-Nearest Neighbor (KNN) method to calculate the distance of all of the points in to all of the points in . The distance between point A and point B can be calculated in Equation (1).
- (4)
- If , it is necessary to add a point at a distance to form a triangle, and both object sets need to add a point. The detection range is set to be 50 m. In order to achieve a better matching effect, the coordinates of the object points selected in this study can be (500,500) and (499,499). Points farther than 50 m away are added to form triangles for matching. These points do not belong to the objects within the perceived range and will not be output. It should be noted that the two points cannot be the same, but the distance between them cannot be too large. After adding points, triangle matching method can be used, which is shown below.
- (5)
- If , the triangle matching method can be used. The diagram of the triangle matching algorithm is shown in Figure 5.
- (6)
- The principle of the triangle matching method can be described from step (1) to step (7). Taking the object in the BEV perspective as an example, this paper only considers the index elements in x and z directions in the KITTI dataset camera coordinate system.
- (1)
- Numbering data to all of the points in and .
- (2)
- Calculate the coordinate difference values , , , between the maximum and minimum values in x and z directions in object sets and object sets by using Equations (2) and (3).
- (3)
- Sort the points in object sets and object sets . Calculate the maximum value of the coordinate difference values. If the maximum value is in the x direction, the object sets and object sets are sorted by the value of x. Otherwise, object sets and object sets are sorted by the value of z.
- (4)
- Randomly select three points to form a triangle in object sets and object sets , and return the index of the point location.
- (5)
- The formed triangles in object sets and object sets are normalized. The side length of each triangle is divided by the shortest side length. This setting can ensure the setting of a uniform threshold for successful matching.
- (6)
- Calculate the error sum of the edges and points for each triangle in object sets with all triangles in object sets . Take any two points in a triangle as an example. As shown in Figure 5, Point and point are corresponding points. The error of point A can be expressed in Equation (4).The error of edge AB can be expressed in Equation (5).The total error (TE) of two triangles can be expressed in Equation (6).
- (7)
- Calculate the minimum value of all trianglestriangle. If is less than the triangle error threshold , the two triangles are matched, and the corresponding points of the triangle sorted by (3) are the matching objects.
- (7)
- Judge the remaining points, and repeat steps (3), (4), and (5) until all points are matched.
3.4. Perception Effectiveness Judgment
- (1)
- The perception results are processed in the DE method. After clustering and statistics, the number of networks detecting the object and the average confidence level of the detected object are calculated in Equation (7), respectively.
- (2)
- Set detection network number threshold and average confidence and .
- (3)
- Judge the uncertainty by using Equation (8).
- (1)
- First, the triangle matching method is used for object matching of multi-source perception results. Objects that are successfully matched are considered to be real objects.
- (2)
- For the results that cannot be matched in each perception source, the DE method is used for judgment.
- (3)
- The final processing results are fused.
3.5. Spatial Uncertainty
4. Experimental Results
4.1. Experiment Settings
4.1.1. PointPillars 3D Objects Detection Network
4.1.2. SMOKE 3D Objects Detection Network
4.1.3. Implementation Details
4.2. Results
4.2.1. Perception Effectiveness Judgement
4.2.2. Spatial Uncertainty
4.2.3. Validation of Perception Uncertainty
4.2.4. Influencing Factors of Uncertainty
- (1)
- The proposed real-time judgment on perception effectiveness has a high accuracy of 92%.
- (2)
- The estimation of spatial uncertainty based on DE is positively correlated to the ground truth error.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
DE | Deep Ensemble |
DNN | Deep neural network |
KNN | K-Nearest Neighbor |
AD | Autonomous driving |
MCD | Monte Carlo Dropout |
MCMC | Markov Chain Monte Carlo |
IOU | Intersection over union |
JIOU | Jaccard IoU |
SOTIF | Safety of the intended functionality |
BSAS_excl | Basic sequential algorithmic scheme with intra-sample exclusivity |
TP | True Positive |
FP | False Positive |
FN | False Negative |
TV | Total Variance |
TE | Total Error |
CNN | Convolutional Neural Network |
DCN | Deformable Convolutional Network |
References
- Chen, L.; He, Y.; Wang, Q.; Pan, W.; Ming, Z. Joint optimization of sensing, decision-making and motion-controlling for autonomous vehicles: A deep reinforcement learning approach. IEEE Trans. Veh. Technol. 2022, 71, 4642–4654. [Google Scholar] [CrossRef]
- Cosgun, A.; Ma, L.; Chiu, J.; Huang, J.; Demir, M.; Anon, A.M.; Lian, T.; Tafish, H.; Al-Stouhi, S. Towards full automated drive in urban environments: A demonstration in gomentum station, california. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 1811–1818. [Google Scholar]
- Chen, J.; Sun, J. Platoon Separation Strategy Optimization Method Based on Deep Cognition of a Driver’s Behavior at Signalized Intersections. IEEE Access 2020, 8, 17779–17791. [Google Scholar] [CrossRef]
- Wang, S.; Jiang, K.; Chen, J.; Yang, M.; Fu, Z.; Wen, T.; Yang, D. Skeleton-based Traffic Command Recognition at Road Intersections for Intelligent Vehicles. Neurocomputing 2022, 501, 123–134. [Google Scholar] [CrossRef]
- Rahman, Q.M.; Corke, P.; Dayoub, F. Run-Time Monitoring of Machine Learning for Robotic Perception: A Survey of Emerging Trends. IEEE Access 2021, 9, 20067–20075. [Google Scholar] [CrossRef]
- Peng, L.; Li, B.; Yu, W.; Yang, K.; Shao, W.; Wang, H. SOTIF Entropy: Online SOTIF Risk Quantification and Mitigation for Autonomous Driving. arXiv 2022, arXiv:2211.04009. [Google Scholar]
- Feng, D.; Harakeh, A.; Waslander, S.L.; Dietmayer, K. A Review and Comparative Study on Probabilistic Object Detection in Autonomous Driving. IEEE Trans. Intell. Transp. Syst. 2022, 23, 9961–9980. [Google Scholar] [CrossRef]
- Feng, D.; Haase-Schutz, C.; Rosenbaum, L.; Hertlein, H.; Glaser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1341–1360. [Google Scholar] [CrossRef] [Green Version]
- Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Al Sallab, A.A.; Yogamani, S.; Pérez, P. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4909–4926. [Google Scholar] [CrossRef]
- Junior: The Stanford entry in the Urban Challenge. J. Field Robot. 2008, 25, 569–597. [CrossRef] [Green Version]
- Held, D.; Guillory, D.; Rebsamen, B.; Thrun, S.; Savarese, S. A Probabilistic Framework for Real-time 3D Segmentation using Spatial, Temporal, and Semantic Cues. In Proceedings of the Robotics: Science and Systems 2016, Ann Arbor, MI, USA, 18–22 June 2016. [Google Scholar]
- Van Amersfoort, J.; Smith, L.; Teh, Y.W.; Gal, Y. Uncertainty estimation using a single deep deterministic neural network. PMLR 2020, 119, 9690–9700. [Google Scholar]
- Meyer, G.P.; Thakurdesai, N. Learning an uncertainty-aware object detector for autonomous driving. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25–29 October 2020; pp. 10521–10527. [Google Scholar] [CrossRef]
- Liu, Y.; Yang, G.; Hosseiny, M.; Azadikhah, A.; Mirak, S.A.; Miao, Q.; Raman, S.S.; Sung, K. Exploring uncertainty measures in Bayesian deep attentive neural networks for prostate zonal segmentation. IEEE Access 2020, 8, 151817–151828. [Google Scholar] [CrossRef]
- Tao, Y.; Sun, H.; Cai, Y. Predictions of deep excavation responses considering model uncertainty: Integrating BiLSTM neural networks with Bayesian updating. Int. J. Geomech. 2022, 22, 04021250. [Google Scholar] [CrossRef]
- Zhang, X.; Chan, F.T.; Mahadevan, S. Explainable machine learning in image classification models: An uncertainty quantification perspective. Knowl. Based Syst. 2022, 243, 108418. [Google Scholar] [CrossRef]
- Miller, D.; Dayoub, F.; Milford, M.; Sunderhauf, N. Evaluating merging strategies for sampling-based uncertainty techniques in object detection. In Proceedings of the IEEE International Conference on Robotics and Automation, Montreal, QC, Canada, 20–24 May 2019; pp. 2348–2354. [Google Scholar] [CrossRef] [Green Version]
- Goswami, S. False Detection (Positives and Negatives) in Object Detection. arXiv 2020, arXiv:abs/2008.06986. [Google Scholar]
- Bogdoll, D.; Nitsche, M.; Zollner, J.M. Anomaly Detection in Autonomous Driving: A Survey. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 4487–4498. [Google Scholar] [CrossRef]
- Jiao, X.; Chen, J.; Wang, Y.; Jiang, K.; Cao, Z.; Yang, M.; Yang, D. Reliable autonomous driving environment model with unified state-extended boundary. IEEE Trans. Intell. Transp. Syst. 2022, 24, 516–527. [Google Scholar] [CrossRef]
- Jiao, X.; Cao, Z.; Chen, J.; Jiang, K.; Yang, D. A General Autonomous Driving Planner Adaptive to Scenario Characteristics. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21228–21240. [Google Scholar] [CrossRef]
- Liu, J.; Wang, H.; Peng, L.; Cao, Z.; Yang, D.; Li, J. PNNUAD: Perception Neural Networks Uncertainty Aware Decision-Making for Autonomous Vehicle. IEEE Trans. Intell. Transp. Syst. 2022, 23, 24355–24368. [Google Scholar] [CrossRef]
- Peng, L.; Wang, H.; Li, J. Uncertainty Evaluation of Object Detection Algorithms for Autonomous Vehicles. Automot. Innov. 2021, 4, 241–252. [Google Scholar] [CrossRef]
- Cao, Z.; Liu, J.; Zhou, W.; Jiao, X.; Yang, D. LiDAR-based Object Detection Failure Tolerated Autonomous Driving Planning System. In Proceedings of the 2021 IEEE Intelligent Vehicles Symposium (IV), Nagoya, Japan, 11–17 July 2021; pp. 122–128. [Google Scholar]
- Yang, L.; Zhang, X.; Wang, L.; Zhu, M.; Zhang, C.F.; Li, J. Mix-Teaching: A Simple, Unified and Effective Semi-Supervised Learning Framework for Monocular 3D Object Detection. arXiv 2022, arXiv:2207.04448. [Google Scholar]
- Pitropov, M.; Huang, C.; Abdelzad, V.; Czarnecki, K.; Waslander, S. LiDAR-MIMO: Efficient Uncertainty Estimation for LiDAR-based 3D Object Detection. In Proceedings of the IEEE Intelligent Vehicles Symposium, Aachen, Germany, 4–9 June 2022; pp. 813–820. [Google Scholar] [CrossRef]
- Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection From Point Clouds. arXiv 2019, arXiv:1812.05784. [Google Scholar]
- Liu, Z.; Wu, Z.; Toth, R. SMOKE: Single-stage monocular 3D object detection via keypoint estimation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 4289–4298. [Google Scholar] [CrossRef]
- Czarnecki, K.; Salay, R. Towards a framework to manage perceptual uncertainty for safe automated driving. In Proceedings of the International Conferrence on Computer Safety, Vasteras, Sweden, 18–21 September 2018; Volume 11094 LNCS, pp. 439–445. [Google Scholar] [CrossRef] [Green Version]
- Feng, D.; Wang, Z.; Zhou, Y.; Rosenbaum, L.; Timm, F.; Dietmayer, K.; Tomizuka, M.; Zhan, W. Labels are Not Perfect: Inferring Spatial Uncertainty in Object Detection. IEEE Trans. Intell. Transp. Syst. 2022, 23, 9981–9994. [Google Scholar] [CrossRef]
- Wu, P.; Chen, S.; Metaxas, D.N. MotionNet: Joint Perception and Motion Prediction for Autonomous Driving Based on Bird’s Eye View Maps. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 11382–11392. [Google Scholar] [CrossRef]
- Mena, J.; Pujol, O.; Vitrià, J. A Survey on Uncertainty Estimation in Deep Learning Classification Systems from a Bayesian Perspective. ACM Comput. Surv. 2022, 54, 1–36. [Google Scholar] [CrossRef]
- Melucci, M. Relevance Feedback Algorithms Inspired by Quantum Detection. IEEE Trans. Knowl. Data Eng. 2016, 28, 1022–1034. [Google Scholar] [CrossRef]
- Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the 2017 Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6403–6414. [Google Scholar]
- Wang, Z.; Feng, D.; Zhou, Y.; Rosenbaum, L.; Timm, F.; Dietmayer, K.; Tomizuka, M.; Zhan, W. Inferring spatial uncertainty in object detection. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25–29 October 2020; pp. 5792–5799. [Google Scholar] [CrossRef]
- Gasperini, S.; Haug, J.; Mahani, M.A.N.; Marcos-Ramiro, A.; Navab, N.; Busam, B.; Tombari, F. CertainNet: Sampling-Free Uncertainty Estimation for Object Detection. IEEE Robot. Autom. Lett. 2022, 7, 698–705. [Google Scholar] [CrossRef]
- Feng, D.; Rosenbaum, L.; Timm, F.; Dietmayer, K. Leveraging heteroscedastic aleatoric uncertainties for robust real-time LiDAR 3D object detection. In Proceedings of the IEEE Intelligent Vehicles Symposium, Paris, France, 9–12 June 2019; pp. 1280–1287. [Google Scholar] [CrossRef] [Green Version]
- Pan, H.; Wang, Z.; Zhan, W.; Tomizuka, M. Towards Better Performance and More Explainable Uncertainty for 3D Object Detection of Autonomous Vehicles. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems, ITSC 2020, Rhodes, Greece, 20–23 September 2020. [Google Scholar] [CrossRef]
- Kendall, A.; Badrinarayanan, V.; Cipolla, R. Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding. arXiv 2015, arXiv:1511.02680. [Google Scholar]
- Michelmore, R.; Wicker, M.; Laurenti, L.; Cardelli, L.; Gal, Y.; Kwiatkowska, M. Uncertainty Quantification with Statistical Guarantees in End-to-End Autonomous Driving Control. In Proceedings of the International Conference on Robotics and Automation, Paris, France, 31 May–31 August 2020. [Google Scholar]
- Shao, W.; Xu, Y.; Peng, L.; Li, J.; Wang, H. Failure Detection for Motion Prediction of Autonomous Driving: An Uncertainty Perspective. arXiv 2023, arXiv:2301.04421. [Google Scholar]
- Huang, Z.; Wu, J.; Lv, C. Efficient deep reinforcement learning with imitative expert priors for autonomous driving. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef]
- Bern, M.; Edelsbrunner, H.; Eppstein, D.; Tan, S. Edge insertion for optimal triangulations. Discret. Comput. Geom. 1993. [Google Scholar] [CrossRef] [Green Version]
- Fekete, S.P. The Complexity of MaxMin Length Triangulation. arXiv 2012, arXiv:1208.0202. [Google Scholar]
Net | Precision | Recall | F1-Score |
---|---|---|---|
PointPillars | 86.8% | 57.8% | 0.6939 |
SMOKE | 76.5% | 47.1% | 0.583 |
Net | NumNET | MeanScore |
---|---|---|
PointPillars | 4.375 | 0.6748 |
SMOKE | 2.8659 | 0.4388 |
Correct Judgment | Wrong Judgment | Failure Diagnosis Rate |
---|---|---|
920 frames | 80 frames | 92% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yang, M.; Jiang, K.; Wen, J.; Peng, L.; Yang, Y.; Wang, H.; Yang, M.; Jiao, X.; Yang, D. Real-Time Evaluation of Perception Uncertainty and Validity Verification of Autonomous Driving. Sensors 2023, 23, 2867. https://doi.org/10.3390/s23052867
Yang M, Jiang K, Wen J, Peng L, Yang Y, Wang H, Yang M, Jiao X, Yang D. Real-Time Evaluation of Perception Uncertainty and Validity Verification of Autonomous Driving. Sensors. 2023; 23(5):2867. https://doi.org/10.3390/s23052867
Chicago/Turabian StyleYang, Mingliang, Kun Jiang, Junze Wen, Liang Peng, Yanding Yang, Hong Wang, Mengmeng Yang, Xinyu Jiao, and Diange Yang. 2023. "Real-Time Evaluation of Perception Uncertainty and Validity Verification of Autonomous Driving" Sensors 23, no. 5: 2867. https://doi.org/10.3390/s23052867
APA StyleYang, M., Jiang, K., Wen, J., Peng, L., Yang, Y., Wang, H., Yang, M., Jiao, X., & Yang, D. (2023). Real-Time Evaluation of Perception Uncertainty and Validity Verification of Autonomous Driving. Sensors, 23(5), 2867. https://doi.org/10.3390/s23052867