1. Introduction
Tunnel infrastructure shows signs of deterioration over time due to construction defects, ageing, unexpected overloading and natural phenomena, possibly leading to problems in structural integrity. Consequently, periodic inspections of concrete tunnels should be conducted to ensure that they are still healthy and safe. Today, these are predominantly performed through periodic visual observations, looking for structural defects such as cracking, spalling and water leakage to identify possible changes in the infrastructure with respect to a previous survey. Strain gauges, displacement meters and other contact measurement methods can be employed to monitor problematic areas more closely within these structures. To conduct such observations, personnel are required to be physically present in the tunnel. Associated with this, there are several drawbacks, including the human presence in hazardous environments and the financial cost involved to train and hire people to do the inspections. In addition, these inspections require a considerable amount of time to perform, leading to longer operation down-times and thus higher monetary losses. Furthermore, the outcome from these inspections is highly dependent on human subjectivity, leading to possible inaccuracies that result in false and missing change detections.
All this has led to an increase in the demand for robotic systems and remote operations to reduce direct human intervention. One possible solution is to use vision-based sensing monitoring systems that can provide reliable objective results. Hence, a substantial effort can be found in the literature on automating inspections using image processing and machine learning methods to detect and classify cracks, structural deformities, and the presence of water along the tunnel linings. Moreover, the imagery can also be used to provide multiview point clouds to visualise the location of these defects and identify issues with surface deformation [
1,
2]. This task is not trivial as tunnel environments are characterised by non-uniform illumination and shadows, deformations, lack of features, dirt, stains, and possible occlusion of parts of the walls due to cables, pipes and other servicing equipment.
Whilst defect identification is essential to automate inspection, regular monitoring of tunnel linings can provide a more informative survey to further automate inspection and analysis. In this paper, we present a tunnel inspection application that uses robotics, computer vision and data fusion to monitor for changes on tunnel linings. The main contributions of this work are:
Integration of a commercial inspection camera system on a robotic platform;
Development of a specular highlight localisation algorithm based on uneven illumination correction and deep learning to remove these artefacts from the images captured by the system;
Implementation and analysis of bi-temporal image fusion techniques for image comparison and change-map generation;
Implementation and evaluation of two decision-level fusion techniques for robust change detection.
The tunnel used to develop this system is within CERN, the European Organisation for Nuclear Research. The considered tunnel is a 27 km long tunnel lying at around 100 m below the ground, hosting the world’s largest particle accelerator, the Large Hadron Collider (LHC).
The remainder of this article is structured as follows.
Section 2 reviews the state-of-the-art with respect to automated tunnel inspection and the techniques used here. The proposed solution is presented in
Section 3.
Section 4 explains the image acquisition part. In
Section 5, pre-processing tasks are described. Bi-temporal image fusion is described in
Section 6.
Section 7 discusses decision-level fusion in the context of change detection, followed by the change map (CM) analysis process presented in
Section 8. A performance evaluation is made in
Section 9. A summary and suggestions for future work conclude this article.
2. Background Information
2.1. General Tunnel Inspection
Research on automated health monitoring of tunnel structures has received significant attention in recent years, as recorded in [
3,
4]. Various solutions that deal with different aspects of automated tunnel inspection were proposed through the use of cost-effective photographic equipment and computer vision. In [
5], an extensive survey of works within the whole image-based tunnel inspection spectrum is presented. This includes tunnel profile monitoring, crack and leakage detection, as well as tunnel surface documentation and visualisation.
2.2. Change Detection
Change detection is a well-researched problem in the fields of video surveillance, remote sensing ([
6,
7]) and medical imaging, amongst others. Reviews of change identification methods are found in [
8,
9]. However, literature on the detection of changes on tunnel linings is still lacking, possibly due to the challenges encountered in this area. Some of these can be referred to in [
10,
11,
12,
13,
14]. A system aimed at supporting structural inspectors to monitor the condition of railway tunnels is presented in [
10]. An array of cameras with uniform lighting is used to capture the image data that are registered and stitched to allow tunnel inspectors to examine for defects with reference to their location. The solution presented also implements a change detection algorithm to attract attention to areas that are considered important, and the inspectors can refer to previous imagery from previous inspections to assess the evolution of the defect. A three-dimensional model of the structure is also created to better contextualise these defects. This paper does not report on the accuracy of the techniques used. A change detection system that relies on computer vision techniques was presented in [
11]. A camera was implemented on a monorail, and the captured images were first processed by a shading correction algorithm before using image mosaicing to correct for offsets between successive inspections. A change detection algorithm was then applied on a survey image and the reference image using image differencing, binary image comparison and optical flow analysis. The solution in [
12] registers new images on a three-dimensional surface model to detect and assist visual inspections. The changes are detected using a probabilistic model that considers different feature maps and a geometric prior to reducing the impact of noise and nuisance sources. The system presented in [
13] builds panoramas of the tunnel surface and registers images from different inspections. A two-channel convolutional neural network is used to determine anomalous changes between the current survey image patch and the corresponding reference. The work in [
14] discusses the application of Procrustes methods in photogrammetry. A method that uses freeform surface modelling and a mask region-based convolutional neural network is presented in [
15] to generate 3D tunnel structures and superimpose the position of cracks on the model. Change detection techniques can be applied to measurements completed at different times.
The goal of any change detection algorithm is to detect significant changes between the new measurement and a reference, where the reference is the previous measurement at the exact same location. However, this is not a trivial task as the accuracy of such algorithms can be hindered by apparent intensity changes that result from camera motion and different lighting conditions. Hence, pre-processing steps involving geometric, radiometric adjustments and semantic segmentation are generally required as a primary stage to provide more robust change detection solutions.
2.3. Data Fusion
Data fusion combines data from different methods for increased reliability, higher redundancy and improved identification. Surveys of different fusion architectures are presented in [
16,
17]. Image fusion is a specific type of data fusion, classified into pixel, feature and decision levels. Image fusion applications can also be categorised by the time, view or modality at which the images are taken. Multiview applications, such as [
18,
19], fuse images from the same modality but from different viewpoints. Images taken at different times are combined using multi-temporal fusion to detect changes between them or to synthesise images not photographed at a desired time, as in [
20,
21]. In multi-modal fusion, images coming from different sensors are combined, such as in [
22,
23].
3. Solution Overview
The proposed solution is illustrated in
Figure 1. The acquisition of images in the tunnel is made by a mobile robotic platform. Pre-processing steps involving radiometric adjustments and specular highlight localisation are applied to minimise false change detections due to camera motion, uneven illumination, and different light reflections. Bi-temporal fusion, involving image differencing, principal component analysis (PCA) and a structural similarity index (SSIM) followed by decision-level image fusion is employed at respective stages to achieve change detection.
4. Image Data Acquisition
4.1. Acquisition System
A camera system [
24] designed to inspect cylindrical environments was identified on the market. The system is composed of a twelve-unit camera rig, as shown in
Figure 2b, two flashlights, an encoder wheel, two batteries and a computer unit with software for camera synchronisation. The twelve industrial cameras mounted on the rig had a 5-megapixel resolution with adapted lenses. During a demo test in the LHC tunnel, this system was integrated on CERNBot [
25], one of the readily available robotic platforms at CERN, as shown in
Figure 2a. The encoder wheel was attached to the CERNBot, as shown in
Figure 2c.
4.2. Dataset
CERNbot was driven parallel to the tunnel wall at a speed of around 0.2 m/s along a section of the LHC tunnel while capturing synchronised images. This image set is referred to as . Changes to the structure were then simulated by markings on the wall. These markings had different resolutions, with the fine cracks being 0.5 mm wide. The CERNbot was again driven along the same section and at the same speed capturing .
Using these data, 3D models were generated and unwarped into orthophotos using scripts run by the company supplying the same camera system [
24]. Using location information from the encoder wheel, orthophotos could be accurately registered, as seen in
Figure 3, such that pixel-based change detection (PBCD) techniques could be applied. Each orthophoto is segmented into ten parts along its height, and each image crop covers 0.5 m of the tunnel length. Such images were used for training and testing of the algorithms used in the proposed tunnel lining change detection solution.
5. Image Pre-Processing
The detected changes should be due to new defects or from the evolution of already existing ones. Other changes caused by lighting sources should be identified as “nuisance”, preventing them from being propagated in a change detection pipeline. Hence, pre-processing needs to be completed to cater for this.
5.1. Uneven Illumination Correction
An uneven amount of light falling on different areas causes non-uniformity in images leading to nuisance when comparing images. To adjust the uneven illumination, we use the shading algorithm in [
26]. A low-pass filter is applied to the original image using a median filter with a large kernel. The illumination corrected image is obtained through a pixel-wise division of the original image by the low-pass filtered image. As observed in
Figure 4c, subtracting the original images generates a difference image full of “white change areas”; however, this is due to uneven illumination. On the other hand, when the images are pre-processed to correct for uneven illumination, their difference image does not have any “white areas” even if there is a change in lighting, as shown in
Figure 4f. More details on how this correction mechanism compares to other solutions can be found in [
26]. Considering these results, this method proves to be an effective pre-processing method to provide useful images for subsequent processing.
5.2. Specular Highlight Localisation
During image acquisition, flashlights cause reflections on metal racks/pipes present on the wall, resulting in specular highlights in the images. Such highlights are not constant in time or place, leading to false detections when subtracting images to identify changes, as shown in
Figure 4. Thus, highlight detection was implemented to localise these regions in the image pair, as displayed in
Figure 5. For this, semantic segmentation using a modified U-Net [
27] architecture is implemented [
28]. Morphological operations and connectivity analysis are then applied to the segmentation images to generate bounding boxes around highlighted areas in the image pair, as illustrated in
Figure 5c. Such masks are later fused with the CM to mask out these false change candidates.
6. Bi-Temporal Image Fusion
Multi-temporal fusion combines data from images of the same scene, acquired at different times. Hence, this approach can be used to identify changes in a scene by comparing images. In this scenario, bi-temporal image fusion is applied between the two temporal images; reference and survey. The reference and survey image pair in
Figure 6 is used in the explanation of the subsequent methods.
The image fusion method used in this work involves image differencing, PCA and SSIM. The image differencing method finds the difference between the reference and survey images. In an ideal scenario, the result of this will yield an all-zero image except for where a change occurred. However, errors in registration and in the data acquisition process will depart from this ideal case. The PCA is applied to the stacked reference and survey images. The components of the PCA yield the information from both images, the difference between them and other components representing noise. The component representing the difference is used in this case. The PCA is selected as it removes correlated features focusing on the important features, which are the changes in this case. The final metric chosen is the SSIM, which measures the similarity of the reference and survey images. In this case, we are looking at the parts of the images that are not similar, representing change.
6.1. Image Difference
This method has been used extensively for change detection in various applications, including background subtraction for movement detection and remote sensing. As outlined in [
29], two images of the same scene taken at separate times
and
are subtracted pixel-wise. After the subtraction, the magnitude of the difference value is compared against a threshold. Pixels with a difference magnitude higher than the pre-defined threshold are classified as “change”, otherwise noted as “no change”. The
CM is generated using:
where
is the image at time
,
is the image at time
and
T is the threshold on the difference magnitude. This method is simple and requires low computation; however, its accuracy depends on the threshold set. As
T is increased, the number of change pixels decreases, implying the elimination of lower difference magnitudes, thus more noise suppression. However, the “valid change” pixels are lost at
in this particular example.
A fixed threshold value cannot satisfy all scenarios; thus, a better approach is to set the threshold automatically depending on the images being compared. The Gaussian valley emphasis (VE) method proposed in [
30] is used, generating a CM with only a few “noise changes” while retaining the “crack change”, as observed in
Figure 7. The standard deviation of the Gaussian window was empirically set to
.
Approaches such as [
31] use multiple frames differencing to mitigate issues of sudden illumination changes and ghosting problems. However, in the tunnel environment, illumination changes at the same locations between surveys are limited and the simple differencing was deemed adequate. Other techniques can be applied to reduce the impact of noise, such as wavelets, where [
32] proposed a threshold function that is based in the multi-layer wavelet transform. Moreover, [
33] uses wavelets to find a trade-off between the removal of noise and extraction of edges in high-speed moving target imagery.
6.2. Principal Component Analysis (PCA)
PCA reduces the dimensionality of a dataset while maintaining the variances. Independent data transformation analysis applies PCA on each of the temporal images separately. The derived principal components are then analysed by applying other change detection techniques, such as image differencing or regression. On the other hand, merged data transformation analysis stacks N temporal images of p channels each, fuses them into a single -channel image and applies PCA on the latter. In this bi-temporal scenario, the merged data approach is used, and the reference and survey images are stacked on each other. The method was investigated in terms of the original RGB images and the pre-processed images, which is the illumination-corrected images. The processing is done on the RGB data as it provides two more dimensions compared to greyscale images and hence more information that helps in detecting changes.
When RGB images (
) are used, the stacked images are merged into a 6-channel image. The first component (
), corresponding to the highest eigenvalues, contains most of the information from both images.
represents the difference between temporal images while later components contain noise information. Experimental results show that PCA is scene-dependent, thus comparison between different data is often difficult to interpret using a fixed condition, implying the need for scenario-dependent thresholds. When considering
, the histogram shape is not clearly defined at its tails, making it difficult to find an adaptive threshold pair. When pre-processed images (
) are used, a stacked 2-channel image is generated. From PCA, the first component
represents the difference between temporal images while
contains most of the information from both images. In this case, when considering
, the “crack change” has a high value (white), the “pipe reflections change” has a low value (black) and the rest of the wall has a medium value (grey). This again implies that the histogram contains change elements at both of its tails implying the need for a threshold pair. In this case, however, as observed in
Figure 8, the histogram shape follows a Gaussian distribution. To automatically find a threshold pair, the statistical process control (SPC) principle [
34] was adopted as it involves binarising an image with a range of pixel values away from the mean pixel where the range is controlled by an input control factor.
The lower and higher thresholds for each image pair under comparison were determined via SPC using:
where
,
are the mean and standard deviation of
, respectively, and
c is a constant whose value of three was empirically set for the considered dataset.
Applying Equation (
5) on the
of the original and
of the pre-processed images generated the CMs in
Figure 9.
6.3. Structural Similarity (SSIM) Index
Detection of changes between two images can be considered as a comparison of similarities between two images [
35]. The larger the difference between the survey image and the reference image is, the smaller the SSIM will be. This metric is based on the human visual system and is in line with what humans perceive as change. SSIM [
36] performs different similarity measurements of luminance, contrast and structure, and thereafter combines them to obtain a single value. Considering two image blocks
A and
B, the SSIM is given by:
where
,
are the mean and
,
are the variance of
A and
B while
is the covariance between
A and
B. Constants
are calculated using:
where
, generally
as set in [
36] and
L is the dynamic range of the pixel values (
for 8-bit greyscale images). Here, SSIM is used as a PBCD method to generate a CM between a reference and survey image. Therefore, considering the neighbourhood around point
in two temporal images
and
then the SSIM at point
is calculated using Equation (
6). The SSIM is normalised to a range of
and thresholded using:
where
represents the difference image and
T is a constant. A fixed threshold value cannot satisfy all scenarios; thus, the Gaussian VE automatic thresholding method is applied. An investigation of the performance in change detection is done using greyscale images, the V channel in HSV images and pre-processed images corrected for uneven illumination. In general, the best results with minimum noise were obtained using greyscale images, as shown in
Figure 10.
7. Decision-Level Fusion
Considering the complementary advantages of the implemented PBCD methods, the generated CMs from image differencing (), PCA () and SSIM () are fused into a single CM using decision-level fusion methods. Different decision-level fusion methods were implemented; however, the PCA-weighted sum and the majority voting gave the best results and are reported in this section.
7.1. PCA-Weighted Sum
The PCA-based fusion algorithm is illustrated in
Figure 11, where
and
are the first three components of one PCA analysis of the stacked CMs. The first three principal components are used as they represent most of the variation in the data. The
PCA is applied on the three CMs, namely
and
. The first principal component,
, measures mainly the impact of the difference CM, the second principal component,
, has a strong association with PCA CM and the third principal component,
, more so represents the SSIM CM. Hence, these
are used as weights of the CMs to give more importance to the CMs that have better captured the change. The summation of these weighted terms generates the fused
CM using:
As shown in
Figure 12, the PCA approach generates few noise pixels while retaining the actual changes; in this case, those belonging to the crack.
7.2. Majority Voting
In the majority voting algorithm, the three different CMs (
,
and
) cast a unit vote and if at least two of the CMs register a change, then the corresponding pixel in the fused CM is assigned 1 (change), otherwise 0 (no change). Similar to the previous method, this fusion approach generates only a few noise pixels while retaining the actual changes, as shown in
Figure 13.
8. Change Map Analysis
At this point, the fused CM may still contain “nuisance change” areas that should not be considered as “changes”. Hence, the CM analysis process illustrated in
Figure 14 was developed.
8.1. Specular Highlights Filtering
Fusion between the inverse specular highlight mask image
and the final
is done through an AND operation defined by:
8.2. Morphological Operations
The filtered fused may contain some small “change areas” coming from image noise and minor registration errors. Here, a morphological closing operation that uses dilation and erosion sequentially, is applied to the fused CM. This joins any change segments by filling gaps, such as in “crack changes”, while at the same time ignores the “noise changes”.
8.3. Connected Components Labelling
Next, connected components labelling with 8-connectivity is used to identify and group neighbouring pixels into “change components”.
8.4. Dimension Filtering
The components are now filtered by their size. A “change component” is only retained if its width and/or height satisfies the corresponding thresholds
. Using the GDAL library [
37], the orthophoto raster scale is obtained, and using the simple proportion principle, the physical dimensions of the segment’s field of view (FoV) are calculated. Using the configurable parameter
representing the minimum dimension for a detected change together with the corresponding image dimension and
, the thresholds are calculated using:
where
W and
H represent the width and the height of the image, respectively.
If a candidate “change component” has a width larger than and/or a height larger than then the component is confirmed as a “change component”.
8.5. Binary Comparison
A further analysis is done to reduce false changes due to reflections, shadows and parallax errors. The images consist of a white background and darker areas where cracks and marks, etc., appear. First, the images are inverted, then the bounding rectangle of each “change candidate” is masked out of both the reference and survey images using the corresponding area in the CM as a mask. The difference in the number of pixels is divided by the total number of mask pixels.
Considering the same example, the difference ratios given in
Figure 15 correspond to the “change candidates” shown in
Figure 16, whose image patches are displayed in
Figure 17. This shows that the difference ratio for component 0, which is the “actual change”, is much larger than for the others. Thus, a threshold is empirically set to filter out the “false changes”. If the ratio is higher than a threshold, this is considered as a “change”; otherwise, it is ignored such that in this case, for example, only “change candidate 0” is considered as a change.
9. Performance Evaluation
To demonstrate the effectiveness of the proposed change detection module, a set of experiments were conducted by simulating different changes such as cracks and other markings on the walls. The experiment covers 60 m of the LHC tunnel. In addition, some markings were also made on the images during post-processing, using a graphical editing software.
For each test scenario, the changed areas are manually marked with a red dot. The change detection output marked with green boxes and indices is analysed and manually compared to the corresponding reference-survey image pair.
An actual “change component” is marked as a true positive (TP). Each actual “change component” that is not detected by the algorithm is added to the false negative (FN) list. On the other hand, an area that is falsely detected as a change, as it does not correspond to any of the actual changes, is noted as a false positive (FP). To quantitatively evaluate the performance of the change detection algorithm, the following metrics are used.
9.1. Evaluation Metrics
The recall is calculated using the true positive rate (TPR), implying the system’s ability to find the changes. The precision is calculated using the positive detection rate (PDR), implying the system’s ability to identify only the actual changes. The
is also calculated to find an optimal blend of both. These metrics are found using:
9.2. Quantitative Analysis
The quantitative results recorded in
Table 1 show that the decision-level fusion by PCA generated a higher precision rate. As the threshold of the final binary comparison was increased from 0.1 to 0.2, the precision value increased from 83.0% to 94.5%. When the majority voting approach was used, precision of 78.8% and 93% was achieved at the same thresholds of 0.1 and 0.2 in the final comparison stage. This implies that the PCA approach distinguished better between actual and nuisance changes.
However, it is also important to evaluate the effectiveness of the algorithm with respect to its ability to find all the data points of interest, in this case, the identified changes. This is given by the recall rate, which had higher values of 83.71% and 81.11% for the majority voting approach with binary comparison stage threshold values of 0.1 and 0.2, respectively. This implies that the majority voting approach could identify more actual changes with fewer misses.
It is beneficial if the algorithm can correctly classify the changes to avoid false alarms; however, it is important that changes due to defects on the tunnel lining are not missed. Hence, a trade-off between precision and recall is essential. This is found by analysing the
, which combines both metrics. As observed in
Table 1, the fusion using a majority voting approach achieved better general performance with respect to the
.
9.3. Qualitative Analysis
In addition to the quantitative results, a qualitative analysis was made on different scenarios with “crack changes”, other defects and also “nuisance changes” caused by varying light conditions and shadows.
In the example presented in
Figure 18, both of the fusion approaches identified the actual changes correctly. However, the majority voting approach gave a more confined bounding box around the “crack change” labelled 1.
Using the reference and survey images in
Figure 19, the change detection algorithm using majority voting correctly identified both of the “crack changes”; however, the connectivity and binary comparison stages following the PCA method incorrectly identified this as a “nuisance change” and, thus, discarded it.
In
Figure 20, another “defect” was simulated on the wall. In this case, both methods correctly identified the change. The final example in
Figure 21 only exhibits “nuisance changes” with respect to the light. Both CMs show white pixels in different areas in the image, implying possible change due to specular highlights, shadows and light changes. However, the CM analysis stage ignored most of these regions except for the small shadow area at the bottom of the image when using PCA-based fusion, generating a “false change”.
Considering both the quantitative and qualitative results, the majority voting approach for the decision-level fusion while using a threshold of 0.2 for the final binary comparison stage is used for the final implementation of the vision-based tunnel lining change detection system.
10. Conclusions
Periodic tunnel structural monitoring and inspection are a necessity. Inspections are predominantly performed through visual observations, which involve looking for structural defects and making sketches for civil engineers to assess them and, in turn, suggest the required maintenance and/or repairs. Associated with this, there are several drawbacks, including personnel exposure to hazardous conditions and outcome subjectivity that is highly dependent on human intervention, which may lead to inaccuracies or misinterpretations. Considering this, a novel tunnel inspection solution to monitor for changes on tunnel linings was proposed. An automatic image data acquisition system integrated on a robotic platform is used to capture tunnel wall images. To alleviate the impact of different light conditions on the change detection algorithm, pre-processing stages were also implemented. These include a shading correction to adjust uneven illumination and highlights localisation to reduce false changes due to flashlight reflections. Subsequently, a new change detection algorithm was developed through a combination of different bi-temporal pixel-based fusion methods and decision-level fusion of change maps. The proposed solution complements current structural health monitoring techniques and provides a better means of tunnel surface documentation.
While providing a step forward, our future work will focus on the improvement of the change detection algorithm and pre-processing stages. Machine learning solutions are producing promising results in the field of computer vision, and such solutions can be added to or replace this change detection algorithm. However, a much larger dataset will be required to train and test such solutions.
Author Contributions
Conceptualization, L.A., C.J.D. and G.V.; methodology, L.A. and C.J.D.; software, L.A.; validation, L.A., C.J.D. and M.D.C.; formal analysis, L.A. and C.J.D.; investigation, L.A. and C.J.D.; resources, M.D.C.; data curation, L.A.; writing—original draft preparation, L.A., C.J.D.; writing—review and editing, all authors; visualization, L.A.; supervision, C.J.D., M.D.C. and G.V.; project administration, G.V.; funding acquisition, M.D.C. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Some or all data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request.
Acknowledgments
The authors would like to thank SITES for providing the ScanTubes® camera system for a demo test in the LHC tunnel. The data collected during this test allowed us to appropriately develop and test our proposed solution.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Liu, X.; Tang, Y.; Chen, M.; Lin, Y.; Huang, X.; Huang, K.; He, Y.; Li, L. Vision-Based Three-Dimensional Reconstruction and Monitoring of Large-Scale Steel Tubular Structures. Adv. Civ. Eng. 2020, 2020, 1236021. [Google Scholar]
- Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision. Robot. Comput. Integr. Manuf. 2019, 59, 36–46. [CrossRef]
- Montero, R.; Victores, J.; Martanez, S.; Jardon, A.; Balaguer, C. Past, present and future of robotic tunnel inspection. Autom. Constr. 2015, 59, 99–112. [Google Scholar] [CrossRef]
- Balaguer, C.; Montero, R.; Victores, J.G.; Martínez, S.; Jardón, A. Towards Fully Automated Tunnel Inspection: A Survey and Future Trends. In Proceedings of the 31st International Symposium on Automation and Robotics in Construction and Mining (ISARC 2014), Sydney, Australia, 9–11 July 2014; pp. 19–33. [Google Scholar]
- Attard, L.; Debono, C.J.; Valentino, G.; Di Castro, M. Tunnel inspection using photogrammetric techniques and image processing: A review. ISPRS J. Photogramm. Remote Sens. 2018, 144, 180–188. [Google Scholar] [CrossRef]
- Neptune, N.; Mothe, J. Automatic Annotation of Change Detection Images. Sensors 2021, 21, 1110. [Google Scholar] [CrossRef] [PubMed]
- Agapiou, A. Multi-Temporal Change Detection Analysis of Vertical Sprawl over Limassol City Centre and Amathus Archaeological Site in Cyprus during 2015–2020 Using the Sentinel-1 Sensor and the Google Earth Engine Platform. Sensors 2021, 21, 1884. [Google Scholar] [CrossRef]
- Lu, D.; Mausel, P.; Brondízio, E.; Moran, E. Change detection techniques. Int. J. Remote Sens. 2004, 25, 2365–2401. [Google Scholar] [CrossRef]
- Radke, R.J.; Andra, S.; Al-Kofahi, O.; Roysam, B. Image change detection algorithms: A systematic survey. IEEE Trans. Image Process. 2005, 14, 294–307. [Google Scholar] [CrossRef]
- Jenkins, M.D.; Buggy, T.; Morison, G. An imaging system for visual inspection and structural condition monitoring of railway tunnels. In Proceedings of the IEEE Workshop on Environmental, Energy, and Structural Monitoring Systems (EESMS), Milan, Italy, 24–25 July 2017; pp. 1–6. [Google Scholar]
- Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Vision-based change detection for inspection of tunnel liners. Autom. Constr. 2018, 91, 142–154. [Google Scholar] [CrossRef]
- Stent, S.; Gherardi, R.; Stenger, B.; Soga, K.; Cipolla, R. An Image-Based System for Change Detection on Tunnel Linings. In Proceedings of the IAPR International Conference on Machine Vision Applications, Kyoto, Japan, 20–23 May 2013; pp. 2–5. [Google Scholar]
- Stent, S.; Gherardi, R.; Stenger, B.; Cipolla, R. Detecting Change for Multi-View, Long-Term Surface Inspection. In Proceedings of the British Mach. Vision Conference (BMVC), Swansea, UK, 7–10 September 2015; pp. 127–139. [Google Scholar]
- Crosilla, F. Procrustes Analysis and Geodetic Sciences. In Geodesy-The Challenge of the 3rd Millennium; Grafarend, E.W., Krumm, F.W., Schwarze, V.S., Eds.; Springer: Berlin/Heidelberg, Germany, 2003; pp. 287–292. [Google Scholar]
- Xu, X.; Yang, H. Vision Measurement of Tunnel Structures with Robust Modelling and Deep Learning Algorithms. Sensors 2020, 20, 4945. [Google Scholar] [CrossRef]
- Ayed, S.B.; Trichili, H.; Alimi, A.M. Data fusion architectures: A survey and comparison. In Proceedings of the 15th International Conference on Intelligent Systems Design and Applications (ISDA), Marrakech, Morocco, 14–16 December 2015; pp. 277–282. [Google Scholar]
- Azimirad, E.; Haddadnia, J.; Izadipour, A. A comprehensive review of the multi-sensor data fusion architectures. J. Theor. Appl. Inf. Technol. 2015, 71, 33–42. [Google Scholar]
- Rajini, K.C.; Roopa, S. A Multi-view Super-Resolution Method with Joint-optimization of Image Fusion and Blind Deblurring. KSII Trans. Internet Inf. Syst. 2018, 12, 2366–2395. [Google Scholar]
- Yan, L.; Fei, L.; Chen, C.; Ye, Z.; Zhu, R. A Multi-View Dense Image Matching Method for High-Resolution Aerial Imagery Based on a Graph Network. Remote Sens. 2016, 8, 799. [Google Scholar] [CrossRef] [Green Version]
- Bovolo, F.; Bruzzone, L. The Time Variable in Data Fusion: A Change Detection Perspective. IEEE Geosci. Remote Sens. Mag. 2015, 3, 8–26. [Google Scholar] [CrossRef]
- Jan, J. Medical Image Processing, Reconstruction and Restoration: Concepts and Methods. CRC Press: Boca Raton, FL, USA, 2006; pp. 481–482. [Google Scholar]
- Adali, T.; Jutten, C.; Hansen, L.K. Multimodal Data Fusion [Scanning the Issue]. Proc. IEEE 2015, 103, 1445–1448. [Google Scholar] [CrossRef]
- Lahat, D.; Adali, T.; Jutten, C. Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef] [Green Version]
- SITES. ScanTubes. Available online: https://www.sites.fr/en/cas-pratique/scantubes-inspection/ (accessed on 10 April 2021).
- Di Castro, M.; Buonocore, L.R.; Ferre, M.; Gilardoni, S.; Losito, R.; Lunghi, G.; Masi, A. A dual arms robotic platform control for navigation, inspection and telemanipulation. In Proceedings of the 16th International Conference on Accelerator and Large Experimental Physics Control Systems (ICALEPCS 2017), Barcelona, Spain, 8–13 October 2017. [Google Scholar]
- Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Image mosaicing of tunnel wall images using high level features. In Proceedings of the 10th International Symposium on Image and Signal Processing and Analysis, Ljubljana, Slovenia, 18–20 September 2017; pp. 141–146. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Comput. and Comput.-Assisted Intervention-MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
- Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Specular highlights detection using a U-Net based deep learning architecture. In Proceedings of the 2020 Fourth International Conference on Multimedia Computing, Networking and Applications, Valencia, Spain, 19–22 October 2020. [Google Scholar]
- İlsever, M.; Ünsalan, C. Pixel-Based Change Detection Methods. In Two-Dimensional Change Detection Methods: Remote Sensing Applications; Springer: London, UK, 2012; pp. 7–21. [Google Scholar]
- Ng, H.; Jargalsaikhan, D.; Tsai, H.; Lin, C. An improved method for image thresholding based on the valley-emphasis method. In Proceedings of the 2013 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, Kaohsiung, Taiwan, 29 October–1 November 2013; pp. 1–4. [Google Scholar]
- Abdelli, A.; Choi, H.J. A four-frames differencing technique for moving objects detection in wide area surveillance. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju, Korea, 13–16 February 2017; pp. 210–214. [Google Scholar]
- Zhang, Y.; Ding, W.; Pan, Z.; Qin, J. Improved Wavelet Threshold for Image De-noising. Front. Neurosci. 2019, 13, 39. [Google Scholar] [CrossRef] [Green Version]
- Han, F.; Liu, B.; Zhu, J.; Zhang, B. Algorithm Design for Edge Detection of High-Speed Moving Target Image under Noisy Environment. Sensors 2019, 19, 343. [Google Scholar] [CrossRef] [Green Version]
- Tsai, D.M.; Huang, T.Y. Automated surface inspection for statistical textures. Image Vis. Comput. 2003, 21, 307–323. [Google Scholar] [CrossRef] [Green Version]
- Zhao, J.; Liu, S.; Wan, J.; Yasir, M.; Li, H. Change Detection Method of High Resolution Remote Sensing Image Based on D-S Evidence Theory Feature Fusion. IEEE Access 2021, 9, 4673–4687. [Google Scholar] [CrossRef]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- GDAL/OGR Contributors. GDAL/OGR Geospatial Data Abstraction Software Library; Open Source Geospatial Foundation: 2020. Available online: https://gdal.org (accessed on 10 April 2021).
Figure 1.
Block diagram of the proposed automatic inspection solution.
Figure 1.
Block diagram of the proposed automatic inspection solution.
Figure 2.
Commercial camera system integrated on the CERNBot. Image capturing system (a), camera rig (b), encoder wheel (c).
Figure 2.
Commercial camera system integrated on the CERNBot. Image capturing system (a), camera rig (b), encoder wheel (c).
Figure 3.
Orthophotos generated from (a) and (b) captured during the demo test.
Figure 3.
Orthophotos generated from (a) and (b) captured during the demo test.
Figure 4.
The original reference (a) and survey (b) images, the difference image (c), the pre-processed reference (d) and survey (e) images, and the difference image of the pre-processed images (f).
Figure 4.
The original reference (a) and survey (b) images, the difference image (c), the pre-processed reference (d) and survey (e) images, and the difference image of the pre-processed images (f).
Figure 5.
Highlight localisation on the reference image (a), survey image (b) and the corresponding highlight mask (c).
Figure 5.
Highlight localisation on the reference image (a), survey image (b) and the corresponding highlight mask (c).
Figure 6.
Change detection in an ideal-world scenario. (a), (b), CM (c) and change (d).
Figure 6.
Change detection in an ideal-world scenario. (a), (b), CM (c) and change (d).
Figure 7.
Image differencing using Gaussian valley emphasis for automatic thresholding.
Figure 7.
Image differencing using Gaussian valley emphasis for automatic thresholding.
Figure 8.
Histogram of normalised from PCA on pre-processed images.
Figure 8.
Histogram of normalised from PCA on pre-processed images.
Figure 9.
Change maps from PCA applied to different images. (a), (b).
Figure 9.
Change maps from PCA applied to different images. (a), (b).
Figure 10.
Change maps from SSIM. Greyscale (a), HSV (b) and corrected (c).
Figure 10.
Change maps from SSIM. Greyscale (a), HSV (b) and corrected (c).
Figure 11.
Diagram of change map fusion by PCA.
Figure 11.
Diagram of change map fusion by PCA.
Figure 12.
Change map fusion by PCA.
Figure 12.
Change map fusion by PCA.
Figure 13.
Change map fusion by majority voting.
Figure 13.
Change map fusion by majority voting.
Figure 14.
Change map analysis process.
Figure 14.
Change map analysis process.
Figure 15.
Difference ratios of “change candidates”.
Figure 15.
Difference ratios of “change candidates”.
Figure 16.
The change candidates corresponding to the example in
Figure 6.
Figure 16.
The change candidates corresponding to the example in
Figure 6.
Figure 17.
Change candidates (a,b) reference and survey patch ‘0’, (c,d) reference and survey patch ‘1’, (e,f) reference and survey patch ‘2’ and (g,h) reference and survey patch ‘3’.
Figure 17.
Change candidates (a,b) reference and survey patch ‘0’, (c,d) reference and survey patch ‘1’, (e,f) reference and survey patch ‘2’ and (g,h) reference and survey patch ‘3’.
Figure 18.
An example showing similar results for both majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 18.
An example showing similar results for both majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 19.
An example showing different detection results from majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 19.
An example showing different detection results from majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 20.
An example showing similar performance of majority voting and PCA solutions on a different simulated defect on the wall. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 20.
An example showing similar performance of majority voting and PCA solutions on a different simulated defect on the wall. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 21.
An example exhibiting lighting changes that are correctly identified as a nuisance and not detected as a change. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 21.
An example exhibiting lighting changes that are correctly identified as a nuisance and not detected as a change. Reference (a), survey (b), majority voting (c) and PCA (d).
Table 1.
Quantitative results from the change detection algorithm.
Table 1.
Quantitative results from the change detection algorithm.
Fusion Method | TH | TP | FP | FN | TPR % | PDR % | F1-Score % |
---|
MV | 0.1 | 149 | 40 | 29 | 83.7 | 78.8 | 81.0 |
MV | 0.2 | 146 | 11 | 34 | 81.1 | 93.0 | 86.7 |
PCA | 0.1 | 137 | 28 | 39 | 77.8 | 83.0 | 80.4 |
PCA | 0.2 | 103 | 6 | 73 | 58.5 | 94.5 | 72.3 |
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).