[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
A Coalitional Distributed Model Predictive Control Perspective for a Cyber-Physical Multi-Agent Application
Next Article in Special Issue
An Automated Pipeline for Dynamic Detection of Sub-Surface Metal Loss Defects across Cold Thermography Images
Previous Article in Journal
Lateral Abdominal Muscles Shear Modulus and Thickness Measurements under Controlled Ultrasound Probe Compression by External Force Sensor: A Comparison and Reliability Study
Previous Article in Special Issue
Deep Learning-Based Object Detection for Unmanned Aerial Systems (UASs)-Based Inspections of Construction Stormwater Practices
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vision-Based Tunnel Lining Health Monitoring via Bi-Temporal Image Comparison and Decision-Level Fusion of Change Maps

1
Department of Communications and Computer Engineering, Faculty of ICT, University of Malta, MSD 2080 Msida, Malta
2
Engineering Department, SMM Group, CERN, CH-1211 Geneva, Switzerland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(12), 4040; https://doi.org/10.3390/s21124040
Submission received: 13 April 2021 / Revised: 18 May 2021 / Accepted: 7 June 2021 / Published: 11 June 2021
Figure 1
<p>Block diagram of the proposed automatic inspection solution.</p> ">
Figure 2
<p>Commercial camera system integrated on the CERNBot. Image capturing system (<b>a</b>), camera rig (<b>b</b>), encoder wheel (<b>c</b>).</p> ">
Figure 3
<p>Orthophotos generated from <math display="inline"><semantics> <mrow> <mi>D</mi> <mi>a</mi> <mi>t</mi> <mi>a</mi> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </semantics></math> (<b>a</b>) and <math display="inline"><semantics> <mrow> <mi>D</mi> <mi>a</mi> <mi>t</mi> <mi>a</mi> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </semantics></math> (<b>b</b>) captured during the demo test.</p> ">
Figure 4
<p>The original reference <math display="inline"><semantics> <msub> <mi>I</mi> <mi>r</mi> </msub> </semantics></math> (<b>a</b>) and survey <math display="inline"><semantics> <msub> <mi>I</mi> <mi>s</mi> </msub> </semantics></math> (<b>b</b>) images, the difference image (<b>c</b>), the pre-processed reference <math display="inline"><semantics> <msub> <mi>C</mi> <mi>r</mi> </msub> </semantics></math> (<b>d</b>) and survey <math display="inline"><semantics> <msub> <mi>C</mi> <mi>s</mi> </msub> </semantics></math> (<b>e</b>) images, and the difference image of the pre-processed images (<b>f</b>).</p> ">
Figure 5
<p>Highlight localisation on the reference image <math display="inline"><semantics> <msub> <mi>I</mi> <mi>r</mi> </msub> </semantics></math> (<b>a</b>), survey image <math display="inline"><semantics> <msub> <mi>I</mi> <mi>s</mi> </msub> </semantics></math> (<b>b</b>) and the corresponding highlight mask (<b>c</b>).</p> ">
Figure 6
<p>Change detection in an ideal-world scenario. <math display="inline"><semantics> <msub> <mi>I</mi> <mi>r</mi> </msub> </semantics></math> (<b>a</b>), <math display="inline"><semantics> <msub> <mi>I</mi> <mi>s</mi> </msub> </semantics></math> (<b>b</b>), CM (<b>c</b>) and change (<b>d</b>).</p> ">
Figure 7
<p>Image differencing using Gaussian valley emphasis for automatic thresholding.</p> ">
Figure 8
<p>Histogram of normalised <math display="inline"><semantics> <msub> <mi>C</mi> <mn>0</mn> </msub> </semantics></math> from PCA on pre-processed images.</p> ">
Figure 9
<p>Change maps from PCA applied to different images. <math display="inline"><semantics> <msub> <mi>C</mi> <mn>1</mn> </msub> </semantics></math> (<b>a</b>), <math display="inline"><semantics> <msub> <mi>C</mi> <mn>0</mn> </msub> </semantics></math> (<b>b</b>).</p> ">
Figure 10
<p>Change maps from SSIM. Greyscale (<b>a</b>), HSV (<b>b</b>) and corrected (<b>c</b>).</p> ">
Figure 11
<p>Diagram of change map fusion by PCA.</p> ">
Figure 12
<p>Change map fusion by PCA.</p> ">
Figure 13
<p>Change map fusion by majority voting.</p> ">
Figure 14
<p>Change map analysis process.</p> ">
Figure 15
<p>Difference ratios of “change candidates”.</p> ">
Figure 16
<p>The change candidates corresponding to the example in <a href="#sensors-21-04040-f006" class="html-fig">Figure 6</a>.</p> ">
Figure 17
<p>Change candidates (<b>a</b>,<b>b</b>) reference and survey patch ‘0’, (<b>c</b>,<b>d</b>) reference and survey patch ‘1’, (<b>e</b>,<b>f</b>) reference and survey patch ‘2’ and (<b>g</b>,<b>h</b>) reference and survey patch ‘3’.</p> ">
Figure 18
<p>An example showing similar results for both majority voting and PCA. Reference (<b>a</b>), survey (<b>b</b>), majority voting (<b>c</b>) and PCA (<b>d</b>).</p> ">
Figure 19
<p>An example showing different detection results from majority voting and PCA. Reference (<b>a</b>), survey (<b>b</b>), majority voting (<b>c</b>) and PCA (<b>d</b>).</p> ">
Figure 19 Cont.
<p>An example showing different detection results from majority voting and PCA. Reference (<b>a</b>), survey (<b>b</b>), majority voting (<b>c</b>) and PCA (<b>d</b>).</p> ">
Figure 20
<p>An example showing similar performance of majority voting and PCA solutions on a different simulated defect on the wall. Reference (<b>a</b>), survey (<b>b</b>), majority voting (<b>c</b>) and PCA (<b>d</b>).</p> ">
Figure 21
<p>An example exhibiting lighting changes that are correctly identified as a nuisance and not detected as a change. Reference (<b>a</b>), survey (<b>b</b>), majority voting (<b>c</b>) and PCA (<b>d</b>).</p> ">
Versions Notes

Abstract

:
Tunnel structural health inspections are predominantly done through periodic visual observations, requiring humans to be physically present on-site, possibly exposing them to hazardous environments. These surveys are subjective (relying on the surveyor experience), time-consuming, and may demand operation shutdown. These issues can be mitigated through accurate automatic monitoring and inspection systems. In this work, we propose a remotely operated machine vision change detection application to improve the structural health monitoring of tunnels. The vision-based sensing system acquires the data from a rig of cameras hosted on a robotic platform that is driven parallel to the tunnel walls. These data are then pre-processed using image processing and deep learning techniques to reduce nuisance changes caused by light variations. Image fusion techniques are then applied to identify the changes occurring in the tunnel structure. Different pixel-based change detection approaches are used to generate temporal change maps. Decision-level fusion methods are then used to combine these change maps to obtain a more reliable detection of the changes that occur between surveys. A quantitative analysis of the results achieved shows that the proposed change detection system achieved a recall value of 81%, a precision value of 93% and an F1-score of 86.7%.

1. Introduction

Tunnel infrastructure shows signs of deterioration over time due to construction defects, ageing, unexpected overloading and natural phenomena, possibly leading to problems in structural integrity. Consequently, periodic inspections of concrete tunnels should be conducted to ensure that they are still healthy and safe. Today, these are predominantly performed through periodic visual observations, looking for structural defects such as cracking, spalling and water leakage to identify possible changes in the infrastructure with respect to a previous survey. Strain gauges, displacement meters and other contact measurement methods can be employed to monitor problematic areas more closely within these structures. To conduct such observations, personnel are required to be physically present in the tunnel. Associated with this, there are several drawbacks, including the human presence in hazardous environments and the financial cost involved to train and hire people to do the inspections. In addition, these inspections require a considerable amount of time to perform, leading to longer operation down-times and thus higher monetary losses. Furthermore, the outcome from these inspections is highly dependent on human subjectivity, leading to possible inaccuracies that result in false and missing change detections.
All this has led to an increase in the demand for robotic systems and remote operations to reduce direct human intervention. One possible solution is to use vision-based sensing monitoring systems that can provide reliable objective results. Hence, a substantial effort can be found in the literature on automating inspections using image processing and machine learning methods to detect and classify cracks, structural deformities, and the presence of water along the tunnel linings. Moreover, the imagery can also be used to provide multiview point clouds to visualise the location of these defects and identify issues with surface deformation [1,2]. This task is not trivial as tunnel environments are characterised by non-uniform illumination and shadows, deformations, lack of features, dirt, stains, and possible occlusion of parts of the walls due to cables, pipes and other servicing equipment.
Whilst defect identification is essential to automate inspection, regular monitoring of tunnel linings can provide a more informative survey to further automate inspection and analysis. In this paper, we present a tunnel inspection application that uses robotics, computer vision and data fusion to monitor for changes on tunnel linings. The main contributions of this work are:
  • Integration of a commercial inspection camera system on a robotic platform;
  • Development of a specular highlight localisation algorithm based on uneven illumination correction and deep learning to remove these artefacts from the images captured by the system;
  • Implementation and analysis of bi-temporal image fusion techniques for image comparison and change-map generation;
  • Implementation and evaluation of two decision-level fusion techniques for robust change detection.
The tunnel used to develop this system is within CERN, the European Organisation for Nuclear Research. The considered tunnel is a 27 km long tunnel lying at around 100 m below the ground, hosting the world’s largest particle accelerator, the Large Hadron Collider (LHC).
The remainder of this article is structured as follows. Section 2 reviews the state-of-the-art with respect to automated tunnel inspection and the techniques used here. The proposed solution is presented in Section 3. Section 4 explains the image acquisition part. In Section 5, pre-processing tasks are described. Bi-temporal image fusion is described in Section 6. Section 7 discusses decision-level fusion in the context of change detection, followed by the change map (CM) analysis process presented in Section 8. A performance evaluation is made in Section 9. A summary and suggestions for future work conclude this article.

2. Background Information

2.1. General Tunnel Inspection

Research on automated health monitoring of tunnel structures has received significant attention in recent years, as recorded in [3,4]. Various solutions that deal with different aspects of automated tunnel inspection were proposed through the use of cost-effective photographic equipment and computer vision. In [5], an extensive survey of works within the whole image-based tunnel inspection spectrum is presented. This includes tunnel profile monitoring, crack and leakage detection, as well as tunnel surface documentation and visualisation.

2.2. Change Detection

Change detection is a well-researched problem in the fields of video surveillance, remote sensing ([6,7]) and medical imaging, amongst others. Reviews of change identification methods are found in [8,9]. However, literature on the detection of changes on tunnel linings is still lacking, possibly due to the challenges encountered in this area. Some of these can be referred to in [10,11,12,13,14]. A system aimed at supporting structural inspectors to monitor the condition of railway tunnels is presented in [10]. An array of cameras with uniform lighting is used to capture the image data that are registered and stitched to allow tunnel inspectors to examine for defects with reference to their location. The solution presented also implements a change detection algorithm to attract attention to areas that are considered important, and the inspectors can refer to previous imagery from previous inspections to assess the evolution of the defect. A three-dimensional model of the structure is also created to better contextualise these defects. This paper does not report on the accuracy of the techniques used. A change detection system that relies on computer vision techniques was presented in [11]. A camera was implemented on a monorail, and the captured images were first processed by a shading correction algorithm before using image mosaicing to correct for offsets between successive inspections. A change detection algorithm was then applied on a survey image and the reference image using image differencing, binary image comparison and optical flow analysis. The solution in [12] registers new images on a three-dimensional surface model to detect and assist visual inspections. The changes are detected using a probabilistic model that considers different feature maps and a geometric prior to reducing the impact of noise and nuisance sources. The system presented in [13] builds panoramas of the tunnel surface and registers images from different inspections. A two-channel convolutional neural network is used to determine anomalous changes between the current survey image patch and the corresponding reference. The work in [14] discusses the application of Procrustes methods in photogrammetry. A method that uses freeform surface modelling and a mask region-based convolutional neural network is presented in [15] to generate 3D tunnel structures and superimpose the position of cracks on the model. Change detection techniques can be applied to measurements completed at different times.
The goal of any change detection algorithm is to detect significant changes between the new measurement and a reference, where the reference is the previous measurement at the exact same location. However, this is not a trivial task as the accuracy of such algorithms can be hindered by apparent intensity changes that result from camera motion and different lighting conditions. Hence, pre-processing steps involving geometric, radiometric adjustments and semantic segmentation are generally required as a primary stage to provide more robust change detection solutions.

2.3. Data Fusion

Data fusion combines data from different methods for increased reliability, higher redundancy and improved identification. Surveys of different fusion architectures are presented in [16,17]. Image fusion is a specific type of data fusion, classified into pixel, feature and decision levels. Image fusion applications can also be categorised by the time, view or modality at which the images are taken. Multiview applications, such as [18,19], fuse images from the same modality but from different viewpoints. Images taken at different times are combined using multi-temporal fusion to detect changes between them or to synthesise images not photographed at a desired time, as in [20,21]. In multi-modal fusion, images coming from different sensors are combined, such as in [22,23].

3. Solution Overview

The proposed solution is illustrated in Figure 1. The acquisition of images in the tunnel is made by a mobile robotic platform. Pre-processing steps involving radiometric adjustments and specular highlight localisation are applied to minimise false change detections due to camera motion, uneven illumination, and different light reflections. Bi-temporal fusion, involving image differencing, principal component analysis (PCA) and a structural similarity index (SSIM) followed by decision-level image fusion is employed at respective stages to achieve change detection.

4. Image Data Acquisition

4.1. Acquisition System

A camera system [24] designed to inspect cylindrical environments was identified on the market. The system is composed of a twelve-unit camera rig, as shown in Figure 2b, two flashlights, an encoder wheel, two batteries and a computer unit with software for camera synchronisation. The twelve industrial cameras mounted on the rig had a 5-megapixel resolution with adapted lenses. During a demo test in the LHC tunnel, this system was integrated on CERNBot [25], one of the readily available robotic platforms at CERN, as shown in Figure 2a. The encoder wheel was attached to the CERNBot, as shown in Figure 2c.

4.2. Dataset

CERNbot was driven parallel to the tunnel wall at a speed of around 0.2 m/s along a section of the LHC tunnel while capturing synchronised images. This image set is referred to as D a t a T 1 . Changes to the structure were then simulated by markings on the wall. These markings had different resolutions, with the fine cracks being 0.5 mm wide. The CERNbot was again driven along the same section and at the same speed capturing D a t a T 2 .
Using these data, 3D models were generated and unwarped into orthophotos using scripts run by the company supplying the same camera system [24]. Using location information from the encoder wheel, orthophotos could be accurately registered, as seen in Figure 3, such that pixel-based change detection (PBCD) techniques could be applied. Each orthophoto is segmented into ten parts along its height, and each image crop covers 0.5 m of the tunnel length. Such images were used for training and testing of the algorithms used in the proposed tunnel lining change detection solution.

5. Image Pre-Processing

The detected changes should be due to new defects or from the evolution of already existing ones. Other changes caused by lighting sources should be identified as “nuisance”, preventing them from being propagated in a change detection pipeline. Hence, pre-processing needs to be completed to cater for this.

5.1. Uneven Illumination Correction

An uneven amount of light falling on different areas causes non-uniformity in images leading to nuisance when comparing images. To adjust the uneven illumination, we use the shading algorithm in [26]. A low-pass filter is applied to the original image using a median filter with a large kernel. The illumination corrected image is obtained through a pixel-wise division of the original image by the low-pass filtered image. As observed in Figure 4c, subtracting the original images generates a difference image full of “white change areas”; however, this is due to uneven illumination. On the other hand, when the images are pre-processed to correct for uneven illumination, their difference image does not have any “white areas” even if there is a change in lighting, as shown in Figure 4f. More details on how this correction mechanism compares to other solutions can be found in [26]. Considering these results, this method proves to be an effective pre-processing method to provide useful images for subsequent processing.

5.2. Specular Highlight Localisation

During image acquisition, flashlights cause reflections on metal racks/pipes present on the wall, resulting in specular highlights in the images. Such highlights are not constant in time or place, leading to false detections when subtracting images to identify changes, as shown in Figure 4. Thus, highlight detection was implemented to localise these regions in the image pair, as displayed in Figure 5. For this, semantic segmentation using a modified U-Net [27] architecture is implemented [28]. Morphological operations and connectivity analysis are then applied to the segmentation images to generate bounding boxes around highlighted areas in the image pair, as illustrated in Figure 5c. Such masks are later fused with the CM to mask out these false change candidates.

6. Bi-Temporal Image Fusion

Multi-temporal fusion combines data from images of the same scene, acquired at different times. Hence, this approach can be used to identify changes in a scene by comparing images. In this scenario, bi-temporal image fusion is applied between the two temporal images; reference and survey. The reference and survey image pair in Figure 6 is used in the explanation of the subsequent methods.
The image fusion method used in this work involves image differencing, PCA and SSIM. The image differencing method finds the difference between the reference and survey images. In an ideal scenario, the result of this will yield an all-zero image except for where a change occurred. However, errors in registration and in the data acquisition process will depart from this ideal case. The PCA is applied to the stacked reference and survey images. The components of the PCA yield the information from both images, the difference between them and other components representing noise. The component representing the difference is used in this case. The PCA is selected as it removes correlated features focusing on the important features, which are the changes in this case. The final metric chosen is the SSIM, which measures the similarity of the reference and survey images. In this case, we are looking at the parts of the images that are not similar, representing change.

6.1. Image Difference

This method has been used extensively for change detection in various applications, including background subtraction for movement detection and remote sensing. As outlined in [29], two images of the same scene taken at separate times t 1 and t 2 are subtracted pixel-wise. After the subtraction, the magnitude of the difference value is compared against a threshold. Pixels with a difference magnitude higher than the pre-defined threshold are classified as “change”, otherwise noted as “no change”. The CM is generated using:
D i f f ( x , y ) = | I ( x , y , t 1 ) I ( x , y , t 2 ) |
C M ( x , y ) = 1 if D i f f ( x , y ) T 0 otherwise
where I ( x , y , t 1 ) is the image at time t 1 , I ( x , y , t 2 ) is the image at time t 2 and T is the threshold on the difference magnitude. This method is simple and requires low computation; however, its accuracy depends on the threshold set. As T is increased, the number of change pixels decreases, implying the elimination of lower difference magnitudes, thus more noise suppression. However, the “valid change” pixels are lost at T 30 in this particular example.
A fixed threshold value cannot satisfy all scenarios; thus, a better approach is to set the threshold automatically depending on the images being compared. The Gaussian valley emphasis (VE) method proposed in [30] is used, generating a CM with only a few “noise changes” while retaining the “crack change”, as observed in Figure 7. The standard deviation of the Gaussian window was empirically set to σ = 5 .
Approaches such as [31] use multiple frames differencing to mitigate issues of sudden illumination changes and ghosting problems. However, in the tunnel environment, illumination changes at the same locations between surveys are limited and the simple differencing was deemed adequate. Other techniques can be applied to reduce the impact of noise, such as wavelets, where [32] proposed a threshold function that is based in the multi-layer wavelet transform. Moreover, [33] uses wavelets to find a trade-off between the removal of noise and extraction of edges in high-speed moving target imagery.

6.2. Principal Component Analysis (PCA)

PCA reduces the dimensionality of a dataset while maintaining the variances. Independent data transformation analysis applies PCA on each of the temporal images separately. The derived principal components are then analysed by applying other change detection techniques, such as image differencing or regression. On the other hand, merged data transformation analysis stacks N temporal images of p channels each, fuses them into a single N × p -channel image and applies PCA on the latter. In this bi-temporal scenario, the merged data approach is used, and the reference and survey images are stacked on each other. The method was investigated in terms of the original RGB images and the pre-processed images, which is the illumination-corrected images. The processing is done on the RGB data as it provides two more dimensions compared to greyscale images and hence more information that helps in detecting changes.
When RGB images ( p = 3 , N = 2 ) are used, the stacked images are merged into a 6-channel image. The first component ( C 0 ), corresponding to the highest eigenvalues, contains most of the information from both images. C 1 represents the difference between temporal images while later components contain noise information. Experimental results show that PCA is scene-dependent, thus comparison between different data is often difficult to interpret using a fixed condition, implying the need for scenario-dependent thresholds. When considering C 1 , the histogram shape is not clearly defined at its tails, making it difficult to find an adaptive threshold pair. When pre-processed images ( p = 1 , N = 2 ) are used, a stacked 2-channel image is generated. From PCA, the first component C 0 represents the difference between temporal images while C 1 contains most of the information from both images. In this case, when considering C 0 , the “crack change” has a high value (white), the “pipe reflections change” has a low value (black) and the rest of the wall has a medium value (grey). This again implies that the histogram contains change elements at both of its tails implying the need for a threshold pair. In this case, however, as observed in Figure 8, the histogram shape follows a Gaussian distribution. To automatically find a threshold pair, the statistical process control (SPC) principle [34] was adopted as it involves binarising an image with a range of pixel values away from the mean pixel where the range is controlled by an input control factor.
The lower and higher thresholds for each image pair under comparison were determined via SPC using:
T l o w = μ c σ
T h i g h = μ + c σ
where μ , σ are the mean and standard deviation of C i , respectively, and c is a constant whose value of three was empirically set for the considered dataset.
C M ( x , y ) = 1 if C i ( x , y ) > T h i g h 1 if C i ( x , y ) < T l o w 0 otherwise
Applying Equation (5) on the C 1 of the original and C 0 of the pre-processed images generated the CMs in Figure 9.

6.3. Structural Similarity (SSIM) Index

Detection of changes between two images can be considered as a comparison of similarities between two images [35]. The larger the difference between the survey image and the reference image is, the smaller the SSIM will be. This metric is based on the human visual system and is in line with what humans perceive as change. SSIM [36] performs different similarity measurements of luminance, contrast and structure, and thereafter combines them to obtain a single value. Considering two image blocks A and B, the SSIM is given by:
S S I M ( A , B ) = ( 2 μ A μ B + c 1 ) ( 2 σ A B + c 2 ) ( μ A 2 + μ B 2 + c 1 ) ( σ A 2 + σ B 2 + c 2 )
where μ A , μ B are the mean and σ A 2 , σ B 2 are the variance of A and B while σ A B is the covariance between A and B. Constants c 1 , c 2 , c 3 are calculated using:
c 1 = ( K 1 L ) 2 , c 2 = ( K 2 L ) 2 , c 3 = c 2 2
where K 1 , K 2 1 , generally K 1 = 0.01 , K 2 = 0.03 as set in [36] and L is the dynamic range of the pixel values ( L = 255 for 8-bit greyscale images). Here, SSIM is used as a PBCD method to generate a CM between a reference and survey image. Therefore, considering the neighbourhood around point ( x , y ) in two temporal images I ( x , y , t 1 ) and I ( x , y , t 2 ) then the SSIM at point ( x , y ) is calculated using Equation (6). The SSIM is normalised to a range of [ 0 , 255 ] and thresholded using:
D ( x , y ) = 1 S S I M ( x , y ) + 1 2
C M ( x , y ) = 1 if D ( x , y ) T 0 otherwise
where D ( x , y ) represents the difference image and T is a constant. A fixed threshold value cannot satisfy all scenarios; thus, the Gaussian VE automatic thresholding method is applied. An investigation of the performance in change detection is done using greyscale images, the V channel in HSV images and pre-processed images corrected for uneven illumination. In general, the best results with minimum noise were obtained using greyscale images, as shown in Figure 10.

7. Decision-Level Fusion

Considering the complementary advantages of the implemented PBCD methods, the generated CMs from image differencing ( C M d i f f ), PCA ( C M P C A ) and SSIM ( C M S S I M ) are fused into a single CM using decision-level fusion methods. Different decision-level fusion methods were implemented; however, the PCA-weighted sum and the majority voting gave the best results and are reported in this section.

7.1. PCA-Weighted Sum

The PCA-based fusion algorithm is illustrated in Figure 11, where P C 0 , P C 1 , and P C 2 are the first three components of one PCA analysis of the stacked CMs. The first three principal components are used as they represent most of the variation in the data. The PCA is applied on the three CMs, namely C M D , C M P C A and C M S S I M . The first principal component, C 0 , measures mainly the impact of the difference CM, the second principal component, C 1 , has a strong association with PCA CM and the third principal component, C 2 , more so represents the SSIM CM. Hence, these C i are used as weights of the CMs to give more importance to the CMs that have better captured the change. The summation of these weighted terms generates the fused CM using:
C M P C A = C M D · C 0 + C M P C A · C 1 + C M S S I M · C 2
As shown in Figure 12, the PCA approach generates few noise pixels while retaining the actual changes; in this case, those belonging to the crack.

7.2. Majority Voting

In the majority voting algorithm, the three different CMs ( C M D , C M P C A and C M s s i m ) cast a unit vote and if at least two of the CMs register a change, then the corresponding pixel in the fused CM is assigned 1 (change), otherwise 0 (no change). Similar to the previous method, this fusion approach generates only a few noise pixels while retaining the actual changes, as shown in Figure 13.

8. Change Map Analysis

At this point, the fused CM may still contain “nuisance change” areas that should not be considered as “changes”. Hence, the CM analysis process illustrated in Figure 14 was developed.

8.1. Specular Highlights Filtering

Fusion between the inverse specular highlight mask image S p e c H ( x , y ) and the final C M M V ( x , y ) is done through an AND operation defined by:
C M f i l t e r e d ( x , y ) = C M M V ( x , y ) S p e c H ( x , y )

8.2. Morphological Operations

The filtered fused C M f i l t e r e d ( x , y ) may contain some small “change areas” coming from image noise and minor registration errors. Here, a morphological closing operation that uses dilation and erosion sequentially, is applied to the fused CM. This joins any change segments by filling gaps, such as in “crack changes”, while at the same time ignores the “noise changes”.

8.3. Connected Components Labelling

Next, connected components labelling with 8-connectivity is used to identify and group neighbouring pixels into “change components”.

8.4. Dimension Filtering

The components are now filtered by their size. A “change component” is only retained if its width and/or height satisfies the corresponding thresholds T W , T H . Using the GDAL library [37], the orthophoto raster scale is obtained, and using the simple proportion principle, the physical dimensions of the segment’s field of view (FoV) are calculated. Using the configurable parameter d m i n representing the minimum dimension for a detected change together with the corresponding image dimension and F o V , the thresholds are calculated using:
T W = d m i n × W F o V W
T H = d m i n × H F o V H
where W and H represent the width and the height of the image, respectively.
If a candidate “change component” has a width larger than T W and/or a height larger than T H then the component is confirmed as a “change component”.

8.5. Binary Comparison

A further analysis is done to reduce false changes due to reflections, shadows and parallax errors. The images consist of a white background and darker areas where cracks and marks, etc., appear. First, the images are inverted, then the bounding rectangle of each “change candidate” is masked out of both the reference and survey images using the corresponding area in the CM as a mask. The difference in the number of pixels is divided by the total number of mask pixels.
Considering the same example, the difference ratios given in Figure 15 correspond to the “change candidates” shown in Figure 16, whose image patches are displayed in Figure 17. This shows that the difference ratio for component 0, which is the “actual change”, is much larger than for the others. Thus, a threshold is empirically set to filter out the “false changes”. If the ratio is higher than a threshold, this is considered as a “change”; otherwise, it is ignored such that in this case, for example, only “change candidate 0” is considered as a change.

9. Performance Evaluation

To demonstrate the effectiveness of the proposed change detection module, a set of experiments were conducted by simulating different changes such as cracks and other markings on the walls. The experiment covers 60 m of the LHC tunnel. In addition, some markings were also made on the images during post-processing, using a graphical editing software.
For each test scenario, the changed areas are manually marked with a red dot. The change detection output marked with green boxes and indices is analysed and manually compared to the corresponding reference-survey image pair.
An actual “change component” is marked as a true positive (TP). Each actual “change component” that is not detected by the algorithm is added to the false negative (FN) list. On the other hand, an area that is falsely detected as a change, as it does not correspond to any of the actual changes, is noted as a false positive (FP). To quantitatively evaluate the performance of the change detection algorithm, the following metrics are used.

9.1. Evaluation Metrics

The recall is calculated using the true positive rate (TPR), implying the system’s ability to find the changes. The precision is calculated using the positive detection rate (PDR), implying the system’s ability to identify only the actual changes. The F 1 s c o r e is also calculated to find an optimal blend of both. These metrics are found using:
T P R ( R e c a l l ) = T P T P + F N × 100 %
P D R ( P r e c i s i o n ) = T P T P + F P × 100 %
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100 %

9.2. Quantitative Analysis

The quantitative results recorded in Table 1 show that the decision-level fusion by PCA generated a higher precision rate. As the threshold of the final binary comparison was increased from 0.1 to 0.2, the precision value increased from 83.0% to 94.5%. When the majority voting approach was used, precision of 78.8% and 93% was achieved at the same thresholds of 0.1 and 0.2 in the final comparison stage. This implies that the PCA approach distinguished better between actual and nuisance changes.
However, it is also important to evaluate the effectiveness of the algorithm with respect to its ability to find all the data points of interest, in this case, the identified changes. This is given by the recall rate, which had higher values of 83.71% and 81.11% for the majority voting approach with binary comparison stage threshold values of 0.1 and 0.2, respectively. This implies that the majority voting approach could identify more actual changes with fewer misses.
It is beneficial if the algorithm can correctly classify the changes to avoid false alarms; however, it is important that changes due to defects on the tunnel lining are not missed. Hence, a trade-off between precision and recall is essential. This is found by analysing the F 1 s c o r e , which combines both metrics. As observed in Table 1, the fusion using a majority voting approach achieved better general performance with respect to the F 1 s c o r e .

9.3. Qualitative Analysis

In addition to the quantitative results, a qualitative analysis was made on different scenarios with “crack changes”, other defects and also “nuisance changes” caused by varying light conditions and shadows.
In the example presented in Figure 18, both of the fusion approaches identified the actual changes correctly. However, the majority voting approach gave a more confined bounding box around the “crack change” labelled 1.
Using the reference and survey images in Figure 19, the change detection algorithm using majority voting correctly identified both of the “crack changes”; however, the connectivity and binary comparison stages following the PCA method incorrectly identified this as a “nuisance change” and, thus, discarded it.
In Figure 20, another “defect” was simulated on the wall. In this case, both methods correctly identified the change. The final example in Figure 21 only exhibits “nuisance changes” with respect to the light. Both CMs show white pixels in different areas in the image, implying possible change due to specular highlights, shadows and light changes. However, the CM analysis stage ignored most of these regions except for the small shadow area at the bottom of the image when using PCA-based fusion, generating a “false change”.
Considering both the quantitative and qualitative results, the majority voting approach for the decision-level fusion while using a threshold of 0.2 for the final binary comparison stage is used for the final implementation of the vision-based tunnel lining change detection system.

10. Conclusions

Periodic tunnel structural monitoring and inspection are a necessity. Inspections are predominantly performed through visual observations, which involve looking for structural defects and making sketches for civil engineers to assess them and, in turn, suggest the required maintenance and/or repairs. Associated with this, there are several drawbacks, including personnel exposure to hazardous conditions and outcome subjectivity that is highly dependent on human intervention, which may lead to inaccuracies or misinterpretations. Considering this, a novel tunnel inspection solution to monitor for changes on tunnel linings was proposed. An automatic image data acquisition system integrated on a robotic platform is used to capture tunnel wall images. To alleviate the impact of different light conditions on the change detection algorithm, pre-processing stages were also implemented. These include a shading correction to adjust uneven illumination and highlights localisation to reduce false changes due to flashlight reflections. Subsequently, a new change detection algorithm was developed through a combination of different bi-temporal pixel-based fusion methods and decision-level fusion of change maps. The proposed solution complements current structural health monitoring techniques and provides a better means of tunnel surface documentation.
While providing a step forward, our future work will focus on the improvement of the change detection algorithm and pre-processing stages. Machine learning solutions are producing promising results in the field of computer vision, and such solutions can be added to or replace this change detection algorithm. However, a much larger dataset will be required to train and test such solutions.

Author Contributions

Conceptualization, L.A., C.J.D. and G.V.; methodology, L.A. and C.J.D.; software, L.A.; validation, L.A., C.J.D. and M.D.C.; formal analysis, L.A. and C.J.D.; investigation, L.A. and C.J.D.; resources, M.D.C.; data curation, L.A.; writing—original draft preparation, L.A., C.J.D.; writing—review and editing, all authors; visualization, L.A.; supervision, C.J.D., M.D.C. and G.V.; project administration, G.V.; funding acquisition, M.D.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Some or all data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

The authors would like to thank SITES for providing the ScanTubes® camera system for a demo test in the LHC tunnel. The data collected during this test allowed us to appropriately develop and test our proposed solution.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, X.; Tang, Y.; Chen, M.; Lin, Y.; Huang, X.; Huang, K.; He, Y.; Li, L. Vision-Based Three-Dimensional Reconstruction and Monitoring of Large-Scale Steel Tubular Structures. Adv. Civ. Eng. 2020, 2020, 1236021. [Google Scholar]
  2. Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision. Robot. Comput. Integr. Manuf. 2019, 59, 36–46. [CrossRef]
  3. Montero, R.; Victores, J.; Martanez, S.; Jardon, A.; Balaguer, C. Past, present and future of robotic tunnel inspection. Autom. Constr. 2015, 59, 99–112. [Google Scholar] [CrossRef]
  4. Balaguer, C.; Montero, R.; Victores, J.G.; Martínez, S.; Jardón, A. Towards Fully Automated Tunnel Inspection: A Survey and Future Trends. In Proceedings of the 31st International Symposium on Automation and Robotics in Construction and Mining (ISARC 2014), Sydney, Australia, 9–11 July 2014; pp. 19–33. [Google Scholar]
  5. Attard, L.; Debono, C.J.; Valentino, G.; Di Castro, M. Tunnel inspection using photogrammetric techniques and image processing: A review. ISPRS J. Photogramm. Remote Sens. 2018, 144, 180–188. [Google Scholar] [CrossRef]
  6. Neptune, N.; Mothe, J. Automatic Annotation of Change Detection Images. Sensors 2021, 21, 1110. [Google Scholar] [CrossRef] [PubMed]
  7. Agapiou, A. Multi-Temporal Change Detection Analysis of Vertical Sprawl over Limassol City Centre and Amathus Archaeological Site in Cyprus during 2015–2020 Using the Sentinel-1 Sensor and the Google Earth Engine Platform. Sensors 2021, 21, 1884. [Google Scholar] [CrossRef]
  8. Lu, D.; Mausel, P.; Brondízio, E.; Moran, E. Change detection techniques. Int. J. Remote Sens. 2004, 25, 2365–2401. [Google Scholar] [CrossRef]
  9. Radke, R.J.; Andra, S.; Al-Kofahi, O.; Roysam, B. Image change detection algorithms: A systematic survey. IEEE Trans. Image Process. 2005, 14, 294–307. [Google Scholar] [CrossRef]
  10. Jenkins, M.D.; Buggy, T.; Morison, G. An imaging system for visual inspection and structural condition monitoring of railway tunnels. In Proceedings of the IEEE Workshop on Environmental, Energy, and Structural Monitoring Systems (EESMS), Milan, Italy, 24–25 July 2017; pp. 1–6. [Google Scholar]
  11. Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Vision-based change detection for inspection of tunnel liners. Autom. Constr. 2018, 91, 142–154. [Google Scholar] [CrossRef]
  12. Stent, S.; Gherardi, R.; Stenger, B.; Soga, K.; Cipolla, R. An Image-Based System for Change Detection on Tunnel Linings. In Proceedings of the IAPR International Conference on Machine Vision Applications, Kyoto, Japan, 20–23 May 2013; pp. 2–5. [Google Scholar]
  13. Stent, S.; Gherardi, R.; Stenger, B.; Cipolla, R. Detecting Change for Multi-View, Long-Term Surface Inspection. In Proceedings of the British Mach. Vision Conference (BMVC), Swansea, UK, 7–10 September 2015; pp. 127–139. [Google Scholar]
  14. Crosilla, F. Procrustes Analysis and Geodetic Sciences. In Geodesy-The Challenge of the 3rd Millennium; Grafarend, E.W., Krumm, F.W., Schwarze, V.S., Eds.; Springer: Berlin/Heidelberg, Germany, 2003; pp. 287–292. [Google Scholar]
  15. Xu, X.; Yang, H. Vision Measurement of Tunnel Structures with Robust Modelling and Deep Learning Algorithms. Sensors 2020, 20, 4945. [Google Scholar] [CrossRef]
  16. Ayed, S.B.; Trichili, H.; Alimi, A.M. Data fusion architectures: A survey and comparison. In Proceedings of the 15th International Conference on Intelligent Systems Design and Applications (ISDA), Marrakech, Morocco, 14–16 December 2015; pp. 277–282. [Google Scholar]
  17. Azimirad, E.; Haddadnia, J.; Izadipour, A. A comprehensive review of the multi-sensor data fusion architectures. J. Theor. Appl. Inf. Technol. 2015, 71, 33–42. [Google Scholar]
  18. Rajini, K.C.; Roopa, S. A Multi-view Super-Resolution Method with Joint-optimization of Image Fusion and Blind Deblurring. KSII Trans. Internet Inf. Syst. 2018, 12, 2366–2395. [Google Scholar]
  19. Yan, L.; Fei, L.; Chen, C.; Ye, Z.; Zhu, R. A Multi-View Dense Image Matching Method for High-Resolution Aerial Imagery Based on a Graph Network. Remote Sens. 2016, 8, 799. [Google Scholar] [CrossRef] [Green Version]
  20. Bovolo, F.; Bruzzone, L. The Time Variable in Data Fusion: A Change Detection Perspective. IEEE Geosci. Remote Sens. Mag. 2015, 3, 8–26. [Google Scholar] [CrossRef]
  21. Jan, J. Medical Image Processing, Reconstruction and Restoration: Concepts and Methods. CRC Press: Boca Raton, FL, USA, 2006; pp. 481–482. [Google Scholar]
  22. Adali, T.; Jutten, C.; Hansen, L.K. Multimodal Data Fusion [Scanning the Issue]. Proc. IEEE 2015, 103, 1445–1448. [Google Scholar] [CrossRef]
  23. Lahat, D.; Adali, T.; Jutten, C. Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef] [Green Version]
  24. SITES. ScanTubes. Available online: https://www.sites.fr/en/cas-pratique/scantubes-inspection/ (accessed on 10 April 2021).
  25. Di Castro, M.; Buonocore, L.R.; Ferre, M.; Gilardoni, S.; Losito, R.; Lunghi, G.; Masi, A. A dual arms robotic platform control for navigation, inspection and telemanipulation. In Proceedings of the 16th International Conference on Accelerator and Large Experimental Physics Control Systems (ICALEPCS 2017), Barcelona, Spain, 8–13 October 2017. [Google Scholar]
  26. Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Image mosaicing of tunnel wall images using high level features. In Proceedings of the 10th International Symposium on Image and Signal Processing and Analysis, Ljubljana, Slovenia, 18–20 September 2017; pp. 141–146. [Google Scholar]
  27. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Comput. and Comput.-Assisted Intervention-MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  28. Attard, L.; Debono, C.J.; Valentino, G.; Castro, M.D. Specular highlights detection using a U-Net based deep learning architecture. In Proceedings of the 2020 Fourth International Conference on Multimedia Computing, Networking and Applications, Valencia, Spain, 19–22 October 2020. [Google Scholar]
  29. İlsever, M.; Ünsalan, C. Pixel-Based Change Detection Methods. In Two-Dimensional Change Detection Methods: Remote Sensing Applications; Springer: London, UK, 2012; pp. 7–21. [Google Scholar]
  30. Ng, H.; Jargalsaikhan, D.; Tsai, H.; Lin, C. An improved method for image thresholding based on the valley-emphasis method. In Proceedings of the 2013 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, Kaohsiung, Taiwan, 29 October–1 November 2013; pp. 1–4. [Google Scholar]
  31. Abdelli, A.; Choi, H.J. A four-frames differencing technique for moving objects detection in wide area surveillance. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju, Korea, 13–16 February 2017; pp. 210–214. [Google Scholar]
  32. Zhang, Y.; Ding, W.; Pan, Z.; Qin, J. Improved Wavelet Threshold for Image De-noising. Front. Neurosci. 2019, 13, 39. [Google Scholar] [CrossRef] [Green Version]
  33. Han, F.; Liu, B.; Zhu, J.; Zhang, B. Algorithm Design for Edge Detection of High-Speed Moving Target Image under Noisy Environment. Sensors 2019, 19, 343. [Google Scholar] [CrossRef] [Green Version]
  34. Tsai, D.M.; Huang, T.Y. Automated surface inspection for statistical textures. Image Vis. Comput. 2003, 21, 307–323. [Google Scholar] [CrossRef] [Green Version]
  35. Zhao, J.; Liu, S.; Wan, J.; Yasir, M.; Li, H. Change Detection Method of High Resolution Remote Sensing Image Based on D-S Evidence Theory Feature Fusion. IEEE Access 2021, 9, 4673–4687. [Google Scholar] [CrossRef]
  36. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. GDAL/OGR Contributors. GDAL/OGR Geospatial Data Abstraction Software Library; Open Source Geospatial Foundation: 2020. Available online: https://gdal.org (accessed on 10 April 2021).
Figure 1. Block diagram of the proposed automatic inspection solution.
Figure 1. Block diagram of the proposed automatic inspection solution.
Sensors 21 04040 g001
Figure 2. Commercial camera system integrated on the CERNBot. Image capturing system (a), camera rig (b), encoder wheel (c).
Figure 2. Commercial camera system integrated on the CERNBot. Image capturing system (a), camera rig (b), encoder wheel (c).
Sensors 21 04040 g002
Figure 3. Orthophotos generated from D a t a T 1 (a) and D a t a T 2 (b) captured during the demo test.
Figure 3. Orthophotos generated from D a t a T 1 (a) and D a t a T 2 (b) captured during the demo test.
Sensors 21 04040 g003
Figure 4. The original reference I r (a) and survey I s (b) images, the difference image (c), the pre-processed reference C r (d) and survey C s (e) images, and the difference image of the pre-processed images (f).
Figure 4. The original reference I r (a) and survey I s (b) images, the difference image (c), the pre-processed reference C r (d) and survey C s (e) images, and the difference image of the pre-processed images (f).
Sensors 21 04040 g004
Figure 5. Highlight localisation on the reference image I r (a), survey image I s (b) and the corresponding highlight mask (c).
Figure 5. Highlight localisation on the reference image I r (a), survey image I s (b) and the corresponding highlight mask (c).
Sensors 21 04040 g005
Figure 6. Change detection in an ideal-world scenario. I r (a), I s (b), CM (c) and change (d).
Figure 6. Change detection in an ideal-world scenario. I r (a), I s (b), CM (c) and change (d).
Sensors 21 04040 g006
Figure 7. Image differencing using Gaussian valley emphasis for automatic thresholding.
Figure 7. Image differencing using Gaussian valley emphasis for automatic thresholding.
Sensors 21 04040 g007
Figure 8. Histogram of normalised C 0 from PCA on pre-processed images.
Figure 8. Histogram of normalised C 0 from PCA on pre-processed images.
Sensors 21 04040 g008
Figure 9. Change maps from PCA applied to different images. C 1 (a), C 0 (b).
Figure 9. Change maps from PCA applied to different images. C 1 (a), C 0 (b).
Sensors 21 04040 g009
Figure 10. Change maps from SSIM. Greyscale (a), HSV (b) and corrected (c).
Figure 10. Change maps from SSIM. Greyscale (a), HSV (b) and corrected (c).
Sensors 21 04040 g010
Figure 11. Diagram of change map fusion by PCA.
Figure 11. Diagram of change map fusion by PCA.
Sensors 21 04040 g011
Figure 12. Change map fusion by PCA.
Figure 12. Change map fusion by PCA.
Sensors 21 04040 g012
Figure 13. Change map fusion by majority voting.
Figure 13. Change map fusion by majority voting.
Sensors 21 04040 g013
Figure 14. Change map analysis process.
Figure 14. Change map analysis process.
Sensors 21 04040 g014
Figure 15. Difference ratios of “change candidates”.
Figure 15. Difference ratios of “change candidates”.
Sensors 21 04040 g015
Figure 16. The change candidates corresponding to the example in Figure 6.
Figure 16. The change candidates corresponding to the example in Figure 6.
Sensors 21 04040 g016
Figure 17. Change candidates (a,b) reference and survey patch ‘0’, (c,d) reference and survey patch ‘1’, (e,f) reference and survey patch ‘2’ and (g,h) reference and survey patch ‘3’.
Figure 17. Change candidates (a,b) reference and survey patch ‘0’, (c,d) reference and survey patch ‘1’, (e,f) reference and survey patch ‘2’ and (g,h) reference and survey patch ‘3’.
Sensors 21 04040 g017
Figure 18. An example showing similar results for both majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 18. An example showing similar results for both majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Sensors 21 04040 g018
Figure 19. An example showing different detection results from majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 19. An example showing different detection results from majority voting and PCA. Reference (a), survey (b), majority voting (c) and PCA (d).
Sensors 21 04040 g019aSensors 21 04040 g019b
Figure 20. An example showing similar performance of majority voting and PCA solutions on a different simulated defect on the wall. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 20. An example showing similar performance of majority voting and PCA solutions on a different simulated defect on the wall. Reference (a), survey (b), majority voting (c) and PCA (d).
Sensors 21 04040 g020
Figure 21. An example exhibiting lighting changes that are correctly identified as a nuisance and not detected as a change. Reference (a), survey (b), majority voting (c) and PCA (d).
Figure 21. An example exhibiting lighting changes that are correctly identified as a nuisance and not detected as a change. Reference (a), survey (b), majority voting (c) and PCA (d).
Sensors 21 04040 g021
Table 1. Quantitative results from the change detection algorithm.
Table 1. Quantitative results from the change detection algorithm.
Fusion MethodTHTPFPFNTPR %PDR %F1-Score %
MV0.1149402983.778.881.0
MV0.2146113481.193.086.7
PCA0.1137283977.883.080.4
PCA0.210367358.594.572.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Attard, L.; Debono, C.J.; Valentino, G.; Di Castro, M. Vision-Based Tunnel Lining Health Monitoring via Bi-Temporal Image Comparison and Decision-Level Fusion of Change Maps. Sensors 2021, 21, 4040. https://doi.org/10.3390/s21124040

AMA Style

Attard L, Debono CJ, Valentino G, Di Castro M. Vision-Based Tunnel Lining Health Monitoring via Bi-Temporal Image Comparison and Decision-Level Fusion of Change Maps. Sensors. 2021; 21(12):4040. https://doi.org/10.3390/s21124040

Chicago/Turabian Style

Attard, Leanne, Carl James Debono, Gianluca Valentino, and Mario Di Castro. 2021. "Vision-Based Tunnel Lining Health Monitoring via Bi-Temporal Image Comparison and Decision-Level Fusion of Change Maps" Sensors 21, no. 12: 4040. https://doi.org/10.3390/s21124040

APA Style

Attard, L., Debono, C. J., Valentino, G., & Di Castro, M. (2021). Vision-Based Tunnel Lining Health Monitoring via Bi-Temporal Image Comparison and Decision-Level Fusion of Change Maps. Sensors, 21(12), 4040. https://doi.org/10.3390/s21124040

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop