Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
Next Article in Journal
FineTea: A Novel Fine-Grained Action Recognition Video Dataset for Tea Ceremony Actions
Previous Article in Journal
Editorial for the Special Issue on “Feature Papers in Section AI in Imaging”
Previous Article in Special Issue
Simultaneous Stereo Matching and Confidence Estimation Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Concrete Crack Detection and Segregation: A Feature Fusion, Crack Isolation, and Explainable AI-Based Approach

by
Reshma Ahmed Swarna
1,2,
Muhammad Minoar Hossain
1,2,
Mst. Rokeya Khatun
2,
Mohammad Motiur Rahman
1 and
Arslan Munir
3,*
1
Department of Computer Science and Engineering, Mawlana Bhashani Science and Technology University, Tangail 1902, Bangladesh
2
Department of Computer Science and Engineering, Bangladesh University, Dhaka 1000, Bangladesh
3
Department of Electrical Engineering and Computer Science, Florida Atlantic University, Boca Raton, FL 33431, USA
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(9), 215; https://doi.org/10.3390/jimaging10090215
Submission received: 13 August 2024 / Revised: 28 August 2024 / Accepted: 30 August 2024 / Published: 31 August 2024
(This article belongs to the Special Issue Image Processing and Computer Vision: Algorithms and Applications)

Abstract

:
Scientific knowledge of image-based crack detection methods is limited in understanding their performance across diverse crack sizes, types, and environmental conditions. Builders and engineers often face difficulties with image resolution, detecting fine cracks, and differentiating between structural and non-structural issues. Enhanced algorithms and analysis techniques are needed for more accurate assessments. Hence, this research aims to generate an intelligent scheme that can recognize the presence of cracks and visualize the percentage of cracks from an image along with an explanation. The proposed method fuses features from concrete surface images through a ResNet-50 convolutional neural network (CNN) and curvelet transform handcrafted (HC) method, optimized by linear discriminant analysis (LDA), and the eXtreme gradient boosting (XGB) classifier then uses these features to recognize cracks. This study evaluates several CNN models, including VGG-16, VGG-19, Inception-V3, and ResNet-50, and various HC techniques, such as wavelet transform, counterlet transform, and curvelet transform for feature extraction. Principal component analysis (PCA) and LDA are assessed for feature optimization. For classification, XGB, random forest (RF), adaptive boosting (AdaBoost), and category boosting (CatBoost) are tested. To isolate and quantify the crack region, this research combines image thresholding, morphological operations, and contour detection with the convex hulls method and forms a novel algorithm. Two explainable AI (XAI) tools, local interpretable model-agnostic explanations (LIMEs) and gradient-weighted class activation mapping++ (Grad-CAM++) are integrated with the proposed method to enhance result clarity. This research introduces a novel feature fusion approach that enhances crack detection accuracy and interpretability. The method demonstrates superior performance by achieving 99.93% and 99.69% accuracy on two existing datasets, outperforming state-of-the-art methods. Additionally, the development of an algorithm for isolating and quantifying crack regions represents a significant advancement in image processing for structural analysis. The proposed approach provides a robust and reliable tool for real-time crack detection and assessment in concrete structures, facilitating timely maintenance and improving structural safety. By offering detailed explanations of the model’s decisions, the research addresses the critical need for transparency in AI applications, thus increasing trust and adoption in engineering practice.

1. Introduction

Concrete is a ubiquitous material in the construction industry, which is valued for its strength and durability. Like all materials, it is susceptible to damage over time, with cracks being a common and critical form of deterioration. Cracks can indicate underlying issues, such as structural weaknesses, material deterioration, and environmental damage, which, if left unaddressed, can lead to severe structural failures and safety hazards for occupants [1,2]. In 2013, a building named Rana Plaza in Bangladesh fell down because of cracks, causing 1134 deaths [3]. Likewise, in 2021, a building in Miami collapsed, mainly due to cracks in its foundation and support columns, resulting in 98 deaths [4]. Many other examples of similar damage exist, not just for buildings but also for dams, roads, and so on [5,6]. So, early detection and precise localization of these cracks are essential for maintaining structural integrity and ensuring safety that allows for timely repairs, preventing further deterioration and reducing repair costs [7]. Traditional methods of crack detection rely on human inspectors, whose assessments can be subjective, inconsistent, and time-consuming, especially for large structures or hard-to-reach areas [8]. To overcome these issues, recent advancements in artificial intelligence (AI) and computer vision offer promising solutions. AI-based crack detection utilizes intelligent mechanisms to analyze images of concrete surfaces, providing automated, accurate, and comprehensive assessments of cracks. These technologies enhance the precision and efficiency of crack detection, minimizing human error and subjectivity, which allows for continuous monitoring without the need for constant human intervention [9,10]. Hence, this research aims to generate an intelligent approach to detect, localize, and explain the cracks in concrete surfaces.
In the past few years, numerous intelligent approaches have emerged for detecting building cracks from images. Akgül [11] developed a fused CNN model by merging two existing architectures known as MobileNetV2 and DenseNet169 for detecting surface cracks in concrete buildings. This model achieved an overall accuracy of 99.87% with fewer features and less complexity compared to other CNN techniques. Additionally, several state-of-the-art (SOTA) CNN models were evaluated in this research. Padsumbiya et al. [12] proposed a CNN-based technique to detect surface cracks within a building from lower-resolution images. This lower-resolution image handling made their method cost-effective and less complex. Additionally, they also detected the crack using an efficient segmentation approach after recognizing the presence of the crack through CNN. Golding et al. [13] evaluated the VGG16 CNN model with different image-processing techniques to identify cracks in concrete structures. The outcomes of this research proved that crack detection did not rely on grayscale or color images. Their results also showed that edge detection and thresholding techniques with CNN reduced the crack recognition ability compared to color images. The best accuracy of this model was 99.54% for color images. To monitor cracks within the building surface structure, Zadeh et al. [14] evaluated various fine-tuned CNN models, including VGG19, ResNet50, InceptionV3, and EfficientNetV2. Their findings demonstrated an impressive accuracy of 99.6% for EfficientNetV2. Regarding detecting cracks in roads and bridges, Kumar et al. [15] generated a CNN model by modifying the LeNet-5 architecture. They assessed their modified LeNet-5 using three concrete crack detection datasets and compared the outcomes with and without employing the principal component analysis (PCA) method. Furthermore, they identified the region of cracks in images. Xu et al. [16] developed a CNN model that included image resolution, multi-scale feature extraction, and complexity minimization to identify the cracks within concrete bridges. Except for pre-training, their CNN provided an overall accuracy of 96.37% for crack identification. Le et al. [17] proposed a CNN model for crack recognition in concrete structures with an overall accuracy of 97.7%. They also demonstrated an integrated framework by merging their intelligent scheme with unmanned aircraft for real-life crack detection. Li and Zhao [18] developed a modified CNN by changing the AlexNet structure for surface crack discovery within the concrete framework. The overall accuracy of this model was 99.09%. This research also integrated their intelligent scheme through a smartphone application to use it in practice. Table 1 summarizes all the relevant works discussed up to this point based on their utilized methods, outcomes, limitations, and future scope. From the analysis of this table, this research enhanced the proposed method to overcome the existing limitations and gaps.
This research utilizes two datasets, each containing images categorized into two groups: crack and non-crack concrete surfaces. These images are preprocessed and proceed through the feature extraction process. As feature extractors, we examine both deep learning (DL) and handcrafted (HC) techniques. From the analysis, we combine the best DL and HC methods to produce a fused feature vector, which is then assessed by feature optimization approaches to identify the most relevant features. A classifier is trained with these relevant features and can recognize the crack and non-crack status of any new concrete surface image. If an image is recognized as containing a crack, our proposed novel algorithm identifies the percentage of the crack within the image and isolates the crack region from other portions of the image. Additionally, two explainable AI (XAI) mechanisms are employed to elucidate the models’ predictions. The key contributions of this research are as follows:
  • Crack recognition: This research utilizes the fusion of features from two techniques: convolutional neural network (CNNs) and curvelet transform, with a feature optimization mechanism to detect the presence of cracks within an image.
  • Crack region segregation: Using image processing operations with the convex hull method, we propose a novel algorithm to separate the crack region and identify the crack percentage from an image.
  • Outcome explanation: We explain the classified images using two popular XAI mechanisms, namely local interpretable model-agnostic explanations (LIMEs) and gradient-weighted class activation mapping++ (Grad-CAM++).
The subsequent sections of this paper are organized as follows: Section 2 details the Materials and Methods used in this research, Section 3 presents the Results and Analysis, and, finally, Section 4 concludes this research.

2. Materials and Methodology

Figure 1 illustrates the research workflow, and Section 2.1 and Section 2.2 elaborate on this research workflow in detail.

2.1. Materials

This research utilizes two datasets for all experiments. The first dataset is the surface crack dataset. This dataset is taken from [19]. This dataset contains images of concrete surfaces, categorized into two types: concrete crack surfaces and non-crack surfaces (i.e., normal concrete surfaces). There are a total of 40,000 images in the dataset, each with a size of 227 × 227 × 3 pixels. Of these, 20,000 images depict cracks, while the remaining 20,000 show non-crack surfaces. Therefore, the dataset is balanced. Table 2 presents samples from this dataset. The second dataset is the bridge crack dataset. The source of this dataset is [20]. This dataset contains images of concrete surfaces from bridges. It includes a total of 6069 images, with 4058 crack images and 2011 non-crack images. The images have a size of 224 × 224 × 3 pixels. Table 2 shows sample images from this dataset.

2.2. Methods

In the preprocessing stage, we resize all images to 224 × 224 × 3 to meet the requirements of the CNN model. We also use this size for the HC feature extraction method to maintain consistency. To improve feature extraction, we apply a median filter to smooth the images. The median filter effectively removes noise while preserving edges, which is necessary for feature extraction [21].
Feature extraction uses techniques to detect and isolate attributes in an image, like edges, textures, shapes, and colors. These features represent the image in a more compact and informative way, helping with image recognition. This research examines both automated (deep learning) and manual (handcrafted) feature extraction methods [22] and fuses these features to identify crack and non-crack images. Figure 2 presents the architecture of the feature extraction process at a glance.
To extract deep features, this research analyzes several CNN models: VGG-16, VGG-19, Inception-V3, and ResNet-50. CNN models automatically extract significant characteristics from images. For both of our datasets, ResNet-50 performs better than the other CNN models in recognizing concrete crack and non-crack surfaces. ResNet-50 is a deep CNN with 50 layers and is part of the ResNet family introduced by [23]. It addresses the challenge of training deep networks by using residual learning. This method employs shortcut connections to bypass one or more layers, thus mitigating the vanishing gradient problem. These connections are called residual blocks. Figure 3 shows the structure of a residual block. The architecture of ResNet-50 includes an initial 7 × 7 convolutional layer, followed by four stages of residual blocks. Each block contains three layers of 1 × 1, 3 × 3, and 1 × 1 convolutions. These stages gradually increase the number of filters, from 64 to 2048. They include batch normalization and ReLU activation functions. The network ends with a global average pooling layer and a fully connected layer, producing output through a softmax function. Since this research uses ResNet-50 as a feature extractor, we consider the output from the layer just before the final fully connected layer to obtain features. Specifically, this is the output of the global average pooling layer, which follows the fourth stage of residual blocks. This global average pooling layer condenses the spatial dimensions of the feature maps into a single 2048-dimensional feature vector for each input image.
The manual feature extraction approach is essentially the HC method. To extract HC features, this research analyzes three techniques: wavelet transform [24], counterlet transform [25], and curvelet transform [26]. Each technique is applied to the concrete surface images, and then features are extracted from the transformed images using a gray-level co-occurrence matrix (GLCM) [27]. For both of our datasets, the curvelet transform performs better than the other HC methods in recognizing concrete crack and non-crack surfaces. Curvelets represent multi-scale image geometric transformation. Curvelets are preferable over other similar techniques since curvelets represent edges, curves, and directionality effectively. This research employs the wrapping-based fast discrete curvelet transform method, as it is the most efficient approach. For an image f[x, y] with height M and width N, if φ[x, y] is the curvelet function and K1 and K2 are the spatial locations of curvelets, then the general expression for the collection of curvelet coefficients is:
C j , θ ,   k 1 ,   k 2 = 0 n < M 0 y < N f [ x , y ]   φ j , θ , k 1 , k 2 x , y
Here, j is the scale and θ is the orientation. For image f[x, y], there exists j x θ number of sub-band images [28]. In our method, we used three scale curvelet transforms with four orientations: 0°, 45°, 90°, and 135°. This means that we have a total of (4 × 3) = 12 sub-band images. GLCM is applied to each of these sub-bands. Then, 13 different features, like contrast, correlation, entropy, etc. [27], are calculated from the GLCM-applied sub-band images.
Based on the outcomes, we select the best CNN and HC techniques to form the fused model. We merge features from ResNet-50 and the curvelet transform to create the final feature vector. This results in a total of 2204 features for any concrete surface image. Of these, 2048 features come from the ResNet-50 model, and the rest come from the curvelet transform.
Feature optimization enhances model performance by reducing feature redundancy and noise, thus improving accuracy and decreasing computational complexity. This process ensures the model focuses on the most relevant and informative features, leading to better predictions and efficiency. We use two popular feature optimization techniques, PCA and linear discriminant analysis (LDA) [29], on the fused features. Both techniques enhance performance, with LDA providing the most efficient outcome. LDA reduces dimensionality by finding a new axis that maximizes the separation between different classes. It projects the data onto a lower-dimensional space while maintaining class separability. LDA computes the mean vectors for each class and the overall mean, then calculates the within-class and between-class scatter matrices. It solves an eigenvalue problem to find the linear combinations of features that best separate the classes. The resulting components are ordered by their ability to discriminate between classes, and the top components are used to reduce the dimensionality.
From the final optimized features of a concrete surface image, the eXtreme gradient boosting (XGB) classifier is used to recognize the crack and non-crack status. This research selects XGB due to its superior outcomes after analyzing four different classifiers: XGB [30], random forest (RF) [31], adaptive boosting (AdaBoost) [32], and category boosting (CatBoost) [33]. The XGB classifier builds an ensemble of decision trees sequentially. Each new tree tries to correct the errors made by the previous trees. It uses a technique called gradient boosting. The algorithm calculates the gradient, which is the difference between the predicted and actual values. The new tree is then trained to minimize this gradient. XGB also uses regularization to prevent overfitting. This means it penalizes more complex models to keep them simple. Additionally, it includes techniques like tree pruning and handling missing values. These features make XGB efficient and accurate in classification tasks.
To make the classification outcome understandable, this research uses two deep explainers: LIME and Grad-CAM++. LIME explains predictions by approximating the model locally with a simpler, interpretable model. When using LIME, it perturbs the input data and observes the changes in the output. It then builds a linear model around the prediction to explain it. LIME provides insights into which features are most important for a specific prediction [34]. Grad-CAM++ generates heatmaps to show which parts of an image influence the model’s decision. It calculates the gradients of the target class concerning the feature maps. It then combines these gradients to produce a weighted map. This map highlights the important regions in the image for the prediction. Grad-CAM++ improves on Grad-CAM by better handling multiple instances of the target object in an image. It provides more precise and detailed visual explanations [35]. LIME and Grad-CAM++ help to interpret and visualize how decisions are made in concrete surface crack recognition.
To localize and identify the exact crack region, this research develops an algorithm, which is presented in Algorithm 1. The algorithm begins by converting the input grayscale image I to a binary image B using a specified threshold T. Morphological operations are applied to B to enhance crack regions. Contours are then detected in the processed binary image M. For each contour, its convex hull Hi is computed. A mask H is generated to isolate crack regions by combining all convex hulls. This mask H is used to extract crack regions R from the original image I. To compute the convex hull from contours, Graham’s scan [36] method is used. Graham’s scan is the most popular technique for convex hulls. It begins by selecting the point with the lowest y-coordinate (and the leftmost if tied) as the starting point. It then sorts all other points based on their polar angle relative to this point. Using a stack, it iteratively adds points to form the convex hull, ensuring that each new point does not create a clockwise turn with the last two points on the stack until all points are processed. In Algorithm 1, the percentage of the image area covered by cracks (Pcrack) is calculated by comparing the area of H to the total area of I. Finally, Pcrack is returned as the output, providing a quantitative measure of crack presence in the image.
Algorithm 1. Crack region isolation using convex hulls
Input:
  • Image I: A grayscale image representing the input image containing cracks.
  • Threshold Value T: A value used to convert the grayscale image to a binary image.
  • Minimum Area Amin: Minimum area threshold for considering a contour as a crack region.
Output:
  • Image I: A grayscale image of the input image containing cracks region with percentage.
Start
1.Convert Image to Grayscale: Let G represent the grayscale image obtained from I.
2.Apply Binary Thresholding: Define a binary image B where each pixel is set to-
B ( x , y ) = 1 , i f   G ( x , y ) T 0 , o t h e r w i s e
3.Enhance Cracks with Morphological Operations: Use morphological operations on B to refine the crack regions. Perform closing and opening operations to smooth and fill gaps in the crack regions.
4.Detect Contours: Identify contours {C1, C2, …, Cn} in the refined binary image M.
5.Calculate Convex Hulls: For each contour Ci, calculate its convex hull Hi
H i = ConvexHull C i
6.Create Convex Hull Mask: Create a mask H where each pixel belongs to one or more convex hulls.
H ( x , y ) = 1 , i f ( x , y )   i s   i n s i d e   a n y   c o n v e x   h u l l   H i 0 , o t h e r w i s e
7.Isolate Crack Regions: Generate an image R by masking I with H. So, R now contains only the crack regions isolated from the original image I.
R ( x , y ) = I ( x , y ) H ( x , y )
8.Calculate Crack Percentage: Determine the percentage of the image area covered by cracks-
P crack   =   Total   number   of   pixels   set   to   1   in   mask   H .   Total   number   of   pixels   in   the   original   image   I × 100
9.Output: Return Pcrack, representing the percentage of the image area covered by cracks.
End

3. Experimental Results and Analysis

This section presents the outcome of this research in detail.

3.1. Criteria for Assessing Performance

To evaluate the performance of the proposed method, we employ 10-fold cross-validation [37] with a training-to-testing ratio of 8:2 on the datasets. The performance metrics used include accuracy, precision, recall, specificity, F1-score, normalized confusion matrix (NCM), and ROC curve [38]. These metrics are detailed below.
  • Accuracy measures the proportion of correctly classified instances.
    A c c u r a c y   = T P + T N T P + T N + F P + F N
    where TP is the number of crack images correctly identified as cracks, TN is the number of non-crack images correctly identified as non-cracks, FP is the number of non-crack images incorrectly identified as cracks, and FN is the number of crack images incorrectly identified as non-cracks from all the testing images.
  • Precision indicates the proportion of true-positive results among all positive predictions.
    P r e c i s i o n   = T P T P + F P
  • Recall measures the ability to correctly identify positive instances.
    R e c a l l   = T P T P + F N
  • Specificity assesses the ability to correctly identify negative instances.
    S p e c i f i c i t y   = T N T N + F P
  • F1-score is the harmonic means of precision and recall.
    F 1 - s c o r e   = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
  • NCM: Tabular form presents the percentage of TP, TN, FP, and FN.
  • ROC curve plots the true-positive rate (recall) against the false-positive rate (1-specificity) at various threshold settings, helping to visualize the trade-off between sensitivity and specificity.

3.2. Outcome of Crack Recognition

Table 3 presents the performance of various techniques for the surface crack dataset. The analysis of Table 3 indicates that the feature extractor ResNet-50 achieves the highest accuracy of 99.38%, surpassing all other individual feature extraction methods. Additionally, HC feature extractors such as wavelet transform, counterlet transform, and curvelet transform perform worse than CNN techniques. Among HC methods, the curvelet transform achieves a peak accuracy of 87.24%, whereas VGG-19, the worst-performing CNN mechanism, achieves 98.36% accuracy. The fusion of the best CNN and HC methods is the combination of ResNet-50 and curvelet transform, which results in an accuracy of 99.63%. With feature optimization using PCA and LDA, this fused mechanism achieves 99.76% and 99.93% accuracy, respectively. Therefore, the fusion of ResNet-50 and curvelet transform with LDA is the ultimate crack recognition approach using the surface crack dataset.
Table 4 delineates the performance metrics of various techniques applied to the bridge crack dataset. Analyzing Table 4 reveals that the ResNet-50 feature extractor attains an apex accuracy of 99.01%, thus outperforming all other standalone feature extraction techniques. In contrast, HC feature extractors, encompassing the wavelet transform, counterlet transform, and curvelet transform, exhibit inferior performance relative to CNN-based techniques. Among HC methods, the curvelet transform registers a maximum accuracy of 89.78%, whereas VGG-16, the least efficacious CNN model, accomplishes an accuracy of 97.53%. The amalgamation of the preeminent CNN and HC methods, that is, the combination of ResNet-50 and curvelet transform, culminates in an accuracy of 99.09%. Subsequent to feature optimization via PCA and LDA, this fused mechanism realizes accuracies of 99.17% and 99.69%, respectively. Consequently, the integration of ResNet-50 and curvelet transform with LDA emerges as the paramount crack recognition strategy for the bridge crack dataset.
This research employs the XGB classifier due to its superior performance compared to other methods, namely RF, AdaBoost, and CatBoost. Figure 4 presents an accuracy comparison of different classifiers using the proposed method (ResNet-50 and curvelet transform with LDA) across both datasets. The analysis of Figure 4 indicates that XGB achieves the highest accuracy, with 99.93% and 99.69% for the surface crack dataset and bridge crack dataset, respectively. AdaBoost holds the second highest results, with 99.44% and 99.51% for the surface crack dataset and bridge crack dataset, respectively. The CatBoost classifier records the third highest accuracy, with 98.99% and 99.42% for the surface crack dataset and bridge crack dataset, respectively. However, the RF classifier yields the poorest performance among the classifiers, with overall accuracies of 99.01% and 98.18% for the surface crack dataset and bridge crack dataset, respectively.
This research uses 10-fold cross-validation to assess the proposed method. This technique is vital for evaluating a model’s robustness and generalizability by dividing the dataset into 10 subsets. The model is trained on nine of these subsets and tested on the remaining one, with this process repeated for each subset. This approach helps to minimize overfitting and provides a more accurate performance measure, as every data point is used for both training and validation. Table 5 and Table 6 show the performance of the proposed method using 10-fold cross-validation for the surface crack dataset and bridge crack dataset, respectively. The analysis of these tables demonstrates that the proposed method exhibits minimal variation and maintains relatively consistent performance.
Figure 5 displays the NCM of the proposed method for both datasets. Figure 5a shows that 99.95% of crack images and 99.90% of non-crack images are correctly identified in the surface crack dataset during testing. Figure 5b indicates that in the bridge crack dataset, 99.75% of crack images and 99.59% of non-crack images are correctly identified during testing. The analysis of Figure 5 also reveals that 0.05% of crack images and 0.10% of non-crack images are misclassified in the surface crack dataset, while 0.25% of crack images and 0.41% of non-crack images are misclassified in the bridge crack dataset.
Figure 6 illustrates the ROC curves for the proposed method applied to two datasets: surface cracks (Figure 6a) and bridge cracks (Figure 6b). The ROC curve is a graphical representation that plots the true-positive rate (TPR) against the false-positive rate (FPR) at various classification thresholds, assessing the model’s ability to distinguish between crack and non-crack images. The diagonal blue line represents a random classifier, which serves as a baseline. In both subfigures, the orange curve is perfectly aligned with the top-left corner of the plot, indicating that the proposed method achieves an AUC (Area Under the Curve) of 1.0 for both datasets. An AUC of 1.0 signifies that the model performs flawlessly, with perfect sensitivity (no false negatives) and specificity (no false positives). This result demonstrates that the proposed method is highly effective and reliable in correctly identifying cracks in both the surface and bridge datasets without any errors, reflecting its robust performance in binary classification tasks.

3.3. Outcome of Crack-Region Segregation

After identifying a crack image using the ResNet-50, curvelet transform, and LDA-based approach, we determine the percentage of the crack in that image using Algorithm 1. Table 7 displays the results of Algorithm 1 for some sample images from both datasets. The outputs in Table 7 demonstrate that our convex hull-driven algorithm not only identifies the percentage of the crack but also isolates the crack region from the other portions of a crack image.

3.4. Outcome Explanation

Table 8 explains the crack images using Grad-CAM++ and LIME for the deep model employed in the proposed method. The outcome of Grad-CAM++ in this table generates more precise and visually detailed heatmaps by considering higher-order derivatives of the loss concerning the feature maps. This makes it particularly effective in highlighting important regions in images where crack features are present. The explanation of LIME in Table 8 highlights regions of the crack image that are most influential in the model’s decision-making process, and this is nothing but the exact crack region.

3.5. Comparative Analysis

The primary objective of this research is to identify the presence of cracks, along with the damage percentage and area, in any concrete structure. We utilize two datasets of concrete structures containing both cracked and normal images. The surface crack dataset pertains to building structures, while the bridge crack dataset pertains to bridge structures. Based on an analysis of several recent techniques, we develop our final model to recognize cracks in images through the fusion of ResNet-50 CNN and curvelet transform-based feature extraction, combined with LDA and XGB classifiers. Our proposed approach effectively identifies cracks in both datasets, performing particularly well with the surface crack dataset. Figure 7 shows a performance comparison of the proposed method in both datasets. Once a crack is identified, the proposed algorithm employing convex hull thresholding, morphological operations, and contour finding mechanisms plots the percentage of the crack in the image. Additionally, this research experiments with various XAI methods to explain the crack recognition process.
Table 9 compares this research with other SOTA methods based on various criteria, such as results, ability to segregate the crack region, crack percentage localization, and outcome explanation capability. The analysis of this table shows that existing methods are mostly developed based on a single dataset, whereas this work uses two different datasets. Decisions based on a single dataset are not sufficient to draw any vital conclusions; hence, the use of two datasets by the proposed method, along with a proper outcome assessment, is a significant contribution.
Table 9 also indicates that the methods by Akgül [11], Padsumbiya et al. [12], Golding et al. [13], Zadeh et al. [14], and Le et al. [17] all use the surface crack dataset, with the best accuracy provided by Akgül [11] at 99.87%. The accuracy of our method for the surface crack dataset is 99.93%, which is better than Akgül’s [11]. For the bridge crack dataset, the best accuracy of 99% is presented by Kumar et al. [15], which is lower than the proposed method’s accuracy of 99.69%. Another method by Li and Zhao [18] uses a private dataset, achieving an accuracy of 99.06%, which is still less than the accuracy of our method. Therefore, in terms of accuracy, the proposed method outperforms all techniques, highlighting another significant contribution.
The analysis in Table 9 shows that while the methods by Padsumbiya et al. [12] and Kumar et al. [15] can segregate the crack region within the image, none of the SOTA methods can detect the percentage of the crack in the image. We developed an algorithm for this purpose. Moreover, none of the techniques in Table 9, except for this research, can explain the outcome through XAI. This research is the first to use the XAI technique with an image-based intelligent crack recognition technique. Therefore, in terms of crack percentage visualization and outcome explanation, the proposed research shows significant contributions.
In addition, some efficient methods that are not fully aligned with this research also offer significant improvements, but the proposed work outperforms these techniques too. For instance, Shang et al. [39] used support vector machine (SVM) and fused CNN Long Short-Term Memory (LSTM) models in signal and image data to detect localized damage in plate structures. They generated the image dataset by converting signal data. Although both SVM and CNN-LSTM achieved 100% accuracy, their work lacks cross-validation and does not provide an explanation for the outcomes. Wu et al. [40] presented an enhanced CNN architecture called MobileNetV2_DeepLabV3 to measure dam crack width from images through segmentation. The model achieves an intersection rate of 83.23%. However, the research does not provide the percentage of the crack or an explanation of the segmentation outcome. In their paper, Wan et al. [41] present various methods for monitoring bridge health. The analysis of these methods reveals a lack of explanatory capability. Ozturk [42] investigates the seismic behavior of two historic temple buildings in the Cappadocia region of Turkey, using dynamic analysis based on ground motion records from recent Turkish earthquakes. The analysis considers the impact of structural walls and highlights that slab discontinuities on the first floors are a major factor in expected structural damage. Significant deformation is observed in the roof domes of the Konakli building, with destructive levels of drift contributing to the anticipated damage. The study cannot ensures any intelligence related to damage recognition. Besides, several other efficient methods [43,44,45,46,47,48,49,50] also provide proficient techniques to identify cracks from images, but none of them is capable of explaining the outcome and capturing the crack percentage.

4. Conclusions

This study presents a novel approach for concrete surface crack detection by integrating deep learning and handcrafted features with a unique convex hull-driven technique for crack percentage isolation. The fusion of ResNet-50 CNN-derived deep features with handcrafted curvelet transform features, followed by optimization using PCA and LDA, and classification through XGBoost, demonstrated superior performance. The model achieved a remarkable accuracy of 99.93% and 99.69% on two distinct datasets, showcasing the effectiveness of the proposed method. Additionally, the incorporation of explainability techniques like LIME and Grad-CAM++ provided deeper insights into the decision-making process of the model, making the predictions more transparent and interpretable. The practical application of this research lies in the automated and accurate detection of cracks in concrete surfaces, which is critical for maintaining structural integrity in civil engineering. The ability to isolate and quantify the crack percentage offers engineers a powerful tool for assessing the extent of damage, thereby enabling more informed decisions regarding maintenance and repairs. This approach not only reduces the need for manual inspections but also minimizes the subjectivity and inconsistency associated with traditional methods, leading to more reliable outcomes in real-world scenarios. For civil engineering professionals, it is recommended to adopt this AI-based crack detection framework as part of regular structural assessments. The method can be integrated into existing infrastructure monitoring systems to enhance the accuracy and reliability of crack detection processes. Additionally, the explainability aspect of the model should be leveraged to ensure that engineers understand the basis of AI predictions, facilitating trust and broader acceptance of AI-driven tools in structural diagnostics. In our future work, we plan to integrate our method with real-time monitoring systems, extending its applicability to diverse structural types and materials. In addition to identifying crack percentages, we will also try to quantify additional parameters of cracks, such as their width and length.

Author Contributions

Conceptualization, R.A.S. and M.M.H.; methodology, R.A.S.; validation, M.M.R. and M.M.R.; formal analysis, M.R.K.; investigation, A.M.; resources, M.M.H.; writing—original draft preparation, R.A.S.; writing—review and editing, A.M. and M.M.R.; visualization, M.R.K.; supervision, M.M.R. and A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source of the datasets is provided in Section 3.1.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kim, B.; Yuvaraj, N.; Preethaa, K.R.S.; Pandian, R.A. Surface crack detection using deep learning with shallow CNN architecture for enhanced computation. Neural Comput. Appl. 2021, 33, 9289–9305. [Google Scholar] [CrossRef]
  2. Islam, M.M.; Hossain, M.B.; Akhtar, M.N.; Moni, M.A.; Hasan, K.F. CNN based on transfer learning models using data augmentation and transformation for detection of concrete crack. Algorithms 2022, 15, 287. [Google Scholar] [CrossRef]
  3. Rana Plaza. (n.d.). Available online: https://cleanclothes.org/campaigns/past/rana-plaza (accessed on 8 June 2024).
  4. Potter, W. Surfside Condo Collapse that Killed 98 Miami Residents Started with Crumbling Pool Deck and Was Exacerbated by Faulty Support Columns in Parking Garage. 8 March 2024. Available online: https://www.dailymail.co.uk/news/article-13174089/champlan-tower-south-florida-condo-partial-collapse-98-dead-faulty-support.html (accessed on 8 June 2024).
  5. Learning from Dam Failures: Analyzing 5 Instructive Examples and Their Lessons. 28 June 2023. Available online: https://asterra.io/resources/dam-failures/ (accessed on 8 June 2024).
  6. Fan, R.; Bocus, M.J.; Zhu, Y.; Jiao, J.; Wang, L.; Ma, F.; Liu, M. Road crack detection using deep convolutional neural network and adaptive thresholding. In Proceedings of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 474–479. [Google Scholar]
  7. Mohan, A.; Poobal, S. Crack detection using image processing: A critical review and analysis. Alex. Eng. J. 2018, 57, 787–798. [Google Scholar] [CrossRef]
  8. Ren, Y.; Huang, J.; Hong, Z.; Lu, W.; Yin, J.; Zou, L.; Shen, X. Image-based concrete crack detection in tunnels using deep fully convolutional networks. Constr. Build. Mater. 2020, 234, 117367. [Google Scholar] [CrossRef]
  9. Luo, X.; Guo, J. AI-Enhanced Algorithm for Structural Health Monitoring—An Image-Based Concrete Crack Detection Method Using Convolutional Neural Networks. Master’s Thesis, Chalmers University of Technology, Göteborg, Sweden, 2021. [Google Scholar]
  10. Munawar, H.S.; Hammad, A.W.; Haddad, A.; Soares, C.A.P.; Waller, S.T. Image-based crack detection methods: A review. Infrastructures 2021, 6, 115. [Google Scholar] [CrossRef]
  11. Akgül, İ. Mobile-DenseNet: Detection of building concrete surface cracks using a new fusion technique based on deep learning. Heliyon 2023, 9. [Google Scholar] [CrossRef] [PubMed]
  12. Padsumbiya, M.; Brahmbhatt, V.; Thakkar, S.P. Automatic crack detection using convolutional neural network. J. Soft Comput. Civ. Eng. 2022, 6, 1–17. [Google Scholar]
  13. Golding, V.P.; Gharineiat, Z.; Munawar, H.S.; Ullah, F. Crack detection in concrete structures using deep learning. Sustainability 2022, 14, 8117. [Google Scholar] [CrossRef]
  14. Shomal Zadeh, S.; Khorshidi, M.; Kooban, F. Concrete Surface Crack Detection with Convolutional-based Deep Learning Models. Int. J. Nov. Res. Civ. Struct. Earth Sci. 2023, 10, 25–35. [Google Scholar] [CrossRef]
  15. Kumar, A.; Kumar, A.; Jha, A.K.; Trivedi, A. Crack detection of structures using deep learning framework. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; pp. 526–533. [Google Scholar]
  16. Xu, H.; Su, X.; Wang, Y.; Cai, H.; Cui, K.; Chen, X. Automatic bridge crack detection using a convolutional neural network. Appl. Sci. 2019, 9, 2867. [Google Scholar] [CrossRef]
  17. Le, T.T.; Nguyen, V.H.; Le, M.V. Development of deep learning model for the recognition of cracks on concrete surfaces. Appl. Comput. Intell. Soft Comput. 2021, 2021, 8858545. [Google Scholar] [CrossRef]
  18. Li, S.; Zhao, X. Image-based concrete crack detection using convolutional neural network and exhaustive search technique. Adv. Civ. Eng. 2019, 2019, 6520620. [Google Scholar] [CrossRef]
  19. Özgenel, Ç.F. Concrete Crack Images for Classification. Mendeley Data, 2. 2019. Available online: https://data.mendeley.com/datasets/5y9wdsg2zt/2 (accessed on 12 August 2024).
  20. Crack-Detection. (n.d.). Available online: https://github.com/tjdxxhy/crack-detection (accessed on 8 June 2024).
  21. Pitas, I.; Venetsanopoulos, A.N.; Pitas, I.; Venetsanopoulos, A.N. Median Filters. Nonlinear Digital Filters: Principles and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1990; pp. 63–116. [Google Scholar]
  22. Salau, A.O.; Jain, S. Feature extraction: A survey of the types, techniques, applications. In Proceedings of the 2019 International Conference on Signal Processing and Communication (ICSC), Noida, India, 7–9 March 2019; pp. 158–164. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Arivazhagan, S.; Ganesan, L. Texture classification using wavelet transform. Pattern Recognit. Lett. 2003, 24, 1513–1521. [Google Scholar] [CrossRef]
  25. Hasan, M.M.; Islam, N.; Rahman, M.M. Gastrointestinal polyp detection through a fusion of contourlet transform and Neural features. J. King Saud. Univ. Comput. Inf. Sci. 2022, 34, 526–533. [Google Scholar] [CrossRef]
  26. Mostafiz, R.; Uddin, M.S.; Jabin, I.; Hossain, M.M.; Rahman, M.M. Automatic brain tumor detection from MRI using curvelet transform and neural features. Int. J. Ambient. Comput. Intell. (IJACI) 2022, 13, 1–18. [Google Scholar] [CrossRef]
  27. Wang, J.S.; Ren, X.D. GLCM based extraction of flame image texture features and KPCA-GLVQ recognition method for rotary kiln combustion working conditions. Int. J. Autom. Comput. 2014, 11, 72–77. [Google Scholar] [CrossRef]
  28. AlZubi, S.; Islam, N.; Abbod, M. Multiresolution analysis using wavelet, ridgelet, and curvelet transforms for medical image segmentation. Int. J. Biomed. Imaging 2011, 2011, 136034. [Google Scholar] [CrossRef]
  29. Hossain, M.M.; Swarna, R.A.; Mostafiz, R.; Shaha, P.; Pinky, L.Y.; Rahman, M.M.; Rahman, W.; Hossain, M.S.; Hossain, M.E.; Iqbal, M.S. Analysis of the performance of feature optimization techniques for the diagnosis of machine learning-based chronic kidney disease. MMach. Learn. Appl. 2022, 9, 100330. [Google Scholar] [CrossRef]
  30. Chen, T. Xgboost: Extreme Gradient Boosting. R Package Version 0.4-2. 2015, Volume 1. Available online: https://cran.ms.unimelb.edu.au/web/packages/xgboost/vignettes/xgboost.pdf (accessed on 12 August 2024).
  31. Rigatti, S.J. Random forest. J. Insur. Med. 2017, 47, 31–39. [Google Scholar] [CrossRef]
  32. Rojas, R. AdaBoost and the Super Bowl of Classifiers a Tutorial Introduction to Adaptive Boosting; Technical Report; Freie University: Berlin, Germany, 2009; Volume 1, pp. 1–6. [Google Scholar]
  33. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. Adv. Neural Inf. Process. Syst. 2018, 31. Available online: https://proceedings.neurips.cc/paper/2018/hash/14491b756b3a51daac41c24863285549-Abstract.html (accessed on 12 August 2024).
  34. Nikam, M.; Ranade, A.; Patel, R.; Dalvi, P.; Karande, A. Explainable Approach for Species Identification using LIME. In Proceedings of the2022 IEEE Bombay Section Signature Conference (IBSSC), Mumbai, India, 8–10 December 2022; pp. 1–6. [Google Scholar]
  35. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 839–847. [Google Scholar]
  36. Sunday, D. The Convex Hull of a 2D Point Set or Polygon. 2006. Available online: http://softsurfer.com/Archive/algorithm (accessed on 8 June 2024).
  37. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  38. Vakili, M.; Ghamsari, M.; Rezaei, M. Performance analysis and comparison of machine and deep learning algorithms for IoT data classification. arXiv 2020, arXiv:2001.09636. [Google Scholar]
  39. Shang, L.; Zhang, Z.; Tang, F.; Cao, Q.; Pan, H.; Lin, Z. Signal Process of Ultrasonic Guided Wave for Damage Detection of Localized Defects in Plates: From Shallow Learning to Deep Learning. J. Data Sci. Intell. Syst. 2023. Available online: http://ojs.bonviewpress.com/index.php/jdsis/article/view/1771 (accessed on 12 August 2024).
  40. Wu, Z.; Tang, Y.; Hong, B.; Liang, B.; Liu, Y. Enhanced Precision in Dam Crack Width Measurement: Leveraging Advanced Lightweight Network Identification for Pixel-Level Accuracy. Int. J. Intell. Syst. 2023, 2023, 9940881. [Google Scholar] [CrossRef]
  41. Wan, S.; Guan, S.; Tang, Y. Advancing bridge structural health monitoring: Insights into knowledge-driven and data-driven approaches. J. Data Sci. Intell. Syst. 2024, 2, 129–140. [Google Scholar] [CrossRef]
  42. Ozturk, B. Seismic behavior of two monumental buildings in historical Cappadocia region of Turkey. Bull. Earthq. Eng. 2017, 15, 3103–3123. [Google Scholar] [CrossRef]
  43. Chen, Y.; Zhu, Z.; Lin, Z.; Zhou, Y. Building surface crack detection using deep learning technology. Buildings 2023, 13, 1814. [Google Scholar] [CrossRef]
  44. Kim, B.C.; Son, B.J. Crack Detection of Concrete Images Using Dilatation and Crack Detection Algorithms. Appl. Sci. 2023, 13, 9238. [Google Scholar] [CrossRef]
  45. Choi, Y.; Park, H.W.; Mi, Y.; Song, S. Crack detection and analysis of concrete structures based on neural network and clustering. Sensors 2024, 24, 1725. [Google Scholar] [CrossRef]
  46. Jin, T.; Zhang, W.; Chen, C.; Chen, B.; Zhuang, Y.; Zhang, H. Deep-Learning-and Unmanned Aerial Vehicle-Based Structural Crack Detection in Concrete. Buildings 2023, 13, 3114. [Google Scholar] [CrossRef]
  47. Khan, M.A.M.; Kee, S.H.; Pathan, A.S.K.; Nahid, A.A. Image Processing Techniques for Concrete Crack Detection: A Scientometrics Literature Review. Remote Sens. 2023, 15, 2400. [Google Scholar] [CrossRef]
  48. Avendaño, J.C.; Leander, J.; Karoumi, R. Image-Based Concrete Crack Detection Method Using the Median Absolute Deviation. Sensors 2024, 24, 2736. [Google Scholar] [CrossRef] [PubMed]
  49. Hacıefendioğlu, K.; Altunışık, A.C.; Abdioğlu, T. Deep Learning-Based Automated Detection of Cracks in Historical Masonry Structures. Buildings 2023, 13, 3113. [Google Scholar] [CrossRef]
  50. Azouz, Z.; Honarvar Shakibaei Asli, B.; Khan, M. Evolution of Crack Analysis in Structures Using Image Processing Technique: A Review. Electronics 2023, 12, 3862. [Google Scholar] [CrossRef]
Figure 1. The workflow of this research.
Figure 1. The workflow of this research.
Jimaging 10 00215 g001
Figure 2. Architecture of the feature extraction technique of this research.
Figure 2. Architecture of the feature extraction technique of this research.
Jimaging 10 00215 g002
Figure 3. A residual block contains a connection that bypasses two layers.
Figure 3. A residual block contains a connection that bypasses two layers.
Jimaging 10 00215 g003
Figure 4. Performance analysis of different classifiers.
Figure 4. Performance analysis of different classifiers.
Jimaging 10 00215 g004
Figure 5. Normalized confusion matrix of the proposed method.
Figure 5. Normalized confusion matrix of the proposed method.
Jimaging 10 00215 g005
Figure 6. ROC curve of the proposed method.
Figure 6. ROC curve of the proposed method.
Jimaging 10 00215 g006
Figure 7. Comparison of the performance of the proposed method in surface crack dataset and bridge crack dataset.
Figure 7. Comparison of the performance of the proposed method in surface crack dataset and bridge crack dataset.
Jimaging 10 00215 g007
Table 1. Summary of relevant prior research.
Table 1. Summary of relevant prior research.
MethodTechniqueAccuracyLimitationFuture Scope
Akgül [11]MobileNetV2 and DenseNet169 based novel CNN99.87%No fold cross-validation and outcome explanationEmploying for images having noise or obstacles
Padsumbiya et al. [12]Four layers sequential CNN model97.8%Lack of comparative outcome analysis and outcome explanationEmploying to multiple surface distresses
Golding et al. [13]VGG16 CNN model99.54%No outcome interpretation and cross-validation. Limited to a particular CNN model onlyEvaluating thresholds of 106 and 101 to images and employing VGC11, VGC19, and AlexNet CNN architectures
Zadeh et al. [14]Fine-tuned EfficientNetV2 CNN model99.6%Single-fold data evaluation and lack of outcome explanation--
Kumar et al. [15]Modified LeNet-5 CNN model99%Inexplicable the causes of crack occurrence and no cross-validation for outcomeEnhancing the model to elucidate the underlying causes of crack occurrence
Xu et al. [16]Lower complex CNN model with high-resolution image managing and multi-scale feature extraction modules96.37%Incapability to draw crack region; lack of outcome evaluation and explanationDeploying the proposed CNN to other convolutional networks as a feature extraction module
Le et al. [17]CNN97.7%Incapability of outcome reasoning and cross-validation as well as lack of model noveltyEmploying for multiple types of cracks; Exploring diverse advanced models
Li and Zhao [18]Modified AlexNet99.06%No outcome interpretation and single-fold data evaluationExploring the model to analyze a broader range of concrete defects across diverse environmental conditions
Table 2. Sample images from the datasets.
Table 2. Sample images from the datasets.
Surface Crack DatasetBridge Crack Dataset
CrackNon-CrackCrackNon-Crack
Jimaging 10 00215 i001Jimaging 10 00215 i002Jimaging 10 00215 i003Jimaging 10 00215 i004
Jimaging 10 00215 i005Jimaging 10 00215 i006Jimaging 10 00215 i007Jimaging 10 00215 i008
Jimaging 10 00215 i009Jimaging 10 00215 i010Jimaging 10 00215 i011Jimaging 10 00215 i012
Table 3. Performance of different techniques for surface crack dataset.
Table 3. Performance of different techniques for surface crack dataset.
TechniqueAccuracyPrecisionRecallSpecificityF1-Score
VGG-1698.5099.1497.5799.2898.35
VGG-1998.3698.3598.0698.6198.20
Inception-V399.0598.8299.0999.0198.96
ResNet-5099.3899.2199.4599.3399.32
Wavelet transform79.2178.2980.1478.3079.20
Counterlet transform83.7882.7585.5781.9884.14
Curvelet transform87.2487.2587.9186.5387.57
ResNet-50 + Curvelet transform99.6399.6799.5399.7299.60
ResNet-50 + Curvelet transform + PCA99.7699.7099.7899.7499.74
ResNet-50 + Curvelet transform + LDA99.9399.9399.9599.9099.94
Table 4. Performance of different techniques in bridge crack dataset.
Table 4. Performance of different techniques in bridge crack dataset.
TechniqueAccuracyPrecisionRecallSpecificityF1-Score
VGG-1697.5398.1696.8698.1997.50
VGG-1998.4397.7099.1697.7198.43
Inception-V398.9299.5098.3699.5098.93
ResNet-5099.0198.5199.5098.5299.00
Wavelet transform84.5187.8081.3687.8884.46
Counterlet transform87.3990.3985.1789.8787.71
Curvelet transform89.7892.9887.5592.3690.18
ResNet-50 + Curvelet transform99.0999.3398.8499.3499.09
ResNet-50 + Curvelet transform + PCA99.1799.5098.8499.5199.17
ResNet-50 + Curvelet transform + LDA99.6999.6699.7599.5999.70
Table 5. Performance of the proposed method evaluated on a fold-by-fold basis for the surface crack dataset.
Table 5. Performance of the proposed method evaluated on a fold-by-fold basis for the surface crack dataset.
FoldAccuracyPrecisionRecallSpecificityF1-Score
fold-199.999.8999.9399.8499.91
fold-299.9899.9810099.9499.99
fold-399.8799.8999.8499.999.86
fold-499.9699.9599.9799.9399.96
fold-599.9699.9699.9899.9399.97
fold-699.9499.9399.9599.9199.94
fold-799.9199.8899.9599.8699.91
fold-899.9599.9599.9599.9499.95
fold-999.9799.9899.9799.9799.97
fold-1099.8899.999.9599.7399.92
Average99.9399.9399.9599.9099.94
Table 6. Performance of the proposed method evaluated on a fold-by-fold basis for bridge crack dataset.
Table 6. Performance of the proposed method evaluated on a fold-by-fold basis for bridge crack dataset.
FoldAccuracyPrecisionRecallSpecificityF1-Score
fold-199.9199.8510099.8199.92
fold-299.8399.8699.8699.7899.86
fold-399.7599.4499.7199.7699.58
fold-499.5899.5399.8898.8999.7
fold-599.5199.8699.3199.7999.58
fold-699.7599.1610099.6599.58
fold-799.4299.8499.0899.8299.46
fold-899.8399.7310099.5699.86
fold-999.6799.6199.8799.3199.74
fold-1099.6799.7499.7499.5499.74
Average99.6999.6699.7599.5999.70
Table 7. Illustration of various stages of crack percentage detection.
Table 7. Illustration of various stages of crack percentage detection.
Sample SourceActual ImageBinary ImageMorphological ImageIsolated Crack Regions
Surface crack datasetJimaging 10 00215 i013Jimaging 10 00215 i014Jimaging 10 00215 i015Jimaging 10 00215 i016
Surface crack datasetJimaging 10 00215 i017Jimaging 10 00215 i018Jimaging 10 00215 i019Jimaging 10 00215 i020
Bridge crack datasetJimaging 10 00215 i021Jimaging 10 00215 i022Jimaging 10 00215 i023Jimaging 10 00215 i024
Bridge crack datasetJimaging 10 00215 i025Jimaging 10 00215 i026Jimaging 10 00215 i027Jimaging 10 00215 i028
Table 8. Explanation of the crack images using XAI techniques.
Table 8. Explanation of the crack images using XAI techniques.
Sample SourceActual ImageGrad-CAM++LIME
Surface crack datasetJimaging 10 00215 i029Jimaging 10 00215 i030Jimaging 10 00215 i031
Surface crack datasetJimaging 10 00215 i032Jimaging 10 00215 i033Jimaging 10 00215 i034
Bridge crack datasetJimaging 10 00215 i035Jimaging 10 00215 i036Jimaging 10 00215 i037
Bridge crack datasetJimaging 10 00215 i038Jimaging 10 00215 i039Jimaging 10 00215 i040
Table 9. Comparison of various methods.
Table 9. Comparison of various methods.
MethodDatasetAccuracyCrack Percentage DetectionCrack Region SegregationUtilization of XAI
Akgül [11]Surface crack dataset99.87%XXX
Padsumbiya et al. [12]Surface crack dataset97.8%XX
Golding et al. [13]Surface crack dataset99.54%XXX
Zadeh et al. [14]Surface crack dataset99.6%XXX
Kumar et al. [15]Bridge crack dataset99%XX
Xu et al. [16]Bridge crack dataset96.37%XXX
Le et al. [17]Surface crack dataset97.7%XXX
Li and Zhao [18]Private dataset of 1455 images of 4160 × 312099.06%XXX
Proposed methodSurface crack dataset99.93%
Bridge crack dataset99.69%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Swarna, R.A.; Hossain, M.M.; Khatun, M.R.; Rahman, M.M.; Munir, A. Concrete Crack Detection and Segregation: A Feature Fusion, Crack Isolation, and Explainable AI-Based Approach. J. Imaging 2024, 10, 215. https://doi.org/10.3390/jimaging10090215

AMA Style

Swarna RA, Hossain MM, Khatun MR, Rahman MM, Munir A. Concrete Crack Detection and Segregation: A Feature Fusion, Crack Isolation, and Explainable AI-Based Approach. Journal of Imaging. 2024; 10(9):215. https://doi.org/10.3390/jimaging10090215

Chicago/Turabian Style

Swarna, Reshma Ahmed, Muhammad Minoar Hossain, Mst. Rokeya Khatun, Mohammad Motiur Rahman, and Arslan Munir. 2024. "Concrete Crack Detection and Segregation: A Feature Fusion, Crack Isolation, and Explainable AI-Based Approach" Journal of Imaging 10, no. 9: 215. https://doi.org/10.3390/jimaging10090215

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop