Cracking is a common pavement distress that would cause further severe problems if not repaired timely, which means that it is important to accurately extract the information of pavement cracks through detection and segmentation. Automated pavement crack detection and segmentation using deep learning are more efficient and accurate than conventional methods, which could be further improved. While many existing studies have utilized deep learning in pavement crack segmentation, which segments cracks from non-crack regions, few studies have taken the exact pavement crack detection into account, which identifies cracks in the images from other objects. A two-step pavement crack detection and segmentation method based on convolutional neural network was proposed in this paper. An automated pavement crack detection algorithm was developed using the modified You Only Look Once 3rd version in the first step. The proposed crack segmentation method in the second step was based on the modified U-Net, whose encoder was replaced with a pre-trained ResNet-34 and the upsample part was added with spatial and channel squeeze and excitation (SCSE) modules. Proposed method combines pavement crack detection and segmentation together, so that the detected cracks from the first step are segmented in the second step to improve the accuracy. A dataset of pavement crack images in different circumstances were also built for the study. The F1 score of proposed crack detection and segmentation methods are 90.58% and 95.75%, respectively, which are higher than other state-of-the-art methods. Compared with existing one-step pavement crack detection or segmentation methods, proposed two-step method showed advantages of accuracy.
Automated pavement crack image segmentation is challenging because of inherent irregular patterns, lighting conditions, and noise in images. Conventional approaches require a substantial amount of feature engineering to differentiate crack regions from non-affected regions. In this paper, we propose a deep learning technique based on a convolutional neural network to perform segmentation tasks on pavement crack images. Our approach requires minimal feature engineering compared to other machine learning techniques. We propose a U-Net-based network architecture in which we replace the encoder with a pretrained ResNet-34 neural network. We use a "one-cycle" training schedule based on cyclical learning rates to speed up the convergence. Our method achieves an F1 score of 96% on the CFD dataset and 73% on the Crack500 dataset, outperforming other algorithms tested on these datasets. We perform ablation studies on various techniques that helped us get marginal performance boosts, i.e., the addition of spatial and channel squeeze and excitation (SCSE) modules, training with gradually increasing image sizes, and training various neural network layers with different learning rates. INDEX TERMS Convolutional neural network, deep learning, fully convolutional network, pavement crack segmentation, U-Net.
Due to medium scattering, absorption, and complex light interactions, capturing objects from the underwater environment has always been a difficult task. Single-pixel imaging (SPI) is an efficient imaging approach that can obtain spatial object information under low-light conditions. In this paper, we propose a single-pixel object inspection system for the underwater environment based on compressive sensing super-resolution convolutional neural network (CS-SRCNN). With the CS-SRCNN algorithm, image reconstruction can be achieved with 30% of the total pixels in the image. We also investigate the impact of compression ratios on underwater object SPI reconstruction performance. In addition, we analyzed the effect of peak signal to noise ratio (PSNR) and structural similarity index (SSIM) to determine the image quality of the reconstructed image. Our work is compared to the SPI system and SRCNN method to demonstrate its efficiency in capturing object results from an underwater environment. The PSNR and SSIM of the proposed method have increased to 35.44% and 73.07%, respectively. This work provides new insight into SPI applications and creates a better alternative for underwater optical object imaging to achieve good quality.
Automated pavement crack segmentation is challenging due to the random shape of cracks, complex background textures and the presence of miscellaneous objects. In this paper, we implemented a Self-Guided Attention Refinement module and incorporated it on top of a Feature Pyramid Network (FPN) to model long-range contextual information. The module uses multi-scale features integrated from different layers in the FPN to refine the features at each layer of the FPN using a self-attention mechanism. The module enables the earlier layers and deeper layers of FPN to suppress noise and increase the crack details, respectively. The proposed network obtains an F1 score of 79.43% and 96.19% on the Crack500 and CFD datasets, respectively. Furthermore, the network also generalizes better than other state-of-the-art methods when tested on uncropped Crack500 and field images using the weights trained on CFD. In addition, ablation tests using the Crack500 dataset are conducted. The Self-Guided Attention Refinement module increases the average F1 score and recall by 0.6% and 0.8% while roughly maintaining the average precision. From the ablation test, the inclusion of the Self-Guided Attention Refinement module allows the network to segment finer and more continuous cracks. Then, the module is incorporated on PANet, DeepLab v3+ and U-Net to verify the improvements made to FPN. The results show that the module improves the F1 score, precision and recall compared to the absence of it. Moreover, the Self-Guided Attention Refinement Module is compared with the Self-Adaptive Sparse Transform Module (SASTM). The results show that the Self-Guided Attention Refinement Module provides a more consistent improvement.
The detection and reconstruction of transparent objects have remained challenging due to the absence of their features and variations in the local features with variations in illumination. In this paper, both compressive sensing (CS) and super-resolution convolutional neural network (SRCNN) techniques are combined to capture transparent objects. With the proposed method, the transparent object’s details are extracted accurately using a single pixel detector during the surface reconstruction. The resultant images obtained from the experimental setup are low in quality due to speckles and deformations on the object. However, the implemented SRCNN algorithm has obviated the mentioned drawbacks and reconstructed images visually plausibly. The developed algorithm locates the deformities in the resultant images and improves the image quality. Additionally, the inclusion of compressive sensing minimizes the measurements required for reconstruction, thereby reducing image post-processing and hardware requirements during network training. The result obtained indicates that the visual quality of the reconstructed images has increased from a structural similarity index (SSIM) value of 0.2 to 0.53. In this work, we demonstrate the efficiency of the proposed method in imaging and reconstructing transparent objects with the application of a compressive single pixel imaging technique and improving the image quality to a satisfactory level using the SRCNN algorithm.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.