Detection of Apparent Defects of Underwater Structures in Turbid Waters Based on Polarization Imaging and Deep Learning

LÜ Zong-jie, LI Jun-jie, ZHANG Xue-wu

Journal of Changjiang River Scientific Research Institute ›› 2025, Vol. 42 ›› Issue (9) : 156-166.

PDF(18104 KB)
PDF(18104 KB)
Journal of Changjiang River Scientific Research Institute ›› 2025, Vol. 42 ›› Issue (9) : 156-166. DOI: 10.11988/ckyyb.20240836
Engineering Safety And Disaster Prevention

Detection of Apparent Defects of Underwater Structures in Turbid Waters Based on Polarization Imaging and Deep Learning

Author information +
History +

Abstract

[Objective] In underwater engineering inspection, the turbid shallow water environment severely hinders the performance of machine vision-based methods for detecting surface defects in underwater structures. To address the challenge of defect detection in turbid water, this study proposes a lightweight three-stage underwater defect detection method that integrates polarization imaging and deep learning techniques. A defect detection model, named PCC-YOLOv7, is developed. [Methods] First, polarization imaging technology was combined with a polarization restoration model to analyze the polarization characteristics of light waves. This approach effectively suppressed scattering interference in turbid water, thereby achieving clear imaging of turbid environments and restoring defect images. Consequently, defect details obscured by scattering particles were reconstructed. Second, the CAA-SRGAN (Coordinate Attention ACON-Super Resolution Generative Adversarial Network) model was introduced. By employing an improved attention mechanism and a generative adversarial network structure, super-resolution processing was performed on the restored images. This yielded high-resolution underwater defect images, providing a high-quality data foundation for subsequent precise detection. Finally, a defect detection model based on CBAM-YOLOv7 was established, where the convolutional block attention module (CBAM) was utilized to enhance the network’s focus on defect features. Leveraging the advanced YOLOv7 object detection framework, common underwater structural defects, including cracks, holes, and spalling can be rapidly and accurately identified. These three sub-models worked collaboratively to form a comprehensive detection system. [Results] For image restoration, the polarization restoration model exhibited superior performance in metrics such as image clarity and color fidelity compared to current restoration methods. The CAA-SRGAN model generated images with notable improvements in detail texture preservation and resolution enhancement. The CBAM-YOLOv7 defect detection model achieved higher accuracy in both defect localization and classification. A comprehensive evaluation of the PCC-YOLOv7 defect detection model revealed an average improvement of 33.5% in mean average precision (mAP0.5, mAP0.75, and mAP0.5-0.95). Compared to existing models, PCC-YOLOv7 significantly enhanced defect detection performance in turbid underwater environments, effectively improving both recognition rate and detection efficiency. [Conclusions] The PCC-YOLOv7 defect detection model innovatively integrates polarization imaging technology with deep learning. Through the collaborative operation of three functionally complementary sub-models, it successfully addresses the challenge of detecting surface defects in underwater structures in turbid water. Compared to existing models, the proposed model demonstrates enhanced adaptability to turbid underwater detection scenarios. It enables stable and efficient detection of surface defects in underwater structures under complex turbid conditions, providing a practical technical solution for the safety assessment and maintenance of underwater structures. Future work may focus on further optimizing the model structure and extending its application to more underwater scenarios.

Key words

turbid water / underwater structure / defect detection / polarization imaging / deep learning / super-resolution reconstruction

Cite this article

Download Citations
LÜ Zong-jie , LI Jun-jie , ZHANG Xue-wu. Detection of Apparent Defects of Underwater Structures in Turbid Waters Based on Polarization Imaging and Deep Learning[J]. Journal of Changjiang River Scientific Research Institute. 2025, 42(9): 156-166 https://doi.org/10.11988/ckyyb.20240836

References

[1]
钮新强, 谭界雄, 田金章. 混凝土面板堆石坝病害特点及其除险加固[J]. 人民长江, 2016, 47(13):1-5.
(NIU Xin-qiang, TAN Jie-xiong, TIAN Jin-zhang. Analysis on CFRD Defect’s Characteristics and Its Reinforcement[J]. Yangtze River, 2016, 47(13):1-5. (in Chinese))
[2]
盛金保, 厉丹丹, 蔡荨, 等. 大坝风险评估与管理关键技术研究进展[J]. 中国科学:技术科学, 2018, 48(10):1057-1067.
(SHENG Jin-bao, LI Dan-dan, CAI Qian, et al. Research Progress and Its Practice of Key Techniques for Dam Risk Assessment and Management[J]. Scientia Sinica (Technologica), 2018, 48(10): 1057-1067. (in Chinese))
[3]
陈永灿, 陈嘉杰, 王皓冉, 等. 大直径长引水隧洞水下检测机器人系统关键技术[J]. 清华大学学报(自然科学版), 2023, 63(7): 1015-1031.
(CHEN Yong-can, CHEN Jia-jie, WANG Hao-ran, et al. Key Technology of Underwater Inspection Robot System for Large Diameter and Long Headrace Tunnel[J]. Journal of Tsinghua University (Science and Technology), 2023, 63(7): 1015-1031. (in Chinese))
[4]
陶显, 侯伟, 徐德. 基于深度学习的表面缺陷检测方法综述[J]. 自动化学报, 2021, 47(5):1017-1034.
(TAO Xian, HOU Wei, XU De. A Survey of Surface Defect Detection Methods Based on Deep Learning[J]. Acta Automatica Sinica, 2021, 47(5):1017-1034. (in Chinese))
[5]
MCGLAMERY B L. A Computer Modelfor Underwater Camera Systems[C]// Proceedings of the SPIE, Volume 208. Doi: 10.1117/12.958279.
[6]
SCHECHNER Y Y, KARPEL N. Recovery of Underwater Visibility and Structure by Polarization Analysis[J]. IEEE Journal of Oceanic Engineering, 2006, 30(3): 570-587.
[7]
LEDIG C, THEIS L, HUSZÁR F, et al. Photo-realistic Single Image Super-resolution Using a Generative Adversarial Network[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). July 21-26, 2017, Honolulu, HI, USA. New York: IEEE Press, 2017: 105-114.
[8]
LIM B, SON S, KIM H, et al. Enhanced Deep Residual Networks for Single Image Super-resolution[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). July 21-26, 2017, Honolulu, HI, USA. New York: IEEE Press, 2017: 1132-1140.
[9]
HOU Q, ZHOU D, FENG J. Coordinate Attention for Efficient Mobile Network Design[C]// 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). June 20-25, 2021. Nashville, TN, USA. IEEE, 2021: 13708-13717.
[10]
李培育, 张雅丽. 基于改进SRGAN模型的人脸图像超分辨率重建[J]. 计算机工程, 2023, 49(4):199-205.
Abstract
传统生成对抗网络模型重建人脸图像时出现过多失真,难以在减少失真的情况下有效提高人脸图像真实感。针对该问题,在生成对抗网络SRGAN模型的基础上,提出一种改进的人脸图像超分辨率重建方法。为提高重建像素点与周围像素点的相关性,将双注意力机制模块嵌入到SRGAN模型的生成器和判别器中,在空间域和通道域中获取更精准的特征依赖关系。同时应用自适应激活函数ACON取代原SRGAN网络中的激活函数,通过动态学习ACON激活函数参数为每个神经元设计不同激活形式,从而提高网络特征表达能力。使用改进SRGAN的人脸图像超分辨率重建算法在CelebA测试集上进行重建实验,结果表明:该算法较原算法PSNR值提高0.675 dB,SSIM值提高0.016,LPIPS值优化0.036,有效减少了重建人脸图像中眼睛等重点部位的失真情况;与其他非生成对抗网络的主流算法相比,LPIPS值最低优化0.107,最高优化0.205,有效提高了重建人脸图像的真实感。
(LI Pei-yu, ZHANG Ya-li. Face Image Super-resolution Reconstruction Based on Improved SRGAN Model[J]. Computer Engineering, 2023, 49(4):199-205. (in Chinese))
Because the traditional generation countermeasure network model exhibits excessive distortion when reconstructing the face image, it is difficult to effectively improve the realism of the face image.Under the condition of reducing the distortion, an improved super-resolution reconstruction method of face images based on SRGAN model is proposed.To increase the correlation between the reconstructed and surrounding pixels, the Dual Attention(DA) mechanism module is embedded in the generator and discriminator of the SRGAN model to obtain more accurate feature dependency in the spatial and channel domains.An adaptive activation function, termed ACON, is used to replace the original activation function in an SRGAN network, and different activation forms are designed for each neuron by dynamically learning the parameters of the ACON activation function, to improve the ability of network feature expression.The reconstruction experiment is performed on the CelebA using the improved SRGAN face image reconstruction algorithm.The results show that compared with the original algorithm, the PSNR value increases by 0.675 dB, the SSIM value increases by 0.016, and the LPIPS value is optimized by 0.036, effectively reducing the distortion of key parts, such as the eyes, in the reconstructed face image.Compared with other mainstream algorithms that do not generate countermeasure networks, the LPIPS value is optimized by 0.107 at the lowest and 0.205 at the highest, effectively improving the realism of the reconstructed face image.
[11]
WOO S, PARK J, LEE J Y, et al. CBAM: Convolutional Block Attention Module[C]// Proceedings of the 15th European Conference on Computer Vision-ECCV 2018. Cham: Springer International Publishing. Munich, Germany, September 8-14, 2018: 3-19.
[12]
ZUIDERVELD K. Contrast Limited Adaptive Histogram Equalization[M]//HECKBERT P S. Graphics Gems IV. Amsterdam, The Netherlands: Elsevier, 1994: 474-485.
[13]
ZHANG W, ZHUANG P, SUN H H, et al. Underwater Image Enhancement via Minimal Color Loss and Locally Adaptive Contrast Enhancement[J]. IEEE Transactions on Image Processing, 2022, 31: 3997-4010.
[14]
AGAIAN S S, PANETTA K, GRIGORYAN A M. Transform-based Image Enhancement Algorithms with Performance Measure[J]. IEEE Transactions on Image Processing, 2001, 10(3): 367-382.
This paper presents a new class of the "frequency domain"-based signal/image enhancement algorithms including magnitude reduction, log-magnitude reduction, iterative magnitude and a log-reduction zonal magnitude technique. These algorithms are described and applied for detection and visualization of objects within an image. The new technique is based on the so-called sequency ordered orthogonal transforms, which include the well-known Fourier, Hartley, cosine, and Hadamard transforms, as well as new enhancement parametric operators. A wide range of image characteristics can be obtained from a single transform, by varying the parameters of the operators. We also introduce a quantifying method to measure signal/image enhancement called EME. This helps choose the best parameters and transform for each enhancement. A number of experimental results are presented to illustrate the performance of the proposed algorithms.
[15]
YANG M, SOWMYA A. An Underwater Color Image Quality Evaluation Metric[J]. IEEE Transactions on Image Processing, 2015, 24(12): 6062-6071.
Quality evaluation of underwater images is a key goal of underwater video image retrieval and intelligent processing. To date, no metric has been proposed for underwater color image quality evaluation (UCIQE). The special absorption and scattering characteristics of the water medium do not allow direct application of natural color image quality metrics especially to different underwater environments. In this paper, subjective testing for underwater image quality has been organized. The statistical distribution of the underwater image pixels in the CIELab color space related to subjective evaluation indicates the sharpness and colorful factors correlate well with subjective image quality perception. Based on these, a new UCIQE metric, which is a linear combination of chroma, saturation, and contrast, is proposed to quantify the non-uniform color cast, blurring, and low-contrast that characterize underwater engineering and monitoring images. Experiments are conducted to illustrate the performance of the proposed UCIQE metric and its capability to measure the underwater image enhancement results. They show that the proposed metric has comparable performance to the leading natural color image quality metrics and the underwater grayscale image quality metrics available in the literature, and can predict with higher accuracy the relative amount of degradation with similar image content in underwater environments. Importantly, UCIQE is a simple and fast solution for real-time underwater video processing. The effectiveness of the presented measure is also demonstrated by subjective evaluation. The results show better correlation between the UCIQE and the subjective mean opinion score.
[16]
PANETTA K, GAO C, AGAIAN S. Human-visual-system-inspired Underwater Image Quality Measures[J]. IEEE Journal of Oceanic Engineering, 2016, 41(3): 541-551.
PDF(18104 KB)

Accesses

Citation

Detail

Sections
Recommended

/