Hostname: page-component-77c78cf97d-7rbh8 Total loading time: 0 Render date: 2026-05-04T13:51:09.338Z Has data issue: false hasContentIssue false

Continuous gradient fusion class activation mapping: segmentation of laser-induced damage on large-aperture optics in dark-field images

Published online by Cambridge University Press:  20 November 2023

Yueyue Han
Affiliation:
School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin, China Research Center of Laser Fusion, China Academy of Engineering Physics, Mianyang, China
Yingyan Huang
Affiliation:
School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin, China
Hangcheng Dong
Affiliation:
School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin, China
Fengdong Chen*
Affiliation:
School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin, China
Fa Zeng
Affiliation:
Research Center of Laser Fusion, China Academy of Engineering Physics, Mianyang, China
Zhitao Peng
Affiliation:
Research Center of Laser Fusion, China Academy of Engineering Physics, Mianyang, China
Qihua Zhu
Affiliation:
Research Center of Laser Fusion, China Academy of Engineering Physics, Mianyang, China
Guodong Liu*
Affiliation:
School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin, China
*
Correspondence to: Fengdong Chen and Guodong Liu, School of Instrumentation Science and Engineering, Harbin Institute of Technology, No. 92 Xidazhi Street, Harbin 150001, China. Email: chenfd@hit.edu.cn (F. Chen); lgd@hit.edu.cn (G. Liu)
Correspondence to: Fengdong Chen and Guodong Liu, School of Instrumentation Science and Engineering, Harbin Institute of Technology, No. 92 Xidazhi Street, Harbin 150001, China. Email: chenfd@hit.edu.cn (F. Chen); lgd@hit.edu.cn (G. Liu)

Abstract

Segmenting dark-field images of laser-induced damage on large-aperture optics in high-power laser facilities is challenged by complicated damage morphology, uneven illumination and stray light interference. Fully supervised semantic segmentation algorithms have achieved state-of-the-art performance but rely on a large number of pixel-level labels, which are time-consuming and labor-consuming to produce. LayerCAM, an advanced weakly supervised semantic segmentation algorithm, can generate pixel-accurate results using only image-level labels, but its scattered and partially underactivated class activation regions degrade segmentation performance. In this paper, we propose a weakly supervised semantic segmentation method, continuous gradient class activation mapping (CAM) and its nonlinear multiscale fusion (continuous gradient fusion CAM). The method redesigns backpropagating gradients and nonlinearly activates multiscale fused heatmaps to generate more fine-grained class activation maps with an appropriate activation degree for different damage site sizes. Experiments on our dataset show that the proposed method can achieve segmentation performance comparable to that of fully supervised algorithms.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial licence (https://creativecommons.org/licenses/by-nc/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use.
Copyright
© The Author(s), 2023. Published by Cambridge University Press in association with Chinese Laser Press
Figure 0

Figure 1 Schematic diagram of the methodology for online capturing images of optics (FODI images) by the FODI system.

Figure 1

Figure 2 An example of the FODI image. (a)–(c) Images of stray light interference. (d), (e) Images of large damage sites. (f), (g) Images of weak damage sites.

Figure 2

Figure 3 The process of class activation mapping.

Figure 3

Figure 4 Examples of class activation maps generated by Grad-CAM and LayerCAM on FODI images. The red arrows point to scattered activated regions. The yellow arrows point to underactivated regions.

Figure 4

Figure 5 The pipeline of CG-Fusion CAM.

Figure 5

Figure 6 The class activation maps of LayerCAM from different stages. The red box shows the feature and gradient maps of some channels from Stage 5.

Figure 6

Figure 7 The original and CG-CAM methods of backpropagating gradients from the max-pooling layer to the convolution layer.

Figure 7

Figure 8 Comparison of LayerCAM and CG-CAM results from different stages. (a), (d) Feature maps. (b), (e) Class activation gradient maps. (c), (f) Channel class activation maps.

Figure 8

Figure 9 Results of LayerCAM from each convolutional layer.

Figure 9

Figure 10 Comparison of the multiscale fusion effect from different stages of CG-CAM. The red arrows point to blurred boundaries.

Figure 10

Figure 11 Examples of typical samples. (a) Background class samples. (b) Damage class samples. (c) Manually produced damage class samples.

Figure 11

Table 1 The classification performance of the VGG-16 model on our dataset.

Figure 12

Table 2 Comparison of baselines and our method under various evaluation metrics.

Figure 13

Figure 12 Comparison of the class activation maps and segmentation results between the baselines and our method. The green areas are the false positive segmentation results. The red areas are the segmentation results containing true damage sites.

Figure 14

Figure 13 The overall damage segmentation results of a large-aperture optic. (a)–(d) Enlarged local images.

Figure 15

Table 3 Comparison of the baseline and our two core algorithms under various evaluation metrics.

Figure 16

Figure 14 Comparison of the class activation maps between LayerCAM and CG-CAM from different stages.

Figure 17

Figure 15 The effect of each step in the nonlinear multiscale fusion algorithm.