Lenses are fundamental elements in many optical applications. However, various aberrations are inevitably present in lenses, which will affect the distribution of focused light intensity and optical imaging. Accurately predicting the aberrations of a lens is of great significance. Nevertheless, quantitatively measuring the aberrations of a lens, especially when multiple aberrations are present simultaneously, is a challenging task. In this paper, we propose a method based on a designed deep residual network called Y-ResNet to measure the astigmatism and coma of a lens simultaneously. The Y-ResNet was trained on the focused image pattern of a Gaussian beam passing through a lens with astigmatism and coma. The trained network can accurately predict the aberration coefficients of the lens with 0.99 specificity, 0.925 precision, 0.9382 recall, and a 0.9406 F1-score achieved on astigmatism and 0.99 specificity, 0.956 precision, 0.98 recall, and a 0.954 F1-score achieved on coma. Specifically, even if only part of the intensity distribution of the light spot is captured, the network can accurately estimate the aberrations of the lens with an accuracy of over 90% on coma and can identify astigmatism aberration features. This paper can provide a feasible method for correcting beam patterns caused by aberration based on deep learning.

The transmission of laser beams through optical systems has important significance for the interaction between lasers and matter as well as for fields such as optical communication.1 It is necessary to obtain a focused beam of light with a specific intensity distribution for laser material processing and other applications. Transmission theory can be used to simulate the intensity distribution of a light beam as it transmits through an ideal optical system.2 However, aberrations in the optical system are typically ignored in such simulations, despite the presence of aberrations that can alter the intensity distribution and impact laser applications.

Astigmatism and coma, as representative forms of off-axis aberrations, have become increasingly important in the design of wide-field optical systems.3 With increasing astigmatism, the sharpness of beam patterns around their edges deteriorates. Astigmatism distorts the ideal intensity distribution into a complex and slender diffraction pattern, which becomes more pronounced as the level of astigmatism increases. Coma, named for its elongated comet-like structure, produces a complex and asymmetric diffraction pattern with a bright center around which residual spots are arranged in a triangular pattern. Addressing the shape distortion of beam patterns resulting from these aberrations is a critical challenge, particularly in imaging and related fields. Additionally, accurately estimating the aberrations is crucial for correcting them and controlling the beam patterns.

While traditional optical field transmission theory can simulate the transmission of a light beam through an optical system with aberrations, acquiring aberration parameters from a known intensity distribution of a beam of light is challenging, particularly when multiple aberrations exist simultaneously. With the increase in computational power and data volume, deep learning technology has rapidly developed and provided a potential solution to address this problem. Deep Convolutional Neural Networks (DCNNs)4,5 have been widely used in various tasks, including face alignment,6 object identification,7,8 image segmentation,9 fault diagnosis,10,11 and target detection.12,13 However, as the number of network layers increases, DCNNs can suffer from degradation.14 In 2015, Microsoft15 developed a new network called Deep Residual Networks (ResNet), which introduces a deep residual learning framework to address the degradation problem and enhance robustness. ResNet has been applied to classification tasks, such as ImageNet and CIFAR-10, and has achieved excellent results, broadening the use of deep learning in both academic and applied research. ResNet is mainly used for classification, such as disease detection and classification,16 garbage image classification,17 and medical image classification.18–20 

Möckl et al.21 simulated point spread function (PSF) images with Zernike aberrations at different focal points using vector diffraction theory. By training the neural network on these 100 000 images, it demonstrated good performance in predicting aberration coefficients. This approach successfully verified its capability in realistic simulations of complex PSF images carrying aberration phase information. Vanberg et al.22 demonstrated that deep CNN can accurately perceive the wavefront of focal plane images by predicting modal Zernike coefficients using two datasets of Zernike coefficients. Ma et al.23 took two sets of datasets of in-focal and out-of-focus intensity images as the input of AlexNet, and the corresponding Zernike coefficients were obtained from the intensity images. Wang et al.24 proposed phase-retrieval deep convolutional neural networks (PRDCNNs) based on the U-Net architecture, which directly measured the aberration structure and showed good fluctuation and robustness more directly and accurately. These provided proof that DCNNs can extract aberration information from intensity pictures for estimation. Methods for estimating aberrations using DCNNs have recently been applied to aberrations caused by optical systems,25,26 biological imaging,27,28 and turbulence.29,30

In order to obtain multiple aberration coefficients simultaneously from a known beam intensity distribution and accurately estimate the aberrations for use in controlling the beam patterns, we propose a new residual network called Y-ResNet to simultaneously estimate and identify these two aberration coefficients from the diffracted intensity patterns. In this study, the dataset of diffraction intensity patterns of a single lens with both coma and astigmatism aberrations is constructed by diffraction integration. The results demonstrate the effectiveness of our method, as it accurately extracts the difference between the two types of aberrations and estimates the categories simultaneously. This approach overcomes the limitations of the Zernike polynomial-based methods and provides more comprehensive and accurate results for aberration estimation. Estimating aberrations using the method of deep learning is more efficient than traditional wave-front aberration31–33 estimation methods. Compared to the previous work with a dataset of 100 000 images,21–23 we obtained high accuracy with a much smaller dataset of 6561 and can simultaneously make separate estimations for two different aberrations in intensity images.

This section outlines the fundamental principles of the propagation of a Gaussian beam through a lens with astigmatism and coma aberrations. Under the paraxial condition,34 the ABCD transfer matrix is used to represent the transmission of a laser beam transmitting through an optical system.

In a Cartesian coordinate system, the complex amplitude of a Gaussian beam in free space located at the source plane (z = 0) can be expressed as follows:
(1)
where (x0, y0) are the position coordinates of the source plane, w0 represents the beam waist of a Gaussian beam, and the value of w0 is 2 × 10−3 m.
Under the paraxial condition, the extended Huygens-Fresnel principle states that the Collins diffraction integral of a Gaussian beam propagating through an optical system can be expressed as follows:
(2)
where (x0, y0) and (x1, y1) denote the position coordinates of the source plane (z = 0) and the diffraction plane (z > 0), respectively. z represents the distance of beam propagation after transmitting through a thin lens, which is 1.5 × 10−1 m. E1(x1,y1,z) is the complex amplitude at the diffraction plane. k = 2π/λ is the wave number, with λ being the wavelength of the Gaussian beam, which is 632.8 × 10−9 m. A, B, and D are the elements of the ABCD transfer matrix. For a Gaussian beam that transmits through an ideal single thin lens and then propagates for z distance in free space (Fig. 1), the ABCD transfer matrix is written as
(3)
where f denotes the focal length of the ideal lens, which is 1 × 10−1 m, and z represents the distance between the lens and the diffraction plane.
FIG. 1.

Illustration of a Gaussian beam focused by a lens with aberrations.

FIG. 1.

Illustration of a Gaussian beam focused by a lens with aberrations.

Close modal
Considering under non-ideal circumstances that a lens with aberrations of astigmatism and coma has a Collins diffraction integral in Eq. (2), it can be described as follows:
(4)
where C6 and C3 are the coefficients of astigmatism and coma, respectively.
We can obtain the intensity distribution of the beam passing through a lens with astigmatism and aberration as follows:
(5)
where E1*(x1, y1, z) is the complex conjugation of E1(x1, y1, z).

We can simulate the light intensity images using the light intensity expression in Eq. (5). We divide the range of astigmatism coefficients and coma coefficients into 81 types, respectively, which means we take 81 values evenly from 1 × 10−5 to 5 × 10−5. The intensity distribution patterns of different aberration coefficients are shown in Fig. 2.

FIG. 2.

Intensity distribution patterns of focused beam with different aberration coefficients.

FIG. 2.

Intensity distribution patterns of focused beam with different aberration coefficients.

Close modal

Deep convolutional networks combine different characteristics, which can be enriched by increasing the depth of the network. However, degradation problems can occur as the depth of the network increases, leading to a decrease in accuracy. In many cases, redundant layers exist in the network, and we hope that these layers can complete identity mapping to ensure that the inputs and outputs are the same. To address this issue, He15 proposed ResNet by replacing several layers of the original network with a residual block. ResNet uses residual blocks, which enable the direct flow of information from one layer to another while bypassing the intermediate layers. In this way, the network can retain the original information and prevent degradation effectively. The residual block acts as a shortcut connection, enabling the network to learn residual functions instead of directly learning the desired output. The introduction of ResNet has significantly improved the performance and accuracy of deep convolutional networks.

Figure 3 illustrates the fundamental structure of a residual block. It introduces “shortcut connections” in addition to the stacked and downward transfer of two convolution layers. This makes the mapping of the output F(x) + x, which does not introduce any additional parameters or increase computational complexity. The response of standard deviations (std) with the residual function is typically closer to 0 compared to the error response without the residual function. The residual formula is given as follows:
(6)
(7)
(8)
where x in Eq. (6) is the input vector, σ represents the Rectified Linear Unit (ReLU) with simplified bias, and W1 and W2 are the abbreviations of two weight layers in Fig. 3. Equation (6) describes the purple path in Fig. 3. y in Eq. (7) is the output vector, and F(x,Wi) is the mapping of the residual that we want to acquire. Equation (7) describes the final output in Fig. 3. Equation (8) is an improvement of Eq. (7), in which, if the dimensions between x and F are not equal, we can match the dimensions by the shortcut connections with a linear unit Ws. We apply the residual formula to every few stacked layers and downward transfer to the other few stacked layers.
FIG. 3.

Residual structure.

FIG. 3.

Residual structure.

Close modal

In order to estimate the aberration coefficients of astigmatism and coma, we have developed a neural network called Y-ResNet based on the classical CNN network known as the deep residual network (ResNet).15  Figure 4(a) illustrates the architecture of the network. The input images of the network are intensity patterns with a single channel with pixels of 224 × 224. To accommodate the network architecture, we pad the input images with the zero-padding layer, resulting in images becoming 230 × 230 pixels. The initial stage of Y-ResNet involves a convolutional layer with a size of 7 × 7 and a stride of 2 × 2 for down sampling; the activation function is set as ReLU. To prevent overfitting and enhance training efficiency, a batch normalization layer (BN) is incorporated. Additionally, a max-pooling layer with a size of 7 × 7 and a stride of 2 × 2 for decreasing the parameters of training is connected with the zero-padding layer as a basic convolution unit. For all the ResNet networks with different layers (18, 34, 50), four “convolutional blocks” are the components of the main structure of the networks. For Resnet-18, each block includes two identity blocks separately, and each identity block comprises two convolution layers, as shown in Fig. 4(c). Each “conv” displayed in Fig. 4 consists of a convolution layer and batch normalization, as shown in Fig. 4(e). For Resnet-34 and Resnet-50, the numbers of “convolutional blocks” are three, four, six, and three, respectively, as shown in Figs. 4(b) and 4(d). In order to extract global information and mitigate overfitting, an average-pooling layer is connected to the “convolution blocks.” This pooling layer helps aggregate information from the previous layers. After passing through all the convolution layers and pooling layers above, information about the input images is fully extracted and selected. The next step involves feeding the selected information into a fully connected layer with a dropout rate of 0.0002. The outputs of the network are two different fully connected layers representing the estimating values of astigmatism coefficients and coma coefficients, respectively.

FIG. 4.

The structure of ResNet to determine two aberration coefficients. (a) Main structure of the network; (b) Y-ResNet-34; (c) Y-ResNet-18; (d) Y-ResNet-50; (e) conv unit.

FIG. 4.

The structure of ResNet to determine two aberration coefficients. (a) Main structure of the network; (b) Y-ResNet-34; (c) Y-ResNet-18; (d) Y-ResNet-50; (e) conv unit.

Close modal

All three ResNet models are trained by Adam as an optimizer with a learning rate of 0.0002 and binary cross-entropy as a loss function. We train the network with 50 epochs and shrink by 0.1 at the 15th and 25th epochs.

Based on the formula in Sec. II A, we explored the diffraction patterns of a Gaussian beam passing through a lens with astigmatism and coma and propagating a certain distance z. In order to provide the datasets to the neural network, we employed MATLAB as a simulation tool. The beam parameters are presented in Table I.

TABLE I.

Basic parameters of the Gaussian beam.

Beam parametersNumerical value
λ (wavelength) 632.8 × 10−9
f (focal length) 2 × 10−1
Z (propagation distance) 1.5 × 10−1
w (beam waist) 2 × 10−3
C6 (Astigmatism coefficient) 1 × 10−4–5 × 10−4 
C3 (Coma coefficient) 1 × 10−4–5 × 10−4 
Beam parametersNumerical value
λ (wavelength) 632.8 × 10−9
f (focal length) 2 × 10−1
Z (propagation distance) 1.5 × 10−1
w (beam waist) 2 × 10−3
C6 (Astigmatism coefficient) 1 × 10−4–5 × 10−4 
C3 (Coma coefficient) 1 × 10−4–5 × 10−4 

The amount of datasets plays a crucial role in determining the performance of neural networks. We employ four different datasets to validate the impact of the dataset amount on the network’s ability to predict aberration coefficients. Group A: 81 types of astigmatism coefficients and 81 types of coma coefficients to jointly construct 6561 intensity patterns. Group B: 67 types of astigmatism coefficients and 67 types of coma coefficients to jointly construct 4489 intensity patterns. Group C: 58 types of astigmatism coefficients and 58 types of coma coefficients to jointly construct 3364 intensity patterns. Group D: 51 types of astigmatism coefficients and 51 types of coma coefficients to jointly construct 2601 intensity patterns. All the intensity pattern images have been set in size to 224 × 224 pixels. The datasets are separated into validation sets and training sets in a 1:9 ratio.

To evaluate the accuracy of the results estimated by the Y-Resnet-34, we adopt four evaluation metrics: Mean Absolute Error (MAE), Mean Squared Error (MSE), Root Mean Square Error (RMSE), and Spearman’s Rank Order Correlation Coefficient (Rxy). The MSE metric reflects the accuracy of signal recovery,35 while Rxy is applicable when the variables are measured on an ordinal scale.36 The reason we chose Resnet-34 as the neural network for prediction is to strike a balance between training time and training precision. The definitions of the four metrics are as follows:
(9)
(10)
(11)
(12)
where yi is the actual value of the i-th, xi is the estimated value of the i-th, n is the number of the data and x̄ and ȳ are the mean values of xi and yi, respectively.
In order to further validate the estimated ability of the network to predict light intensity images, we define the accuracy as
(13)
where yec is the number of correct estimations and Ye is the number of all estimations.

The better the Y-Resnet-34 performance, the lower the values of MAE, RMSE, and MSE, and conversely, the higher the values of Rxy and accuracy.

Figures 5 and 6 illustrate the results of four different metrics for the four datasets. It can be clearly seen from Fig. 5 that the values of the three types of errors in Group A are lower than those of the other three groups. Figure 6 also shows that Group A exhibits higher Rxy values compared to the other groups for two aberration coefficients. Based on the analysis, the training samples with 6561 demonstrate better performance with Y-Resnet-34.

FIG. 5.

Measurement standard of classification on Y-ResNet-34. (a) Astigmatism coefficients; (b) Coma coefficients.

FIG. 5.

Measurement standard of classification on Y-ResNet-34. (a) Astigmatism coefficients; (b) Coma coefficients.

Close modal
FIG. 6.

Rxy on two aberration coefficients with four groups on Y-ResNet-34.

FIG. 6.

Rxy on two aberration coefficients with four groups on Y-ResNet-34.

Close modal

Figure 7 demonstrates the accuracy of two types of aberration with Y-ResNet-34. For astigmatism, the network achieves a preferable performance, whereas for coma, the representation accuracy is relatively lower, with the highest accuracy being only 84.8%. That means the performance of the network in estimating coma coefficients is lower than that in astigmatism coefficients.

FIG. 7.

Accuracy of astigmatism and coma on four groups on Y-ResNet-34.

FIG. 7.

Accuracy of astigmatism and coma on four groups on Y-ResNet-34.

Close modal

In Sec. III A, we found a suitable input sample (input 6561 images of 81 classes for each aberration) for determining the aberration coefficients, and now we input this suitable input sample into different networks (Y-Resnet-18, Y-ResNet-34, and Y-ResNet-50) to verify the network’s prediction accuracy of the two aberration parameters. Simultaneously, in order to optimize the issue of low prediction accuracy for coma, we have implemented an optimization approach. Specifically, we cropped the input image to a certain size that only maintain the bright pattern and remove the black borders, as shown in Fig. 8.

FIG. 8.

Intensity patterns before and after cropped. (a) Before cropped; (b) after cropped.

FIG. 8.

Intensity patterns before and after cropped. (a) Before cropped; (b) after cropped.

Close modal
In this section, we use different evaluation metrics to express the accuracy of neural network predictions and evaluate the classification task’s performance. Different from the one used in Sec. III A, these evaluation metrics, which are widely used in classification tasks to evaluate the effectiveness of the network model, are defined as
(14)
(15)
(16)
(17)

These metrics consist of the arithmetic operation of four sets: (a) True Positive (TP): predict aberration coefficients correctly. (b) True Negative (TN): predict not a specific type of aberration coefficients correctly. (c) False Negative (FN): predict aberration coefficient incorrectly. (d) False Positive (FP): predict not a specific type of aberration coefficients incorrectly. The higher value of specificity indicates that the classifier is more capable of identifying negative and positive samples; the higher values of recall and precision indicate that the classifier is more capable of identifying positive samples; and the F1-score combines the precision and recall of the classifier, which is a good balance of the performance evaluation. These four evaluation functions measure the performance of the network using both positive and negative samples.

We employ Y-ResNet-18, Y-ResNet-34, and Y-ResNet-50 on the same datasets to verify if the coefficients are well predicted. Four metrics for characterizing two aberration coefficients by three networks are shown in Fig. 9.

FIG. 9.

Results analysis of four metrics with three Networks. (a) Astigmatism coefficients; (b) coma coefficients.

FIG. 9.

Results analysis of four metrics with three Networks. (a) Astigmatism coefficients; (b) coma coefficients.

Close modal

Figure 9 implies that the Y-ResNet-18 network achieves better results in the classification of aberration coefficients compared to Y-ResNet-34 and Y-ResNet-50. That means for 6561 intensity patterns, fewer layers can better fit the dataset. Moreover, the evaluation metrics of specificity, precision, recall, and F1-score maintain a significantly high value of over 0.954 for coma coefficients and over 0.925 for astigmatism coefficients. For aberration classification on both Y-ResNet-34 and Y-ResNet-50, the values of specificity are above 0.99.

In the classification process, it is possible for FP and FN to be close to zero, leading to limitations in precision and recall measurements. To overcome this limitation, the F1-score is introduced, which is the harmonic mean of precision and recall. The F1-score is highest when precision and recall are close to each other.

Based on the values of these metrics, it can be concluded that the network exhibits excellent performance in predicting the values of each aberration coefficient. The high values of specificity, precision, recall, and F1-score indicate that the network is able to accurately classify the aberration coefficients, further emphasizing its effectiveness in this task.

In some circumstances, the detector cannot capture the distribution of intensity patterns completely. In view of this situation, the ability to recognize a partial intensity pattern is crucial. Therefore, in this section, we cut the intensity pattern into different sizes to test and verify the capability of the designed ResNet for the recognition of aberration coefficients. This analysis will help assess the model’s performance and its ability to extract meaningful features from incomplete or partial intensity patterns, providing insights into its robustness in practical applications.

Figure 10 illustrates the cropped intensity pattern images with different sizes, where each subsequent image is obtained by clipping 1/7 of the original image. It can be observed that as the intensity pattern size decreases, the distortion caused by astigmatism gradually reduces, while the distortion caused by coma increases. To analyze the accuracy of the network in predicting astigmatism and coma coefficients using different cropped sizes, the cropped intensity patterns are input into the Y-ResNet-50 model for training. The learning rate, loss function, and optimizer remain the same as in Sec. III B.

FIG. 10.

Aberration estimation of cropped intensity patterns.

FIG. 10.

Aberration estimation of cropped intensity patterns.

Close modal

As the size of the intensity pattern decreases, the accuracy of network predictions gradually decreases to 90%. This indicates that as the intensity pattern size decreases, the captured image features also diminish, resulting in a decrease in prediction accuracy. Furthermore, Fig. 10 also demonstrates that the accuracy of astigmatism coefficients noticeably decreases with the cropping of intensity patterns. This observation aligns with our understanding that cropped intensity patterns provide less information for accurately estimating astigmatism coefficients.

Overall, these findings highlight the impact of intensity pattern size on network accuracy and emphasize the importance of considering the size and quality of captured intensity patterns to achieve reliable predictions of astigmatism and coma coefficients.

In this paper, we propose a neural network called Y-ResNet for accurately predicting the aberration of a Gaussian beam as it propagates through a single lens over a certain distance. The Y-ResNet network is specifically designed to extract and estimate aberration coefficient features from intensity patterns. Unlike traditional ResNet models, Y-ResNet has an output layer with two parallel branches, enabling separate classification of astigmatism coefficients and coma coefficients.

The results demonstrate the high accuracy of astigmatism coefficient estimation, with a rate of over 96% achieved with different dataset sizes. Additionally, after removing irrelevant information from the intensity patterns, the estimation accuracy of coma coefficients reaches as high as 98% with different layers of ResNet. This proves that it is feasible to extract and estimate the aberration coefficients from the intensity patterns. To further evaluate the ability of the network to predict aberrations, we conducted experiments with intensity pattern cropping of various sizes. The accuracy of aberration prediction gradually decreases as the cropping size becomes smaller, but it consistently remains above 90%. This suggests that even with smaller cropped patterns, the neural network can still extract relevant features, albeit with reduced influence from astigmatism.

These findings demonstrate the effectiveness of the Y-ResNet model in predicting aberration coefficients and highlight the model’s robustness in handling intensity patterns of different sizes. The proposed approach holds promise for accurate aberration estimation and contributes to the advancement of aberration analysis in Gaussian beam propagation through lenses. We have only estimated the aberrations carried by the intensity images at the simulation level and have not confirmed the feasibility of the method experimentally. Future work could focus on changing different types of incident beams for different applications or try to estimate three or more types of aberrations simultaneously.

This work was supported by the Scientific Research Fund of Liaoning Provincial Education Department (Grant No. LJKMZ20220620) and the 2023 Central Government guidance for local science and technology development funds (basic research on free exploration) (Grant No. 2023JH6/100100066).

The authors have no conflicts to disclose.

Jinyang Jiang: Conceptualization (equal); Formal analysis (equal); Software (equal); Writing – original draft (equal). Xiaoyun Liu: Funding acquisition (equal); Writing – original draft (equal). Yonghao Chen: Conceptualization (equal); Formal analysis (equal). Siyu Gao: Formal analysis (equal). Ying Liu: Formal analysis (equal). Yueqiu Jiang: Funding acquisition (lead).

The data that support the findings of this study are available from the corresponding author upon reasonable request.

1.
L. C.
Andrews
and
R. L.
Phillips
,
Laser Beam Propagation through Random Media
(
SPIE Press
,
Bellingham, WA
,
2005
).
2.
A. P.
Dzyuba
, “
Optical phase retrieval with the image of intensity in the focal plane based on the convolutional neural networks
,”
J. Phys.: Conf. Ser.
1368
(
2
),
022055
(
2019
).
3.
A. K.
Gupta
and
K.
Singh
, “
Partially coherent far-field diffraction in the presence of primary astigmatism
,”
Can. J. Phys.
56
(
12
),
1539
1544
(
1978
).
4.
Y.
Dong
and
Y.
Wu
, “
Adaptive cascade deep convolutional neural networks for face alignment
,”
Computer Stand. Interfaces
42
,
105
112
(
2015
).
5.
V. P.
Lutsiv
, “
Convolutional deep-learning artificial neural networks
,”
J. Opt. Technol.
82
,
499
508
(
2015
).
6.
S.
Zhang
,
H.
Yang
, and
Z.
Yin
, “
Multiple deep convolutional neural networks averaging for face alignment
,”
J. Electron. Imaging
24
(
3
),
033013
(
2015
).
7.
H.
Chao
,
Y.
He
,
J.
Zhang
, and
J.
Feng
, “
GaitSet: Regarding gait as a set for cross-view gait recognition
,”
Proc. AAAI Conf. Artif. Intell.
33
,
8126
8133
(
2019
).
8.
C.
Li
,
X.
Wei
,
H.
Guo
,
W.
He
,
X.
Wu
,
H.
Xu
, and
X.
Liu
, “
Recognition of the internal situation of aircraft skin based on deep learning
,”
AIP Adv.
11
(
10
),
105216
(
2021
).
9.
Y.
Li
and
X.
Li
, “
Automatic segmentation using deep convolutional neural networks for tumor CT images
,”
Int. J. Pattern Recognit. Artif. Intell.
37
(
03
),
2352003
(
2023
).
10.
Z.
Zhou
,
H.
Wang
,
Z.
Li
, and
W.
Chen
, “
Fault diagnosis of rolling bearing based on deep convolutional neural network and gated recurrent unit
,”
J. Adv. Mech. Des., Syst., Manuf.
17
(
2
),
JAMDSM0017
(
2023
).
11.
Z.
Xing
,
Y.
Liu
,
Q.
Wang
, and
J.
Li
, “
Multi-sensor signals with parallel attention convolutional neural network for bearing fault diagnosis
,”
AIP Adv.
12
(
7
),
075020
(
2022
).
12.
Y. S.
Balcioglu
,
B. C.
Sezen
, and
C.
Cubukcu Cerasi
, “
Solar cell busbars surface defect detection based on deep convolutional neural network
,”
IEEE Latin Am. Trans.
21
(
2
),
242
250
(
2023
).
13.
C.
Li
,
G.
Cui
,
W.
Zhang
et al, “
Defect detection in vehicle mirror nonplanar surfaces with multi-scale atrous single-shot detect mechanism
,”
AIP Adv.
11
(
7
),
075202
(
2021
).
14.
K.
He
and
J.
Sun
, “
Convolutional neural networks at constrained time cost
,” in
IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
(
IEEE
,
Boston, MA
,
2015
), pp.
5353
5360
.
15.
K.
He
,
X.
Zhang
,
S.
Ren
, and
J.
Sun
, “
Deep residual learning for image recognition
,” in
IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
(
IEEE
,
Las Vegas, NV
,
2016
), pp.
770
778
.
16.
M.
Ahmed
and
A.
Ahmed
, “
Palm tree disease detection and classification using residual network and transfer learning of inception ResNet
,”
PLoS One
18
(
3
),
e0282250
(
2023
).
17.
X.
Ma
,
Z.
Li
, and
L.
Zhang
, “
An improved ResNet-50 for garbage image classification
,”
Tech. Gaz.
29
(
5
),
1552
1559
(
2023
).
18.
S.
Showkat
and
S.
Qureshi
, “
Efficacy of transfer learning-based ResNet models in chest X-ray image classification for detecting COVID-19 pneumonia
,”
Chemom. Intell. Lab. Syst.
224
,
104534
(
2022
).
19.
M.
Yamin
,
A.
M Basahel
,
M.
Abusurrah
,
S.
M Basahel
,
S.
Nandan Mohanty
, and
E.
Laxmi Lydia
, “
Automated artificial intelligence empowered white blood cells classification model
,”
Comput., Mater. Continua
75
(
1
),
409
425
(
2023
).
20.
Z.
Zhu
,
M.
Attique Khan
,
S.
Wang
, and
Y.
Zhang
, “
RBEBT: A ResNet-based BA-ELM for brain tumor classification
,”
Comput., Mater. Continua
74
(
1
),
101
111
(
2023
).
21.
L.
Möckl
,
P. N.
Petrov
, and
W. E.
Moerner
, “
Accurate phase retrieval of complex 3D point spread functions with deep residual neural networks
,”
Appl. Phys. Lett.
115
(
25
),
251106
(
2019
).
22.
P. O.
Vanberg
et al, “
Machine learning for image-based wavefront sensing
,” in
33rd Conference on Neural Information Processing Systems (NeurIPS). Machine Learning and the Physical Sciences Workshop (Vancouver, Canada,
2019
), p.
107
.
23.
H.
Ma
,
H.
Liu
,
Y.
Qiao
,
X.
Li
, and
W.
Zhang
, “
Numerical study of adaptive optics compensation based on Convolutional Neural Networks
,”
Opt. Commun.
433
,
283
289
(
2019
).
24.
Y.
Wang
,
H.
Wang
,
Y.
Li
,
C.
Hu
,
H.
Yang
, and
M.
Gu
, “
High-accuracy, direct aberration determination using self-attention-armed deep convolutional neural networks
,”
J. Microsc.
286
,
13
21
(
2022
).
25.
Y.
Niu
,
Z.
Gao
,
C.
Gao
,
J.
Zhao
, and
X.
Wang
, “
Interferometric wavefront sensing system based on deep learning
,”
Appl. Sci.
10
(
23
),
8460
(
2020
).
26.
B. P.
Cumming
and
M.
Gu
, “
Direct determination of aberration functions in microscopy by an artificial neural network
,”
Opt. Express
28
(
10
),
14511
14521
(
2020
).
27.
B.
Zhang
,
J.
Zhu
,
K.
Si
, and
W.
Gong
, “
Deep learning assisted zonal adaptive aberration correction
,”
Front. Phys.
8
,
621966
(
2021
).
28.
Y.
Jin
,
Y.
Zhang
,
L.
Hu
,
H.
Huang
,
Q.
Xu
,
X.
Zhu
,
L.
Huang
,
Y.
Zheng
,
H. L.
Shen
,
W.
Gong
, and
K.
Si
, “
Machine learning guided rapid focusing with sensor-less aberration corrections
,”
Opt. Express
26
(
23
),
30162
30171
(
2018
).
29.
K.
Wang
,
M. M.
Zhang
,
J.
Tang
,
L.
Wang
,
L.
Hu
,
X.
Wu
,
W.
Li
,
J.
Di
,
G.
Liu
, and
J.
Zhao
, “
Deep learning wavefront sensing and aberration correction in atmospheric turbulence
,”
PhotoniX
2
(
1
),
8
(
2021
).
30.
A. B.
Siddik
,
S.
Sandoval
,
D.
Voelz
,
L. E.
Boucheron
, and
L.
Varela
, “
Deep learning estimation of modified Zernike coefficients and recovery of point spread functions in turbulence
,”
Opt. Express
31
(
14
),
22903
22913
(
2023
).
31.
C.
Rodríguez
and
N.
Ji
, “
Adaptive optical microscopy for neurobiology
,”
Curr. Opin. Neurobiol.
50
,
83
91
(
2018
).
32.
J. W.
Cha
,
J.
Ballesta
, and
P. T. C.
So
, “
Shack-Hartmann wavefront-sensor-based adaptive optics system for multiphoton microscopy
,”
J. Biomed. Opt.
15
(
4
),
046022-1
046022-10
(
2010
).
33.
J.
Liang
,
B.
Grimm
,
S.
Goelz
, and
J. F.
Bille
, “
Objective measurement of wave aberrations of the human eye with the use of a Hartmann–Shack wave-front sensor
,”
J. Opt. Soc. Am. A
11
(
7
),
1949
1957
(
1994
).
34.
J.
Alda
,
J.
Alonso
, and
E.
Bernabeu
, “
Characterization of aberrated laser beams
,”
J. Opt. Soc. Am. A
14
(
10
),
2737
(
1997
).
35.
I. D.
Schizas
,
G. B.
Giannakis
, and
Z.
Luo
, “
Distributed estimation using reduced-dimensionality sensor observations
,”
IEEE Trans. Signal Process.
55
(
8
),
4284
4299
(
2007
).
36.
M. T.
Puth
,
M.
Neuhäuser
, and
G. D.
Ruxton
, “
Effective use of Spearman’s and Kendall’s correlation coefficients for association between two measured traits
,”
Anim. Behav.
102
,
77
84
(
2015
).