Introduction

In recent years, artificial-intelligence technology has been widely used in production processes to reduce labor and quickly respond to process variations. In particular, machine learning and deep learning methods have been introduced into process control and excelled in quickly detecting and correctly identifying abnormal process conditions. For example, Lee and Kim1, Wang et al.2, Chen and Yu3, Kim and Ha4, Yeganeh et al.5, and Sabahno and Amiri6 proposed control charts based on machine or deep learning to improve detection ability. Zhang et al.7, Yu and Liu8, Maged et al.9, and Yu et al.10,11 achieved excellent performance in failure detection in high-dimensional or complex processes using deep learning methods. Moreover, Zan et al.12, Lu et al.13, Yu and Zhang14, Lee et al.15, and Xue et al.16 used machine and deep learning methods to recognize abnormal control chart patterns.

To test the lifetime of electronic products, practitioners may set a test termination time at which to halt the test in order to reduce time and cost. As some tested units may not have failed by the termination time, practitioners can only obtain incomplete lifetime records, which are known as type-I right-censored data. This is a challenge for practitioners because the detection efficiency for process variation of traditional Shewhart-type and EWMA-type control charts decrease for incomplete data. Thus, some researchers have proposed novel control charts to monitor right-censored data10,11,17,18. Steiner and Mackay19 estimated the mean lifetime of incomplete data with a CEV and combined it with an \(\overline{X}\) chart of a lower control limit (LCL) to detect decreases of the lifetime for highly right-censored data with assumptions of normality. Then, Steiner and Mackay20 used Shewhart-type CEV control charts for censored data with non-normality conditions. They next Steiner and Mackay21 combined the EWMA and CEV to increase the detection ability for highly right-censored data. Lee22 determined the design parameters of the CEV \(\overline{X}\) control chart with a minimum cost perspective.

Zhang and Chen23 assumed a Weibull distribution for the data and detected decreases and increases in the average lifetime for censored data using two single-sided EWMA CEV control charts. Tsai and Lin24, Raza et al.25, and Biozone and Wang26 also developed EWMA-type CEV control charts and achieved good monitoring performance in non-normal censored data for their proposed charts. Raza et al.27 and Raza and Siddiqi28 established a double EWMA (DEWMA) control chart to detect shifts of the scale parameter for gamma censored data. Raza and Siddiqi28, Raza et al.29, and Ali et al.30 proposed similar contributions to the DEWMA CEV control chart. Ali et al.31 and Ahmed et al.32,33 developed novel control charts to monitor Weibull and generalized exponential (GE) censored data by replacing CEV with CM and conditional standard deviation (CSD). In addition, Lee et al.34 and Zhao and Wu35 implemented the EWMA CEV chart for multiple censored data and window-censored data, respectively. However, to the authors’ knowledge, there have been no relevant studies monitoring right-censored data using deep learning methods.

Convolutional neural networks (CNN) and long short-term memory (LSTM) networks are common methods of deep learning that have been successfully combined with control chart technology to monitor processes and recognize abnormal patterns in control charts7,14,36. Thus, these techniques may be used to improve the efficiency of a control chart for censored data. From the literature, most previous works developed CEV and CM charts for normal, GE or Weibull censored data. Gamma data is also often observed in reliability life tests and is one of the commonly used lifetime distributions because changing its shape parameter can create distributions with different degrees of skewness. Therefore, this study combines a CNN and LSTM, respectively, with the EWMA CEV and CM charts to monitor gamma type-I right-censored data.

The remainder of this study is organized as follows. In “Methodology” section presents the methodology. In “Proposed control charts” section describes the proposed control chart and in “Performance comparison” section investigates its performance using a statistical simulation. In “Real-world case study” section details implementation of the proposed method for a case study. Finally, in “Conclusions” section provides some conclusions and directions for future development.

Methodology

EWMA charts for gamma type-I censored data

Let \(U = \left\{ {u_{1} ,u_{2} , \ldots ,u_{m} } \right\}\) be a gamma random variable and \(a_{0}\) and \(b_{0}\) be the shape and scale parameters of an in-control process, respectively. The censoring rate for the gamma lifetimes can be represented as \(Pc = 1 - F_{Ga} \left( {u = c_{T} {|}a_{0} ,b_{0} } \right)\), where \(F_{Ga} \left( { \cdot {|}a_{0} ,b_{0} } \right)\) is the cumulative distribution function (CDF) of the gamma distribution with parameters \(a_{0}\) and \(b_{0}\), where \(c_{T}\) is the censoring time. The CEV of the gamma distribution is:

$$Cev = E\left( {U{|}u \ge c_{T} } \right) = \frac{{a_{0} b_{0} \left[ {1 - F_{Ga} \left( {u = c_{T} {|}a_{0} + 1,b_{0} } \right)} \right]}}{{1 - F_{Ga} \left( {u = c_{T} {|}a_{0} ,b_{0} } \right)}},$$
(1)

and the CM of the gamma distribution is

$$CM = F_{Ga}^{ - 1} \left( {0.5 - 0.5F_{Ga} \left( {u = c_{T} {|}a_{0} ,b_{0} } \right){|}a_{0} ,b_{0} } \right),$$
(2)

where \(F_{Ga}^{ - 1} \left( { \cdot {|}a_{0} ,b_{0} } \right)\) is an inverse of CDF of the gamma distribution with \(a_{0}\) and \(b_{0}\). The derivations of Eqs. (1) and (2) are shown in Supplementary. Practitioners take \(n\) samples and measure their lifetime values using the reliability life testing method. The sample mean of size \(n\) can be obtained by \(\overline{X} = \sum\nolimits_{i = 1}^{n} {x_{i} } /n\). Let \(u_{i}\) be the lifetime of the i-th testing sample. Then, the \(x_{i}\) of the i-th testing sample in the \(\overline{X}\)’s formula is:

$$x_{i} = \left\{ {\begin{array}{*{20}l} {u_{i} } \hfill & {\quad for\;u_{i} \le c_{T} } \hfill \\ {Cd} \hfill & {\quad for\;u_{i} > c_{T} } \hfill \\ \end{array} } \right.,$$
(3)

where \(Cd = Cev\) for CEV \(\overline{X}\) statistic or \(CM\) for CM \(\overline{X}\) statistic. The in-control mean \(M_{0}\) and variance \(V_{0}\) can be expressed as follows:

$$\begin{aligned} & M_{0} = a_{0} b_{0} \;\;{\text{and}} \\ & V_{0} = a_{0} b_{0}^{2} . \\ \end{aligned}$$

Let \(\lambda\) be the smoothing parameter of the EWMA chart. Zhang and Chen23 showed that the EWMA statistic at period j for monitoring the mean decrease is:

$$E_{j} = min\left\{ {M_{0} ,\lambda \overline{{X_{j} }} + \left( {1 - \lambda } \right)E_{j - 1} } \right\}.$$
(4)

For an EWMA chart with CEV or CM used to monitor a process mean, an LCL is set to signal the mean reduction because practitioners always focus on the detection of average lifetime reduction. The appearance of an assignable cause leads to a decrease in the process mean, indicating an out-of-control condition. Let \(M_{1} = \delta \times M_{0}\) be the mean of an out-of-control state where the process variance is unchanged and \(\delta\) be a mean shift size that can be obtained by \(\delta = M_{1} /M_{0}\). The gamma shape parameter \(a_{1}\) and scale parameter \(b_{1}\) in an out-of-control state can be obtained by solving the following system of simultaneous equations:

$$\left\{ {\begin{array}{*{20}l} {M_{1} = a_{1} b_{1} } \hfill \\ {V_{0} = a_{1} b_{1}^{2} } \hfill \\ \end{array} } \right.,$$

The solutions for \(a_{1}\) and \(b_{1}\) are, respectively, as follows:

$$\begin{aligned} a_{1} & = M_{1}^{2} /V_{0} \;\;{\text{and}} \\ b_{1} & = V_{0} /M_{1} . \\ \end{aligned}$$
(5)

The primary metric employed for evaluating the effectiveness of control charts is the average run length (ARL)3,15,23,30,34. In an in-control process, a larger ARL signifies a reduced false-alarm rate, while in an out-of-control state, a smaller ARL indicates quicker detection of mean reduction.

CNN

The main advantage of CNNs is that they can effectively capture local features in the data and perform feature extraction for classification or regression prediction, while maintaining the spatial hierarchy of features. Some of the main features and working principles of CNN are laid out below.

The convolutional layer is a fundamental element of CNNs. It operates by applying convolutional kernels (also referred to as filters) over the input data or image through a sliding process, resulting in the generation of feature maps. This operation supports the identification of local features. The output of a convolutional layer \(l\) can be expressed by:

$$\zeta_{l}^{CL} = f_{RL} \left( {b_{l}^{CL} + \zeta_{l - 1}^{CL} \times \omega_{l - 1}^{CL} } \right),$$

where \(f_{RL} \left( \cdot \right)\) is a rectified linear unit (ReLU) activation function, \(l\) is the \(l\)-th layer of the CNN, \(\omega_{l - 1}^{CL}\) is the filter kernel at layer \(l - 1\), and \(b_{l}^{CL}\) is the bias vector at layer \(l\).

The pooling layer plays a crucial role in CNNs by decreasing the dimensionality of feature maps while retaining vital information. The most common pooling operation is max pooling, which selects the highest value within specific regions, thus reducing the feature map’s dimensions. The pooling output at layer \(l\) is given by:

$$Mp_{j} = max\left( {\zeta_{l}^{CL} } \right).$$

After the convolutional and pooling layers, CNNs frequently incorporate fully connected layers to carry out ultimate classification or regression tasks. These layers are responsible for transforming the extracted features into the network’s final output. CNNs commonly comprise a series of convolutional and pooling layers stacked in an interleaved fashion. This layered architecture empowers the network to acquire knowledge about image characteristics spanning diverse levels of abstraction7,8.

LSTM

An LSTM is a deep learning neural network architecture that is an improvement of over traditional recurrent neural networks (RNN). It was specially designed to process sequence data, such as speech recognition, natural language processing, time-series analysis, and other applications. LSTM supports sequence data processing by introducing the three key gating mechanisms described below.

The forget gate determines whether to forget the previous memory information. It uses a sigmoid function to output a value between 0 and 1, controlling whether past memories are retained. The forget gate \({f}^{fg}\) can be expressed as:

$$f^{fg} = f_{sf} \left( {b^{fg} + \omega^{fg} \left[ {\zeta_{t - 1} ,\vartheta_{t} } \right]} \right),$$

where \(f_{sf} \left( \cdot \right)\) is a sigmoid function, \(b^{fg}\) is the bias vector of the forget gate, \(\omega^{fg}\) is the weight vector of the forget gate, \(\zeta_{t - 1}\) is the output vector of the previous step, and \(\vartheta_{t}\) is the input vector of the current step.

The input gate determines how new memory information is added to the LSTM unit. It uses a sigmoid function to determine which information needs updating and uses the tanh function to create a memory cell vector. Let \(f^{ig}\) be the formula of the input gate and \(\epsilon_{t}\) be the memory cell vector of the current step, as follows:

$$\begin{aligned} & f^{ig} = f_{sf} \left( {b^{ig} + \omega^{ig} \left[ {\zeta_{t - 1} ,\vartheta_{t} } \right]} \right)\;\;{\text{and}} \\ & \epsilon_{t} = f^{fg} \times \epsilon_{t - 1} + f^{ig} \times tanh\left( {b^{\epsilon} + \omega^{\epsilon} \left[ {\zeta_{t - 1} ,\vartheta_{t} } \right]} \right), \\ \end{aligned}$$

where \(b^{ig}\) is the bias vector of the input gate, \(\omega^{ig}\) is the weight vector of the input gate, \(tanh\left( \cdot \right)\) is the tanh function, and \(b^{\epsilon}\) and \(\omega^{\epsilon}\) are the bias and weight vectors of the memory cell vector for the current step, respectively.

The output gate determines which memory information will be outputted to the next time step. Similar to the forget gate and input gate, the output gate uses a sigmoid function to control the output. The output gate \(f^{og}\) is:

$$f^{og} = f_{sf} \left( {b^{og} + \omega^{og} \left[ {\zeta_{t - 1} ,\vartheta_{t} } \right]} \right),$$

where \(b^{og}\) is the bias vector of the output gate and \(\omega^{og}\) is the weight vector of the output gate. The output vector of the current step \(\zeta_{t}\) is:

$$\zeta_{t} = f^{og} \times tanh\left( {\epsilon_{t} } \right).$$

The dimensions of \(\zeta_{t}\) as the number of hidden units affects the LSTM network’s computing efficiency and effectiveness. LSTM uses these gates to control the flow of information and update memory, thus helping to address the gradient vanishing problem in RNNs and allowing them to better handle long sequences9,36.

Proposed control charts

This section proposes a procedure to set and implement the control chart based on deep learning networks with EWMA CEV or CM statistic, as shown in Fig. 1. This procedure includes two parts: setting up the control chart and implementing the control chart.

Figure 1
figure 1

Flowchart for the proposed control chart.

The purpose of setting up the control chart is to find the optimal threshold value according to a specific in-control ARL value to maintain the monitoring effect of the control chart. Before setting up a control chart, practitioners can determine an in-control ARL value and choose an initial threshold value based on their own experience. In the literature, the in-control ARL value used for control chart performance comparison is typically 200 or 370.4.

According to the quality characteristic parameters and control parameters such as \(a_{0}\), \(b_{0}\), \(\lambda\), \(Pc\), and \(n\), the in-control gamma type-I censored data can be generated using the Monte Carlo method and then the in-control EWMA statistics \(E_{j}\) for the gamma type-I censored data can be calculated using Eq. (4). It is noted that \(E_{j}\) is the EWMA CEV statistic for \(Cd = Cev\) or the EWMA CM statistic for \(Cd = CM\). First, 10,000 \(E_{j}\) are generated. Next, let \(\left\{ {E_{t - 1} ,E_{t} } \right\}\) be the training data set and be inputted into the CNN or LSTM network for training. As the training data set, the input layer of the CNN or LSTM network must use one-dimensional sequence data. The input and output vectors have dimensions \(1 \times \left( {m - 1} \right)\).

Practitioners are more concerned about decreases in the average lifetime, so the threshold value \(\eta\) as a LCL of traditional control charts is set to detect such decreases. After the network is trained, the estimation \(\widehat{{E_{t} }}\) can be outputted and the residual value can be computed by \(E_{t} - \widehat{{E_{t} }}\). If the j-th residual value is less than the threshold value \(\eta\), then the number of points outside the threshold value (OC) = OC + 1. The above procedure is repeated 10,000 times to obtain total number of points outside the threshold value and the in-control ARL value is 10,000/OC (Note that the false alarm rate is OC/10,000).

If the simulated in-control ARL value is not equal to the specific in-control ARL value, then the threshold value \(\eta\) is adjusted and the above simulation procedure is repeated until the simulated ARL value equals the specific ARL value. In this way, the optimal threshold value \(\eta^{*}\) can be obtained and implemented to monitor the process.

In the implementation of the control chart for process monitoring, practitioners import a trained CNN or LSTM network, set up the optimal threshold value \(\eta^{*}\), and then apply the following steps:

  1. (1)

    Collect lifetime data for period t − 1 and calculate the statistic \(E_{t - 1}\).

  2. (2)

    Input \(E_{t - 1}\) into this trained network to predict the statistic \(\widehat{{E_{t} }}\) of period t.

  3. (3)

    After the lifetime data for period t is obtained, compute the actual value \(E_{t}\).

  4. (4)

    Let the error value for period t be \(E_{t} - \widehat{{E_{t} }}\).

If the error value is less than the optimal threshold value \(\eta^{*}\), then the process indicates an out-of-control condition; otherwise, the process is in control.

Figure 2 shows the simulation process for an out-of-control ARL value. Practitioners can give a shift size value \(\delta\) to calculate the parameters \(a_{1}\) and \(b_{1}\) of the out-of-control state shown in Fig. 2 using Eq. (5).

Figure 2
figure 2

Flowchart for simulation of an ARL value using the proposed control chart.

Some studies typically considered two different types of performance: zero-state (ZS) and steady-state (SS)37,38. ZS performance assumes that a shift occurs at the beginning of the process to measure the out-of-control ARL value. SS performance shows the out-of-control ARL for control charts to identify a process shift for control statistics to reach a static distribution.

For the simulated data generation of ZS and SS conditions, assume that the process has been continuously run for \(\pi\) sampling periods and maintained in the in-control state, \(\pi\) represents the length of the process to reach SS condition. The process occurs the mean shift between \(\pi\)th and \(\pi\) + 1st sampling, and then SS ARL represents the expected value of the number of samples obtained from the occurrence of this mean shift to when the chart indicates an out-of-control signal. In the data generation process, the in-control data of \(\pi\) periods is first generated. After the in-control data, 10,000 out-of-control data are generated. The EWMA statistics are calculated for the data of \(\pi\) + 10,000 using Eq. (4), and the last 10,000 EWMA statistics are taken to simulate the out-of-control ARL values of SS condition according to Fig. 2. For ZS condition, the simulated data can be generated with a setting of \(\pi = 0\).

MATLAB R2023a provides a deep learning toolbox that can easily implement the processes of Figs. 1 and 2. This study codes the simulation processes using MATLAB R2023a to investigate the performance of the proposed control charts.

Performance comparison

Hereafter, ‘CNN chart’ and ‘LSTM chart’ represent the control charts based on the CNN and LSTM networks, respectively. This section compares the ARL performance of the CNN, LSTM, and EWMA charts with CEV or CM for gamma type-I censored data. The parameters of the gamma distribution are set as \(a_{0}\) = 1, 2, and 4 and \(b_{0} = 1\) for comparison, while the skewness values (Sk) of the parameter case are as shown in Table 1. Smaller \(a_{0}\) values indicate greater skewness of the gamma distribution.

Table 1 Skewness values of the gamma distribution.

When using the CNN network, some network parameters, such as the stacked numbers of convolutional and pooling layers, kernel size, and number of kernels, must be determined first to achieve good training and testing results. For the LSTM chart, the stacked numbers of LSTM layers and the number of hidden units must also be decided. Based on the literature, this study uses the trial-and-error method to set these network parameters as per Table 23,39.

Table 2 Structure and parameters of deep learning networks for the proposed control charts.

In the performance comparison, the sample size is fixed at \(n = 5\), which is standard practice for sampling and plotting control charts. The smoothing parameter \(\lambda\) of the EWMA statistic is set at 0.1 and 0.223,27,29. \(Pc\) is 0.2, 0.5, and 0.8 for lower, moderate, and higher censoring rates, respectively. The shift sizes are \(\delta = 0.8\) and 0.7 for small shifts; \(\delta = 0.6\) and 0.5 for moderate shifts; and \(\delta = 0.2\) for large shifts. This study set the in-control ARL value at 200 to measure the out-of-control ARL values under ZS and SS conditions.

Considering that different trained CNN or LSTM networks will have different ARL values, this study trained 100 networks for each condition of process parameters (\(a_{0}\), \(b_{0}\), \(Pc\) and \(\lambda\)) and then selected a trained network with the smallest ARL value for comparison from the 100 trained networks.

Table 3 shows the LCL and the optimal threshold value \(\eta^{*}\) for the six comparison charts under ZS condition (\(\pi\) = 0). The out-of-control ARL values for the six control charts were simulated according to the above conditions and the ARL values are compared in Table 4. The bold cells indicate the control chart with the best detection efficiency for a specific shift size.

Table 3 Design parameter values of six control charts for comparison under ZS condition.
Table 4 ZS ARL values of six control charts.

As the skewness coefficient value decreases, the detection efficiency of the six control charts decreases. As \(Pc\) increases, the detection efficiency of the EWMA chart decreases, and the detection efficiency of the CNN and LSTM charts changes irregularly. The CNN chart exhibits the best detection ability for most shift sizes. The LSTM chart is significantly worse than the other charts for all shift sizes. Comparing the performance of CEV and CM, CNN with CEV is better than CNN with CM when the skew coefficient value is small. As the skew coefficient value becomes larger, the effect of CNN with CM is better than CNN with CEV. EWMA and LSTM charts are not affected by the skewness coefficient. EWMA and LSTM charts with CEV has better performance than EWMA and LSTM charts with CM.

For processes that often occur the mean shift in the initial stage, if the skewness coefficient of lifetime distribution is large, the EWMA CM statistic should be used to train the CNN network and implement monitoring. On the contrary, the EWMA CEV statistic should be considered to train the CNN network and implement monitoring.

This study considers \(\pi\) = 100 and 1000 to measure the ARL value of SS condition. Table 5 shows the LCL and \(\eta^{*}\) of six control charts for \(\pi\) = 1000. Table 7 exhibits the SS ARL values for \(\pi\) = 100 and 1000, respectively.

Table 5 Design parameter values of six control charts for comparison under SS condition.

As Table 6, CNN chart outperforms EWMA and LSTM for most shift sizes in the \(\pi\) = 1000. With only some cases of \(\delta\) = 0.8, the EWMA chart is better than the CNN chart. LSTM chart in most cases of \(\delta\) ≤ 0.5 has better detection efficiency than EWMA chart, but it's still not as efficient as CNN chart. The detection efficiency of LSTM charts in other shift sizes is worse than that of EWMA and CNN charts. The detection efficiency of the EWMA, CNN and LSTM charts decreases as \(Pc\) increases or the skewness coefficient of lifetime distribution becomes large for most shift sizes. EWMA, CNN and LSTM charts with CEV have better detection efficiency than EWMA, CNN and LSTM charts with CM in these cases of \(Pc=0.2\) for all \({a}_{0}\) values and gamma parameters (\({a}_{0}\) = 4, \({b}_{0}\) = 1) for all \(Pc\) values.

Table 6 SS ARL values of six control charts with \(\pi = 1000\).

Comparing the ARL values of Tables 4 and 6 for ZS (\(\pi\) = 0) and \(\pi\) = 1000, the EWMA charts with CEV and CM perform worse detection efficiency under ZS condition than under SS condition. In the LSTM charts with CEV and CM, the detection efficiency increases as \(\pi\) increases for most shift sizes, but the efficiency of some small shift sizes performs irregular changes. The CNN charts with CEV and CM have more excellent performance under ZS condition than under SS condition for most shift sizes. As \(\pi\) increases, the detection ability of the CNN charts with CEV and CM in most shift sizes slightly reduced but, in most cases of \(\delta\) = 0.8, the detection efficiency is significantly reduced.

When the mean shifts occur after the process has been running for a long time, CNN chart will be the best choice for gamma type-I censored data unless there is a need to detect tiny shift sizes. The CNN chart with CEV is suitable for the gamma censored data of lower rates or smaller skewness values, and the CNN chart with CM is recommended for monitoring the moderately, and highly censoring gamma data of larger skewness values.

Real-world case study

A reliability life test for a liquid–crystal display module (LCM) was conducted at a temperature of 70 °C with 80% relative humidity. Based on historical data analysis, the lifetime distribution of an LCM is known to follow a gamma distribution with shape parameter \({a}_{0}=5.72\) and scale parameter \({b}_{0}=0.48\). To save testing time and cost, practitioners use the censoring rate \(Pc=0.8\) to conduct the test and the censoring time \({c}_{T}\) is found to be 1.76 h. The skewness coefficient value of this lifetime distribution is 0.35, which approximates a symmetrical distribution, therefore EWMA CEV statistics is selected for monitoring the LCM’s lifetime. The CEV of this lifetime distribution can be calculated as 3.09. According to quality inspection regulations, five units of each batch of LCMs must be randomly sampled to test lifetime values. Because the EWMA chart based on the CNN network has better performance than the EWMA charts and the EWMA chart on the LSTM network, practitioners developed a CNN-based EWMA chart with CEV using \(\lambda =0.1\) and in-control ARL = 200. As shown in Fig. 1, practitioners trained a CNN network using the EWMA CEV statistics in Eq. (4) and obtained the optimal threshold value \({\eta }^{*}=-0.186\) for the in-control ARL value of 200.

In line with the above, practitioners tested five units from each batch under conditions of 70 °C and 80% humidity, and halted testing when the test time reached 1.76 h. The EWMA statistics (Eq. (4)) were inputted into the well-trained CNN network to predict the statistic of the next period. Table 7 showed the lifetime data of testing units for 30 batches. In the 101st batch, only one tested unit failed at 1.23 h and other four tested units did not fail. The lifetime of four unfailed units is recorded as CEV. The actual value \({E}_{j}\) in Table 7 can be obtained by using Eq. (4). As shown in Table 7, the EWMA CEV statistic of the 101st batch was 2.75, so the practitioners inputted 2.75 into the well-trained CNN network, which outputted the predicted value of 2.82 for the 102nd batch. After the 102nd batch was produced and the life tests of the five units reached the termination time of 1.76 h, the actual value \({E}_{j}\) of this batch was determined to be 2.75 by using Eq. (4) and the error value of this batch was − 0.07.

Table 7 Predicted and error values of the CNN network for the LCM lifetime test.

In Table 7, other error values were obtained with the same method. Figure 3a plots these error values for the CNN-based control chart. It can be seen that the error values of batches 119–130 were below the \({\eta }^{*}\) value, and so this chart indicates the variation at the 119th batch. Figure 3b also shows the EWMA CEV chart with LCL = 2.56 and its in-control ARL value is approximately 200. The EWMA CEV chart signals the variation at the 121st batch, detecting the same variation more slowly than the CNN-based EWMA chart with CEV.

Figure 3
figure 3

Control charts for monitoring the LCM’s censored data.

Conclusions

The combination of deep learning methods and control charts has greatly improved the efficiency of process monitoring. However, poor efficiency in the monitoring of high type-I censored data using control charts is a challenge for practitioners. This study proposed a control chart based on deep learning methods with EWMA CEV and CM statistics to detect the mean lifetime reduction for gamma type-I censored data. The ZS and SS ARL values of the proposed charts were also measured. Comparing the ZS and SS ARL values of the EWMA chart and the two EWMA charts based on deep learning methods with CEV and CM, CNN-based EWMA chart outperforms other control charts under ZS condition. For SS condition, the EWMA charts based on CNN with CEV and CM outperformed the other charts for various skewness coefficient values and censoring rates for most shift sizes. The EWMA charts with CEV and CM was slightly better than the CNN-based EWMA charts with CEV and CM for a few tiny shift sizes. The EWMA charts based on LSTM with CEV and CM consistently had the worst performance under ZS and SS conditions. In addition, a real-world case study showed that the CNN-based EWMA chart detected mean lifetime reduction more efficiently than the traditional EWMA CEV chart.

For the gamma censored data of lower rates or smaller skewness coefficient values, the CNN-based EWMA chart with CEV is the best choice, and the CNN-based EWMA chart with CM is recommended monitoring the moderately, and highly censoring data of heavily skewed gamma distribution. Future work could extend current approaches to combine CUSUM CEV and CM statistics with deep learning methods to monitor the censored data with normal or non-normal distributions. In addition, there are opportunities to combine multiple deep learning methods to build control charts.