Abstract
Progressive-stress accelerated life testing (PSALT) is a specialized experimental method that evaluates the longevity of a product under continuously fluctuating stress levels. Due to the constraints of testing equipment and expenses, the lifetime data collected by PSALT are typically censored. This paper introduces the PSALT model that utilizes Type-II unified progressive hybrid censoring to address this data characteristic, specifically when the lifespan of test units follows a truncated Cauchy power exponential (TCPE) distribution. The distribution’s scale parameter follows the inverse power law, and the cumulative exposure model is relevant for the effects of differing stress levels. The estimation methods for the TCPE parameters and the acceleration factor are examined, including maximum likelihood and Bayesian estimation techniques. Bayesian estimates are generated using the Markov chain Monte Carlo technique based on symmetric and asymmetric loss functions. The highest posterior density intervals are assessed, as well as asymptotic confidence intervals. A simulation study is conducted to evaluate the efficacy of the proposed point and interval estimators. Ultimately, a real data set is applied to the TCPE distribution, and the proposed estimators are assessed.
Similar content being viewed by others
Introduction
Scientific and technological advancements have resulted in the development of intricate and highly durable products, including lamp LEDs, computers, and silicone seals. Product quality is contingent upon reliability; consequently, manufacturers allocate substantial resources to testing, design, and other initiatives that guarantee consistent functionality. The ideal scenario would entail an abundance of life-testing data. However, a challenge arises with highly reliable products, as their extended lifespans frequently result in very few or even zero failures within a reasonable testing period under normal operating conditions.
Accelerated life tests (ALTs) have become a critical process that enables researchers to describe the life characteristics of products (systems or components), causing them to fail more rapidly than the normal operating condition. Higher stress loadings in ALT can be applied in various ways, with the three most common being constant-stress, step-stress, and progressive-stress. Nelson1 and Meeker and Escobar2 provided further information about ALTs. In constant-stress ALT, the products are tested under a constant stress level until the test terminates. Test termination can be determined based on the requirements, such as by establishing a threshold for the number of failures or by establishing a time limit for the test. Several authors investigated constant stress ALTs, for example, see Hassan3, Hassan and Al-Ghamdi4, Lin et al.5, Shi et al.6, Hassan et al.7, Ling8, Nassar and Alam9, Wang et al.10, Yousef et al.11, and Alomani et al.12.
Step-stress ALT products with increasing stress levels step by step by fixing the stress change criteria, which can be either the stress changing time or the number of failures. Several authors, including Ling and Hu13, Klemenc and Nagode14, Alam and Ahmed15, Chen et al.16, Yousef et al.17, and Hassan et al.18 have considered the study of step-stress ALT. In a progressive-stress ALT (PSALT), the amount of stress placed on the test product gradually increases over time. A ramp test is an ALT with linearly increasing stress. Starr and Endicolt19 evaluated capacitors using the ramp method. Solomon et al.20 and Chan21 applied the ramp test to insulations and integrated circuits, respectively. For PSALT, see22,23,24,25,26.
Predicting remaining useful life (RUL) for products or systems is usually the ultimate aim of such life-testing and dependability analysis. Recent studies, for example, investigate RUL prediction using sophisticated stochastic process models for multivariate and dependent degradation data27 or for systems displaying multi-phase degradation28. Moreover, data-driven solutions including federated learning with adaptive sampling are under development for challenging uses including aircraft engine RUL prediction29.
Censoring is widely used in life and reliability tests, as it is a result of the constraints of time and budget. Two of the most frequently employed censoring schemes are Type-I and Type-II. The hybrid censored scheme is a combination of the Type-I censored scheme and the Type-II censored scheme. The Type-I hybrid censoring and Type-II hybrid censoring schemes were also proposed by Epstein30,31. It is a more advanced type of censoring that combines Type-I and Type-II censoring, where the experiment will be terminated at \(min(Z_{m:n}, \tau )\). Merging the Type-II progressive censoring and hybrid censoring schemes produce the so called progressive hybrid censoring scheme (PHCS), which allows us to remove the reliable units at any time, and the experiment ends at \(min(Z_{m:n}, \tau )\), see Kundu and Joarder32 and Childs et al.33. However, the PHCS has the drawback of not being applicable when only a few failures are likely to occur before time \(\tau\). Cho et al.34 proposed a Type-I generalized PHCS, while Lee et al.35 proposed a Type-II generalized PHCS. While generalized PHCS outperform Type-I and Type-II PHCS, they do have significant drawbacks. To address some of the shortcomings of these schemes, GĂłrny and Cramer36 developed Type-II unified UPHCS (Type-II UPHCS), a general type of generalized PHCS. The lifetime experiment will be completed under Type-II UPHCS with a minimum of k number of unit failures, which ensures that the statistical inference is carried out more effectively. The UHCS has been studied for a variety of lifetime distributions, as demonstrated by the research of Jeon and Kang37, Shahrastani and Makhdoom38, and Lone and Panahi39.
Suppose n identical items are put to the test in a life test. One way to characterize the Type-II UPHCS is as follows: Let integer \(k,m\in {1,2,...,n}\) and \(\tau _{1}, \tau _{2} \in (0,\infty )\) be prefixed such that \(\tau _{1}<\tau _{2}\) and \(k<m\) with \(R=(R_{1}, R_{2},..., R_{m})\) is also prefixed integers satisfying \(n=m+R_{1}+R_{2}+...+R_{m}\). When the first unit fails, \(R_{1}\) is randomly removed from the remaining units. In the event of a second failure \(R_{2}\), the remaining units are removed and so on. Experiment termination occurs at \(min \{max(Z_{m:m:n},\tau _{1}),\tau _{2}\}\) if the \(k^{th}\) failure occurs prior to time \(\tau _{1}\). In the event that the \(k^{th}\) failure occurs between \(\tau _{1}\) and \(\tau _{2}\), the experiment is terminated at \(min(Z_{m:m:n}, \tau _{2})\), and if the \(k^{th}\) failure occurs after \(\tau _{2}\), the experiment is terminated at \(Z_{k:n}\). Through this censoring methodology, we can ensure that the experiment will conclude within a maximum duration of \(\tau _{2}\), with a minimum of k failures; otherwise, we can assure precisely k failures. Let \(C_{1}\) and \(C_{2}\) represent the numbers of observed failures by time \(\tau _{1}\) and \(\tau _{2}\), respectively. Furthermore, \(c_{1}\) and \(c_{2}\) represent the observed values of \(C_{1}\) and \(C_{2}\), respectively.
We have one of the following types of observations under the Type-II UPHCS described above: the experiment ends at \(min \{max(Z_{m:m:n},\tau _{1}),\tau _{2}\}\) if the \(k^{th}\) failure occurs before time \(\tau _{1}\). In that case, we have the first three subcases in Table 1. The experiment will terminate at \(min \{max(Z_{k:m:n},\tau _{2}),Z_{m:m:n}\}\) if time \(\tau _{1}\) occurs before the \(k^{th}\). The remaining three subcases are listed in Table 1. As a result, the six scenarios outlined in Table 1 are encountered within the Type-II UPHCS framework. Furthermore, Figure 1 illustrates the Type-II UPHCS mechanism.
If \(\textbf{Z}\) is a Type-II UPHCS from a distribution with the probability density function (PDF) f(z) and the cumulative distribution function (CDF) F(z), then the likelihood function based on the Type-II UPHCS can be written as
Thus, these scenarios may be amalgamated to yield the following:
where
The motivation for employing UPHCS stems from its flexibility and efficiency in handling real-life constraints in life-testing experiments. Unlike Type-I censoring (fixed time) or Type-II censoring (fixed number of failures), UPHCS incorporates both a minimum number of failures (k) and dual time limits (\(\tau _1, \tau _2\)), allowing experiments to terminate early without sacrificing statistical information.
Moreover, UPHCS enhances the classical PHCS by enabling better control over test duration and ensuring sufficient failure data, even when failure events are sparse. It also supports progressive removals, allowing units to be withdrawn at specific failure times, thereby improving resource efficiency.
To highlight its advantage, we compare UPHCS with other schemes in Table 2.
Furthermore, we conducted a simulation comparing the performance of UPHCS with PHCS and Type-II schemes under the same sample size and censoring proportion. The results, summarized in “Simulation study”, show that UPHCS consistently achieves lower bias and mean squared error (MSE) in parameter estimation. This demonstrates the practical and statistical superiority of UPHCS in accelerated life testing experiments.
The novelty of this paper is that there is no previous study related to truncated Cauchy power exponential (TCPE) distribution under PSALT with Type-II UPHCS. Our goal is to investigate the estimation problem using the classical and Bayesian approaches. The Markov chain Monte Carlo (MCMC) method is employed to obtain Bayesian estimates under distinct loss functions. This paper employs a real-world data analysis to illustrate the practical application of the proposed methods. Additionally, a simulation study is conducted to gauge the efficacy of a variety of estimators with an emphasis on bias, MSE for point estimation, and length for confidence intervals. All of these contributions collectively enhance the methodological tools for the analysis of intricate reliability data using ALTs.
The remainder of this paper is organized as follows. The model description and its assumption are given in “Model description and test assumption”. The maximum likelihood estimates (MLEs) are evaluated in “Maximum likelihood estimation”. The Bayes estimates are developed using MCMC technique in “Bayesian analysis”. The highest posterior density (HPD) interval, asymptotic confidence interval (ACI), and Bootstrap confidence interval (BCI) are created in “Interval estimation”. A Monte Carlo simulation is carried out in “Simulation study” to examine how well the suggested methods work. In “An application”, a real data set is analyzed as an example of the discussed methods. In “Summary and conclusion”, some final observations are summarized.
Model description and test assumption
In this section, an overview of the model under consideration is provided. Also, PSALT model is constructed according to the assumptions outlined in the second subsection.
Truncated cauchy power exponential distribution
The truncated Cauchy power-G (TCP-G) family was designed to be rich and contain many new distributions that may be interesting from a statistical perspective (with different supports, numbers of parameters, properties, etc). Here, we concentrate on one defined member of the TCP-G family using the exponential distribution as the baseline. It is referred to as the TCPE distribution for the purposes of this study.
Recently, Aldahlan et al.40 proposed a new class of distributions called the TCP-G class. It is an extension of the Cauchy distribution. The CDF, PDF, and hazard rate function (HR), respectively, are defined as:
where \(G(z;\theta )\) and \(g(z;\theta )\) are the CDF and PDF of a baseline continuous distribution with \(\theta\) as parameter vector and \(\lambda\) is a shape parameter, respectively.
The CDF and PDF of the exponential distribution are defined by \(G(z;\theta )=1-e^{-\theta z}, z> 0\) and \(g(z;\theta )=\theta e^{-\theta z}\). Therefore, by substituting this CDF, PDF, and HR into (2), the CDF, PDF, and HR of TCPE(\(\lambda , \theta\)) can be defined as
and
where \(\lambda\) and \(\theta\) are shape and scale parameters, respectively. Figure 2 presents the PDF and HR plots of the TCPE distribution. The PDF of the TCPE distribution can take different shapes, including several asymmetric shapes, unimodal, and reversed J-shaped. The HRF can be increasing, decreasing, or constant.
Data structure and assumptions
The effect of stress changes from one stress level to another is formed by using a linear cumulative exposure model, for more details, see Nelson41. When the applied stress V is a function of time, \(V = V(z)\), and affects the scale parameter \(\theta\) of the underlying distribution, then \(\theta\) becomes a function of time, \(\theta (z) = \theta (V(z))\).
A test unit’s CDF under progressive stress V(z) can be expressed as
where the linear cumulative exposure model is;
For estimation under PSALT based on TCPE distribution, we make the following assumptions.
-
(1)
Under normal operating conditions, a unit’s lifetime is provided by TCPE distribution.
-
(2)
With two parameters, \(\eta\) and \(\zeta\), the parameter \(\theta\) follows the inverse power law. This implies that \(\theta (z)=\frac{1}{\eta [V(z)]^{\zeta }}\), where \(\eta > 0\) and \(\zeta > 0\) are parameters to be estimated.
-
(3)
The progressive stress V(z) has a direct relationship with time and a constant rate \(s\), i.e., \(V(z)=sz, s >0\).
-
(4)
The n items to be tested are split into \(\hbar (\ge 2)\) groups during the testing process. The progressive stress is applied to each group, which contains \(n_{u}\) items. In light of this, when \(u = 1,..., \hbar\), \(V_{u}(z)=s_{u}z, \qquad 0< s_{1}<...<s_{\hbar }\).
-
(5)
The failure mechanisms of a unit are the same regardless of the stress level \(V_{u}, u=1,...,\hbar\).
-
(6)
The statistically independent random variables \(Z_{u,1}, Z_{u,2},..., Z_{u,n_{u}}\) denote the \(n_{u}\) failure times in group u for \(u = 1,..., \hbar\).
By applying Assumptions 2 and 4, the cumulative exposure model (6) can be reformulated as follows:
The CDF (5) of TCPE distribution under progressive-stress \(V_{u}(z)\), is given by
and the corresponding PDF is given by
Maximum likelihood estimation
This section discusses the MLE of the parameters \(\zeta , \eta ,\) and \(\lambda\) under PSALT when the data are Type-II UPHCS. Let \(\underline{Y} = (y_{1,c^{*},n},...,y_{uv,c^{*},n})\) under the stress level \(V_{u}(z), u=1,...,\hbar\) denotes the Type-II UPHCS of size \(c^{*}\) from a sample of size n drawn from a TCPE(\(\lambda , \theta\)). Upon substituting (8) and (9) in (1), the likelihood function of the set of parameters \(\vartheta =( \eta , \zeta , \lambda )\) can be expressed as:
We replace \(y_{uv,c{*},n}\) by \(y_{uv}\) for simplicity of notation. Hence, we can rewrite (10) as follows:
where
and \(\psi (\tau _{1})\) is the same as in (12) with \(y_{uv}=\tau _{1}\). Similarly for \(\psi (\tau _{2})\).
Then, the log-likelihood function can be written as
The log-likelihood equations for the parameters \(\eta , \zeta\) and \(\lambda\) are, respectively, given by
where
and \(\psi _{2}(\tau _{1}), \psi ^{*}_{2}(\tau _{1}), \psi _{3}(\tau _{1})\) are the same as in (15) with \(y_{uv}=\tau _{1}\). Similarly for \(\psi _{2}(\tau _{2}), \psi ^{*}_{2}(\tau _{2})\) and \(\psi _{3}(\tau _{2})\). Since the likelihood equations of \(\eta , \zeta ,\) and \(\lambda\) are nonlinear and solving them analytically is difficult, some numerical techniques such as the Newton–Raphson method can be used to obtain the MLEs of the distribution parameters \(\eta , \zeta ,\) and \(\lambda\), namely \(\hat{\eta }, \hat{\zeta },\) and \(\hat{\lambda }\).
Bayesian analysis
In this section, Bayesian analysis is considered with different prior and posterior distributions.
Prior and posterior distribution
To precede the Bayesian analysis, prior distributions for each unknown parameter must be taken into consideration. Here, in order to make the Bayesian analysis comparable with the likelihood-based analysis formed in “Maximum likelihood estimation”, we assume that the three parameters \(\eta , \zeta ,\) and \(\lambda\) have gamma prior, thus
Assuming that the parameters \(\eta , \zeta\), and \(\lambda\) are independent, the joint prior PDF is given by
Together with the likelihood function in (11), by using the Bayesian theorem, the joint posterior density of \(\eta , \zeta\), and \(\lambda\) can be written as follows
Symmetric and asymmetric loss functions
Two different types of loss functions for Bayesian estimation are examined in this study. The initial one is the squared error loss function (SELF), a symmetric function that values both overestimation and underestimation equally when estimating parameters. The second option is the linear exponential loss function (LLF), which is asymmetric and provides various weights according to overestimation and underestimation.
The Bayes estimate of the function of parameters \(D=D(\vartheta )\) based on SELF is given by
The Bayes estimate under LLF of D is given by
where \(\nu \ne 0\) is the shape parameter of the LLF.
It should be noted that the Bayes estimates of D in (18) and (19) are analogous to a ratio of two multiple integrals that cannot be reduced analytically. In this regard, using an approximation technique to calculate requested estimates is recommended, as discussed in the following subsection.
MCMC method
Here, the MCMC method is employed to generate samples from the posterior distribution and then compute the Bayes estimates of D under PSALT. The conditional posterior distributions of \(\eta , \zeta\), and \(\lambda\) are derived using the joint posterior density function in (17).
Since the conditional posterior distributions of the parameters \(\eta ,\zeta\), and \(\lambda\) are not reducible into the forms of some well-known distributions, we use Metropolis–Hastings algorithm, see Upadhyay and Gupta42. To evaluate Bayes estimates of \(D=D(\lambda , \eta , \zeta )\) under SELF and LLF, the following procedure is utilized.
-
(1)
Assign some initial values for \(\eta ,\zeta\) and \(\lambda\).
-
(2)
Set \(j=1\).
-
(3)
Generate \(\eta \sim N(\eta _{j},\sigma _{11}), \zeta \sim N(\zeta _{j},\sigma _{22})\) and \(\lambda \sim N(\lambda _{j},\sigma _{33})\) where \(\sigma\) denotes the variance-covariance matrix.
-
(4)
Calculate
$$\begin{aligned} T_{j}=\frac{\pi ( \eta _{j}, \zeta _{j}, \lambda _{j} |\underline{\text {Y}})}{\pi ( \eta _{j-1}, \zeta _{j-1},\lambda _{j-1}|\underline{\text {Y}})}. \end{aligned}$$ -
(5)
Accept \(( \eta _{j}, \zeta _{j},\lambda _{j})\) with probability \(min(1,T_{j})\).
-
(6)
Repeat steps (3) to (5) \(\mathbb {M}\) times to obtain \(\mathbb {M}\) number of samples for the parameters \((\eta , \zeta , \lambda )\).
-
(7)
Compute the Bayes estimates under SELF and LLF of \(\eta ,\zeta ,\) and \(\lambda\), using (18), (19) as
$$\begin{aligned} \left. \begin{aligned} \tilde{\eta }_{SELF}&=\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}\eta _{j},\\ \tilde{\zeta }_{SELF}&=\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}\zeta _{j},\\ \tilde{\lambda }_{SELF}&=\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}\lambda _{j}. \end{aligned}\right\} \qquad \left. \begin{aligned} \tilde{\eta }_{LLF}&=\frac{-1}{\nu }\log \bigg [\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}e^{-\nu \eta _{j}}],\\ \tilde{\zeta }_{LLF}&=\frac{-1}{\nu }\log \bigg [\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}e^{-\nu \zeta _{j}}],\\ \tilde{\lambda }_{LLF}&=\frac{-1}{\nu }\log \bigg [\frac{1}{\mathbb {\acute{M}}}\sum _{j=1}^{\mathbb {\acute{M}}}e^{-\nu \lambda _{j}}]. \end{aligned}\right\} \end{aligned}$$(21)where \(\mathbb {\acute{M}}=\mathbb {M}-M_{0}\) and \(M_{0}\) is also known as burn-in sample.
Interval estimation
In this section, we construct ACIs using MLE’s normality property, the BCIs, and HPD credible intervals for the parameters \(\eta , \zeta\), and \(\lambda\).
Approximate confidence intervals
Since the MLEs of the unknown parameters, cannot be derived in a closed form, it is difficult to calculate the exact distribution of the MLEs. As a result, the exact CIs of the parameters cannot be obtained. Hence, the ACI for the parameters \(\lambda , \eta\), and \(\zeta\) are developed based on large sample approximation.
The ACIs of \(\vartheta =( \eta , \zeta , \lambda )\) based on the observed Fisher information matrix can be computed in accordance with the asymptotic normality results of the MLEs. The observed information matrix, which can be obtained by deriving the unknown parameters of the log-likelihood function, can be applied to approximate the asymptotical variance-covariance of the MLEs.
Then the approximate asymptotic variance-covariance matrix is
The MLEs of parameters approximately follow the multivariate normal distribution with mean \(\vartheta\) and variance-covariance matrix \(Cov(\hat{\vartheta })\), namely \(\hat{\vartheta }\sim N(\vartheta , Cov(\hat{\vartheta }))\). Therefore, for arbitrary \(0<\epsilon <1\), the \(100(1-\epsilon )\%\) ACIs of the unknown parameters can be expressed as follows:
where \(V(\hat{\vartheta })\) is the asymptotic variance of \({\hat{\vartheta }}\), \(\rho _{\frac{\epsilon }{2}}\) is the \((1-\epsilon )\) quantile of the standard normal distribution N(0, 1).
Bootstrap confidence intervals
A resampling technique for statistical inference is the bootstrap. Confidence interval estimation uses frequently (see Efron43 for more details). In this part, we create confidence intervals for the unknown parameters \(\eta , \zeta ,\) and \(\lambda\) using the parametric bootstrap approach. We presented the percentile bootstrap (BP) and bootstrap-T (BT) CI parametric bootstrap approaches.
Percentile bootstrap CI
-
1.
Compute the MLE of PSALT for parameters of TCPE distribution.
-
2.
Generate a bootstrap samples using \(\eta ,\zeta\) and \(\lambda\) to obtain the bootstrap estimate of \(\eta\) say \(\hat{\eta }^{*b}\) , \(\zeta\) say \(\hat{\zeta }^{*b}\) and \(\lambda\) say \(\hat{\lambda }^{*b}\) using the bootstrap sample.
-
3.
Repeat step (2) \(\mathcal {B}\) times to have \((\hat{\eta }^{*b(1)},\hat{\eta }^{*b(2)},\ldots ,\hat{\eta }^{*b(\mathcal {B})}), (\hat{\zeta }^{*b(1)},\hat{\zeta }^{*b(2)},\ldots \hat{\zeta }^{*b(\mathcal {B})})\) and
\((\hat{\lambda } ^{*b(1)},\hat{\lambda }^{*b(2)},\ldots ,\hat{\lambda }^{*b(\mathcal {B})})\).
-
4.
Arrange all values of \(\hat{\eta }^{*b}, \hat{\zeta }^{*b}\) and \(\hat{\lambda }^{*b}\) obtained in step 3 in an ascending order as \((\hat{\eta }^{*b[1]},\hat{\eta }^{*b[2]},\ldots ,\hat{\eta }^{*b([\mathcal {B}])}), (\hat{\zeta }^{*b[1]},\hat{\zeta }^{*b[2]},\ldots \hat{\zeta }^{*b[\mathcal {B}]})\) and \((\hat{\lambda } ^{*b[1]},\hat{\lambda }^{*b[2]},\ldots ,\hat{\lambda }^{*b[\mathcal {B}]})\).
-
5.
Two sided \(100(1 - \epsilon )\%\) BP confidence intervals for the unknown parameters \(\eta ,\zeta ,\lambda\) are given by \([\hat{\eta }^{*b([\mathcal {B}\frac{\epsilon }{2}])},\hat{\eta }^{*b([\mathcal {B}(1-\frac{\epsilon }{2})])}], [\hat{\zeta }^{*b([\mathcal {B}\frac{\epsilon }{2}])}, \hat{\zeta }^{*b([\mathcal {B}(1-\frac{\epsilon }{2})])}]\) and \([\hat{\lambda }^{*b([\mathcal {B}\frac{\epsilon }{2}])},\hat{\lambda }^{*b([\mathcal {B}(1-\frac{\epsilon }{2})])}]\).
Bootstrap-T CI
-
1.
Same as the steps (1,2) in BP.
-
2.
Compute the t-statistic of \(\vartheta\) as \(T_{k}=\frac{\hat{\vartheta }^{*b}-\hat{\vartheta }}{\sqrt{V(\hat{\vartheta }^{*b})}}\) where \(V(\hat{\vartheta }^{*b})\) is asymptotic variances of \(\hat{\vartheta }^b\) and it can be obtained using the Fisher information matrix.
-
3.
Repeat steps 2-3 \(\mathcal {B}\) times and obtain \((T_{k}^{(1)}, T_{k}^{(2)}, \ldots , T_{k}^{(\mathcal {B})})\).
-
4.
Arrange the values obtained in step 3 in an ascending order as \((T_{k}^{[1]}, T_{k}^{[2]}, \ldots , T_{k}^{[\mathcal {B}]})\).
-
5.
A two side \(100(1 - \epsilon )\%\) BT confidence intervals for the unknown parameters \(\eta ,\zeta ,\) and \(\lambda\) are given by \([\hat{\eta }+ T_k^{[\mathcal {B}\frac{\epsilon }{2}]}\sqrt{V(\hat{\eta })},\hat{\eta }+ \sqrt{V(\hat{\eta })} T_k^{[\mathcal {B}(1-\frac{\epsilon }{2})]}]\), \([\hat{\zeta }+ T_k^{[\mathcal {B}\frac{\epsilon }{2}]}\sqrt{V(\hat{\zeta })},\hat{\zeta }+ \sqrt{V(\hat{\zeta })} T_k^{[\mathcal {B}(1-\frac{\epsilon }{2})]}]\) and \([\hat{\lambda }+ T_k^{[\mathcal {B}\frac{\epsilon }{2}]}\sqrt{V(\hat{\lambda })},\hat{\lambda }+ \sqrt{V(\hat{\lambda })} T_k^{[\mathcal {B}(1-\frac{\epsilon }{2})]}]\).
Highest posterior density credible interval
The 95% HPD credible interval represents the range within which the true parameter value lies with 95% probability, based on the posterior distribution. Unlike equal-tailed intervals, the HPD interval is the shortest interval that contains the most probable values of the parameter, making it particularly informative for skewed or multimodal distributions. We compute the HPD interval using the posterior samples generated via MCMC methods. To compute the Bayesian HPD credible intervals of any function of \(\vartheta\), set the credible level to \(100(1-\epsilon )\%\) and apply the MCMC method described in “MCMC method” based on its point estimation. The following are the specific steps:
-
(1)
To get \(\mathbb {M}\) groups of samples, repeat Steps (1)-(6) of the MCMC method sampling in “MCMC method”.
-
(2)
Arrange \(\vartheta _{[j]}\), as \(\eta _{[1]}, \eta _{[2]},...,\eta _{[\mathbb {\acute{M}}]},\zeta _{[1]},\zeta _{[2]},...,\zeta _{[\mathbb {\acute{M}}]}\) and \(\lambda _{[1]},\lambda _{[2]},...,\lambda _{[\mathbb {\acute{M}}]}\), where \(\mathbb {\acute{M}}\) denotes the length of the generated simulation.
-
(3)
The \(100(1-\epsilon )\%\) HPD credible interval of \(\vartheta\) are acquired as:
$$\begin{aligned} \big (\vartheta _{[\mathbb {\acute{M}}\frac{\epsilon }{2}]}, \vartheta _{[\mathbb {\acute{M}}(1-\frac{\epsilon }{2})]} \big ). \end{aligned}$$For more details on HPD intervals, see Chen and Shao44, El-Saeed45, and Kruschke46.
Simulation study
A simulation analysis is considered in this section to compare the effectiveness of the traditional maximum likelihood and Bayesian estimation techniques under various schemes of Type-II UPHCS. The statistical program R-package was used to carry out extensive computations. To generate samples under Type-II UPHCS, different values of r, k, \(\tau _1\), and \(\tau _2\) have been determined as illustrated in the steps below:
-
The parameter values \((\lambda , \zeta , \eta )\) in this simulation are (\(\lambda =0.7, \zeta =1.4, \eta =1.6\)), (\(\lambda =0.7, \zeta =1.4, \eta =0.5\)), (\(\lambda =0.7, \zeta =0.4, \eta =0.5\)) and (\(\lambda =2, \zeta =1.4, \eta =2.5\)), and sample sizes \(n_1=50\), \(n_2=40\), \(n_3=30\), and \(n_4=20\) (\(\hbar =4\)) are taken into consideration.
-
Generate four samples from a uniform distribution (0, 1).
-
Generate TCPE samples by using the inverse of CDF, but this inverse hasn’t existed in mathematical form. So the iterative algorithm has been used by ’uniroot’ function in R-packages. Return four samples have TCPE distribution.
-
Select censoring size \(m_u\); \(u=1,\dots ,\hbar\) by using relative censoring size r as 0.6, and 0.8, where \(m_u=r\times n_u\), and \(k_u=\frac{m_u}{2}\).
-
Select the censoring scheme removal as:
Scheme 1: \(R_{m_u}=n_u-m_u\), and \(R_w=0~; w=1:m_u-1\) where \(u=1,\dots ,\hbar\).
Scheme 2: \(R_{1_u}=n_u-m_u\), and \(R_w=0~; w=2:m_u\) where \(u=1,\dots ,\hbar\).
Scheme 3: \(R_{1_u}=\frac{n_u-m_u}{2}\), \(R_{m_u}=\frac{n_u-m_u}{2}\) and \(R_w=0~; w=2:m_u-1\) where \(u=1,\dots ,\hbar\).
-
The algorithm technique in26 has produced a Type-II PC from TCPE distribution with a size of \(n_u\) and a censoring size of \(m_u\).
-
If the generated Type-II PC data is \((z_{1,m_u,n_u}, z_{2,m_u,n_u}, \dots , z_{m_u,m_u,n_u}),\) then \(z_{m_u,m_u},n_u <\tau _1\), we set \(R_{m_u} = 0\) and use the transformation which was suggested by Ng et al.47.
-
Now, we have \(m_u\) observation as Type-II PC data as the following
\((z_{1,m_u,n_u}, z_{2,m_u,n_u}, \dots , z_{m_u,m_u,n_u}, z_{m_u+1,n_u},\dots , z_{m_u+R{m_u},n_u})\). Then, as stated in “Model description and test assumption”, we determined the experiment’s end time and the matching Type-II UPHC data that were observed.
For the point estimate, we calculated the MSE and estimated bias for MLE and the Bayesian estimates of \(\lambda , \zeta ,\) and \(\eta\), using the gamma informative prior under SELF, LLF (with \(\nu =-0.5\)), and LLF (with \(\nu =0.5\)). Using 5,000 simulations, we also create the length of ACIs (LAICs), length of Bootstrap-P CIs (LBPCI), length of Bootstrap-T CIs (LBTCI), coverage probabilities (CP), and Bayesian credible ranges for \(\lambda , \zeta\) and \(\eta\) for 95% CIs. We can use the estimate and variance-covariance matrix of the MLE approach to get appropriate and superior values for the independent joint prior’s hyper-parameters. The estimated hyper-parameters can be calculated as follows by equating the mean and variance of gamma priors as the following equations:
where, I is the number of iteration and \(\hat{\vartheta }^1=\hat{\lambda }\), \(\hat{\vartheta }^2=\hat{\zeta }\), and \(\hat{\vartheta }^3=\hat{\eta }\) .
Classical estimators can be assessed numerically by using an appropriate iterative approach, such as the Nelder-Mead (N-M) method, but they cannot be derived in explicit form. One can use the R statistical programming language software by creating the “maxLik” package, which uses the N-M approach of maximisation in the maximum likelihood computations, developed by Henningsen and Toomet48, to calculate the desired MLE of \(\lambda , \zeta ,\) and \(\eta\) for any given data set. We simulated 12,000 MCMC samples and disregarded the first 2000 iterations as burn-in to obtain the Bayes point estimates along with their HPD interval estimates of the same unknown parameters using the “coda” tools of the R programming language.
The following comments can be concluded based on Tables 3, 4, 5 and 6:
Table 3 illustrates that the MLE of the parameter \(\lambda\) demonstrates a substantial negative bias, particularly at the higher censoring rate (\(r = 0.8\)). This bias is significantly mitigated by Bayesian estimators. The limitations of relying on asymptotic approximations in this scenario are underscored by the poor coverage provided by ACI, particularly for parameter \(\lambda\), as indicated by scheme 1. Additionally, scheme 3 generates less bias than schemes 1 and 2. Furthermore, LLF performs well when censoring is low and performs poorly when censoring is high. Consequently, this may suggest that the LLF will perform poorly on smaller datasets where censoring will increase.
Table 4 indicates that the performance of MLE for \(\eta\) is significantly subpar, with a high MSE, particularly when \(r=0.8\). Significant enhancements in MSE are provided by the Bayesian approach. Additionally, the coverage probabilities of both bootstrap methods are nearly 95% . The decision between BP and PT is arbitrary due to the similar interval lengths and coverage rates. It is important to note that the coverage rates of HPDs in the table for a variety of scenarios are less than 90%. This suggests that HPDs provide subpar estimates in this dataset.
Table 5 summarizes that scheme 3 exhibits a reasonable MSE and outperforms schemes 1 and 2. Schemes 1 and 2 are generally more biased than scheme 3. Across all schemes, Bootstrap confidence intervals maintain acceptable performance. In contrast to the SELF, LLF with \(\nu <0\) and \(\nu >0\) exhibit superior bias performance. The coverage and bias of all estimators are reasonable.
Table 6 illustrates that the asymptotic assumptions are less reliable for this parameter due to the higher MSE for \(\eta\) under MLE. In terms of improved performance, Bayesian methods offer substantial improvements in terms of reduced bias and MSE. Scheme 1 expands the credible ranges of HPD. Scheme 3 demonstrates an unacceptable low coverage rate and a substantial MSE. It is important to note that the loss function selection had a minimal impact on the tested values. This may imply that any loss function can be reasonably applied.
Across all tables, the censoring rate increases the bias and MSE for all estimators and potentially degrades the coverage probabilities of ACIs across all tables. Consequently, it is necessary to consider the censoring rate. Changes in the censoring level are less sensitive to the Bayesian estimates, which increases their resilience. More specifically, MLEs are comparable in situations where censoring is low (\(<=30\%\)), but they are not comparable when censoring levels are high (\(>=50\%\)). Estimates that are calculated using Bayesian estimators are more reliable. The loss function selection has a moderate effect, as LLF may occasionally decrease MSE but may potentially impact coverage. It is contingent upon the specific application and the parameter value that is the subject of the practical implications. Generally, the HPD credible intervals are shorter than the ACIs; however, this is not always present. The reliability of the HPD intervals (as determined by coverage) should be meticulously evaluated. In order to reduce bias, employ LLF with \(\nu =-0.5\) to achieve low censoring. In the event of high censorship, SELF is the superior option for robustness.
An application
Now, using an actual data set analysed in49 and50, our aim is to illustrate the point and interval estimation techniques mentioned in this paper. The information shows the failure times (in hours) of electrolytic capacitors placed under two sets of Type-II UPHCS and having a size of 32 volts and 22 microfarads. 30 units make up each testing group. The following are the failure times (in hours):
First group (\(s_1 = 5.0417\)): “7.21, 10.24, 10.26, 10.37, 10.51, 10.56, 11.25, 11.28, 11.29, 11.35, 12.23, 12.25, 12.36, 12.57, 13.03, 13.04, 13.05, 13.27, 13.46, 13.49, 14.23, 14.45, 15.00, 15.43, 15.47, 16.55, 17.07, 17.21, 17.23, 18.49”.
Second group (\(s_2 = 5.833\)): “7.36, 7.55, 7.57, 8.00, 8.23, 8.46, 9.02, 9.03, 9.04, 9.22, 9.32, 9.34, 9.49, 10.28, 10.53, 11.33, 11.34, 11.54, 12.16, 12.53, 12.55, 13.20, 14.06, 14.21, 14.21, 14.21, 16.24, 16.41, 17.53, 21.26”.
Figure 3 discussed box plots, total time on test (TTT), hazard rates, and strip plots for each data set. The box plot and the strip chart were used to explain the data and ensure that there are no outliers and to see whether the data tends to the right or left side. Through these graphics, we notice that there are no outliers and that the data tends to the right side.
Before continuing, it is determined whether the TCPE distribution with CDF (8) is appropriate for fitting the aforementioned data using the statistical test of Kolmogorov-Smirnov (K-S) and its corresponding P-value for each group. The aforementioned information and CDF (8) can be used to demonstrate that the TCPE parameter estimates in Table 7 maximize the likelihood function of parameters. The MLEs of parameters and their standard error (SE) are calculated. Also, measures values of Akaike information criterion (AIC), the consistent AIC (CAIC), Bayesian IC (BIC), and Hannan-Quinn IC (HQIC). Table 7 provides the K-S, Cramér–von Mises (CVM), Anderson–Darling (AD) test statistics, and the P-value are provided. By the result in Table 7, we note that data are fitted for this model where P-values are larger than 0.05. Also, these results are better than the results presented in50 for the Rayleigh and Weibull distributions.
The profile likelihood plots in Figures. 4 and 5 clearly demonstrate that the MLEs are unique and correspond to the highest values of the likelihood function, confirming the presence of well-defined peaks for each parameter. The TCPE distribution fits the data better than the Weibull distribution (see50), as indicated by the P-value for the corresponding TCPE distribution being higher than that calculated for Weibull distribution. Additionally, Figure. 6 shows that the TCPE distribution fits these data rather well.
Point estimates of the parameters based on the Type-II UPHCS introduced from the provided real data sets have been derived and summarized in Table 9 using data selected of Type-II UPHCS as obtained in Table 8.
By results in Table 9, we note that scheme 2 has smaller values of ACI, CACI, BIC, and HQIC.
Summary and conclusion
In many different sectors, accelerated life testing has been used to rapidly collect failure time data for test units in a significantly shorter period than testing under typical operating circumstances. A PSALT under Type-II UPHC is taken into consideration in this article when the lifetime of test units follows TCPE distribution. The scale parameter of the distribution follows the inverse power law, and the cumulative exposure model accounts for the effect of fluctuating stress. Using Type-II UPHC, the maximum likelihood estimates are compared with the Bayesian estimates of the unknown parameters based on symmetric and asymmetric loss functions via MCMC technique. We also provide some interval estimators of the unknown parameters including asymptotic intervals, bootstrap intervals, and highest posterior density intervals. Simulations are used to compare the accuracy of the maximum likelihood estimates with the Bayesian estimates. Also, to evaluate the effectiveness of the proposed confidence intervals for various parameter values and sample sizes. Analysis to real data has been examined.
Data availability
The data used to support the findings of this study are included within the article.
References
Nelson, W.B. Accelerated testing: statistical models, test plans, and data analysis. JWS., (2009).
Meeker, W. & Escobar, L. Statistical Methods for Reliability Data (Wiley, New York, NY, USA, 1998).
Hassan, A. S. Estimation of the generalized exponential distribution parameters under constant-stress partially accelerated life testing using Type I censoring. The Egypt. Stat. J. 51(2), 48–62 (2007).
Hassan, A. S. & Al-Ghamdi, A. S. Optimun step stress accelerated life testing for Lomax distribution. J. Appl. Sci. Res. 5(12), 2153–2164 (2009).
Lin, C. T., Hsu, Y. Y., Lee, S. Y. & Balakrishnan, N. Inference on constant stress accelerated life tests for log-location-scale lifetime distributions with type-I hybrid censoring. JSCS 89(4), 720–749 (2019).
Shi, X., Lu, P. & Shi, Y. Reliability estimation for hybrid system under constant-stress partially accelerated life test with progressively hybrid censoring. Recent Pat. Eng. 14(1), 82–94 (2020).
Hassan, A. S., Pramanik, S., Maiti, S. & Nassr, S. G. Estimation in Constant Stress Partially Accelerated Life Tests for Weibull Distribution Based on Censored Competing Risks Data. Ann. Sci. 7(1), 45–62 (2020).
Ling, M. H. Optimal Constant-Stress Accelerated Life Test Plans for One-Shot Devices with Components Having Exponential Lifetimes under Gamma Frailty Models. Mathematics 10(5), 840 (2022).
Nassar, M. & Alam, F. Analysis of Modified Kies Exponential Distribution with Constant Stress Partially Accelerated Life Tests under Type-II Censoring. Mathematics 10(5), 819 (2022).
Wang, X., Wang, B. X., Liang, W. & Jiang, P. Inference for constant stress accelerated life test under the proportional reverse hazards lifetime distribution. Qual. Reliab. Eng. Int. 38(8), 4223–4235 (2022).
Yousef, M. M., Alyami, S. A. & Hashem, A. F. Statistical inference for a constant-stress partially accelerated life tests based on progressively hybrid censored samples from inverted Kumaraswamy distribution. PLoS ONE 17(8), e0272378 (2022).
Alomani, G. A., Hassan, A. S., Al-Omari, A. I. & Almetwally, E. M. Different estimation techniques and data analysis for constant-partially accelerated life tests for power half-logistic distribution. Sci. Rep. 14(1), 20865 (2024).
Ling, M. H. & Hu, X. W. Optimal design of simple step-stress accelerated life tests for one-shot devices under Weibull distributions. Reliab. Eng. Syst. Saf. 193, 106630 (2020).
Klemenc, J. & Nagode, M. Design of step-stress accelerated life tests for estimating the fatigue reliability of structural components based on a finite-element approach. Fatigue Fract. Eng. Mater. Struct. 44(6), 1562–1582 (2021).
Alam, I. & Ahmed, A. Inference on maintenance service policy under step-stress partially accelerated life tests using progressive censoring. J. Stat. Comput. Simul. 92(4), 813–829 (2022).
Chen, L. S., Liang, T. & Yang, M. C. Designing efficient Bayesian sampling plans based on the simple step-stress test under random stress-change time for censored data. J. Stat. Comput. Simul. 93(7), 1104–1129 (2023).
Yousef, M. M., Alsultan, R. & Nassr, S. G. Parametric inference on partially accelerated life testing for the inverted Kumaraswamy distribution based on Type-II progressive censoring data. Math. Biosci. Eng. 20(2), 1674–1694 (2023).
Hassan, A., Hagag, A. E., Metwally, N. & Sery, O. Statistical Analysis of Inverse Weibull based on Step-Stress Partially Accelerated Life Tests with Unified Hybrid Censoring Data. Comput. J. Math. Stat. Sci. 4(1), 162–185. https://doi.org/10.21608/cjmss.2024.319502.1072 (2025).
Starr, W. T. & Endicolt, H. S. Progressive stress-a new accelerated approach to voltage endurance. Trans. AIEE, Part III: Power Appar. Syst. 80(3), 515–522 (1961).
Solomon, P., Klein, N. & Albert, M. A statistical model for step and ramp voltage breakdown tests in thin insulators. Thin Solid Films 35(3), 321–326 (1976).
Chan, C. K. A proportional hazard approach to accelerate SiO2 breakdown voltage & time distributions. IEEE Trans. Reliab. 39, 147–150 (1990).
Chen, W. H., Yang, F., Qian, P., Pan, J. & He, Q. C. Novel accelerating life test method and its application by combining constant stress and progressive stress. Chin. J. Mech. Eng. 31(1), 1–8 (2018).
Wang, R., Gu, B. & Xu, X. Reliability statistical analysis about products of Birnbaum-Saunders fatigue life distribution for life test and progressive stress accelerated life test. Commun. Stat. Simul. Comput. 47(8), 2304–2331 (2018).
Hashem, A. F., Alyami, S. A. & Abdel-Hamid, A. H. Inference for a Progressive-Stress Model Based on Ordered Ranked Set Sampling under Type-II Censoring. Mathematics 10(15), 2771 (2022).
Abushal, T. A. & Abdel-Hamid, A. H. Inference on a new distribution under progressive-stress accelerated life tests and progressive type-II censoring based on a series-parallel system. AIMS Math. 7(1), 425–454 (2022).
Mahto, A. K. et al. Bayesian estimation and prediction under progressive-stress accelerated life test for a log-logistic model. Alex. Eng. J. 101, 330–342 (2024).
Xu, A., Wang, R., Weng, X., Wu, Q. & Zhuang, L. Strategic integration of adaptive sampling and ensemble techniques in federated learning for aircraft engine remaining useful life prediction. Appl. Soft Comput. 175, 113067 (2025).
Xu, A., Fang, G., Zhuang, L., & Gu, C. A multivariate student-t process model for dependent tail-weighted degradation data. IISE Trans. 1–17 (2024).
Zhuang, L., Xu, A., Wang, Y. & Tang, Y. Remaining useful life prediction for two-phase degradation model based on reparameterized inverse Gaussian process. Eur. J. Oper. Res. 319(3), 877–890 (2024).
Epstein, B. Truncated life tests in the exponential case. Ann. Math. Stat. 555–564 (1954).
Epstein, B. Estimation from life test data. Technometrics 2(4), 447–454 (1960).
Kundu, D. & Joarder, A. Analysis of Type-II progressively hybrid censored data. Comput. Stat. Data Anal. 50(10), 2509–2528 (2006).
Childs, A., Chandrasekar, B., & Balakrishnan, N. Exact likelihood inference for an exponential parameter under progressive hybrid censoring schemes. In Statistical models and methods for biomedical and technical systems (pp. 319-330). Birkhäuser Boston, (2008).
Cho, Y., Sun, H. & Lee, K. Exact likelihood inference for an exponential parameter under generalized progressive hybrid censoring scheme. Stat. Methodol. 23, 18–34 (2015).
Lee, K., Sun, H. & Cho, Y. Exact likelihood inference of the exponential parameter under generalized Type II progressive hybrid censoring. J. Korean Stat. Soc. 45(1), 123–136 (2016).
Górny, J. & Cramer, E. Modularization of hybrid censoring schemes and its application to unified progressive hybrid censoring. Metrika (Springer) 81(2), 173–210 (2018).
Jeon, Y. E. & Kang, S. B. Estimation of the Rayleigh distribution under unified hybrid censoring. Austrian J. Stat. 50(1), 59–73 (2021).
Shahrastani, S.Y., & Makhdoom, I. Estimating E-Bayesian of parameters of inverse Weibull distribution using an unified hybrid censoring scheme. Pak. J. Stat. Oper. Res. 113–122 (2021).
Lone, S.A., & Panahi, H. Estimation procedures for partially accelerated life test model based on unified hybrid censored sample from the Gompertz distribution. Eksploatacja i Niezawodność 24(3), (2022).
Aldahlan, M. A., Jamal, F., Chesneau, C., Elgarhy, M. & Elbatal, I. The truncated Cauchy power family of distributions with inference and applications. Entropy 22(3), 346 (2020).
Nelson, W. B. Accelerated testing: statistical models, test plans, and data analysis. JWS, (1990).
Upadhyay, S. K. & Gupta, A. A Bayes analysis of modified Weibull distribution via Markov chain Monte Carlo simulation. J. Stat. Comput. Simul. 80(3), 241–254 (2010).
Efron, B. Bootstrap methods: another look at the jackknife. In Breakthroughs in statistics 569–593 (Springer, New York, NY, 1992).
Chen, M. H. & Shao, Q. M. Monte Carlo estimation of Bayesian credible and HPD intervals. J. Comput. Graph. Stat. 8(1), 69–92 (1999).
El-Saeed, A. R. & Almetwally, E. M. On Algorithms and Approximations for Progressively Type-I Censoring Schemes. Stat. Anal. Data Min. ASA Data Sci. J. 17(6), e11717 (2024).
Kruschke, J. Doing Bayesian data analysis: A tutorial with R, JAGS, and Stan, (2014).
Ng, H., Kundu, D. & Chan, P. S. Statistical analysis of exponential lifetimes under an adaptive Type-II progressive censoring scheme. Naval Res. Log. (NRL) 56(8), 687–698 (2009).
Henningsen, A. & Toomet, O. maxLik: A package for maximum likelihood estimation in R. Comput. Stat. 26, 443–458 (2011).
Rong Hua, W. & FEI, H. L. Statistical inference of Weibull distribution for tampered failure rate model in progressive stress accelerated life testing. J. Syst. Sci. Complex. 14(2), 237–243 (2004).
Hashem, A. F. & Abdel-Hamid, A. H. Statistical Prediction Based on Ordered Ranked Set Sampling Using Type-II Censored Data from the Rayleigh Distribution under Progressive-Stress Accelerated Life Tests. J. Math. 2023(1), 5211682 (2023).
Acknowledgements
Ongoing Research Funding Program, (ORF-2025-969), King Saud University, Riyadh, Saudi Arabia.
Author information
Authors and Affiliations
Contributions
Conceptualization, M.M.Y., A.S.H, E.M.A., M.N. and A.H.M.; methodology, M.M.Y., A.S.H. and M.N.; software, A.H.M., E.M.A. and M.M.Y.; validation, M.M.Y., A.H.M. and A.S.H.; formal analysis, E.M.A., M.N. and M.M.Y.; investigation, A.S.H., M.M.Y. and E.M.A.; resources, A.H.M., A.S.H. and M.N.; data curation, E.M.A. and M.M.Y.; writing–original draft preparation, M.M.Y., A.S.H. and E.M.A.; writing–review and editing, M.M.Y., A.S.H., M.N. and A.H.M. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Yousef, M.M., Hassan, A.S., Almetwally, E.M. et al. Utilizing unified progressive hybrid censored data in parametric inference under accelerated life tests. Sci Rep 15, 32567 (2025). https://doi.org/10.1038/s41598-025-06927-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-025-06927-5