Masked Causes of Failure in Series Systems:
A Likelihood Framework
Abstract
We develop a general likelihood framework for estimating component reliability from series system data when the component cause of failure is masked. The framework applies to any parametric specification of component hazard functions, including covariate-dependent hazards. Three sufficient conditions on the masking mechanism—that the candidate set contains the true cause, that masking probabilities are symmetric across candidates, and that masking probabilities are independent of the component lifetime parameters—allow the unknown masking distribution to be eliminated from the likelihood. We present the resulting likelihood contributions for exact failures with masked cause, right-censored, left-censored, and interval-censored observations. The framework serves as a foundation for distribution-specific inference, and we provide a summary of instantiations for five common lifetime distribution families.
1 Introduction
Estimating the reliability of individual components within a series system is a fundamental problem in reliability engineering (Agustin, 2011). A series system fails when any one of its components fails, so the system lifetime is determined by the weakest component. In many practical settings, only the system-level failure time is observable—the specific component that caused the failure may be unknown or only partially identified. This masking of the failure cause arises naturally in industrial diagnostics, field warranty data, and accelerated life testing, where post-failure inspection is infeasible, costly, or imprecise.
A common diagnostic outcome is a candidate set: a subset of components that plausibly contains the failed component. When the candidate set is a proper subset of all components but not a singleton, the failure cause is partially masked. When the candidate set is the full component set, the cause is fully masked. When it is a singleton, the cause is exactly identified.
The purpose of this paper is to provide a self-contained reference for the likelihood framework for masked failure data in series systems. We present the likelihood under three sufficient conditions (C1–C2–C3) on the masking mechanism that allow the unknown distribution of candidate sets to be eliminated from the likelihood function. The resulting likelihood is expressed entirely in terms of component reliability and hazard functions, enabling maximum likelihood estimation for any parametric specification of component hazard functions.
The framework is deliberately general: we derive the general likelihood structure in terms of component hazard functions, without specializing to any particular distributional form. This work grew out of an earlier master’s project (Towell, 2023b) that developed the likelihood model for Weibull series systems with simulation studies; the present paper extracts and generalizes the core likelihood framework. Distribution-specific treatments—including derivations of score equations, Fisher information, and simulation studies—are deferred to companion papers that cite the present work. In Section 7, we provide hazard function specifications for five common families (Exponential, Weibull, Pareto, Log-normal, and Gamma), enabling practitioners to apply the framework directly.
1.1 Related Work
The C1–C2–C3 conditions and the basic masked-data likelihood have a substantial history in the reliability literature. Miyakawa (1984) introduced the conditions for analyzing incomplete competing risks data. Usher and Hodgson (1988) formulated the MLE problem for masked series system data under these conditions, and Usher et al. (1993) derived exact maximum likelihood estimates for exponential components. Guess et al. (1991) established that the conditions hold in many practical diagnostic scenarios and developed component reliability estimation under partial failure information. Sarhan (2001) extended reliability estimation to broader settings with masked system life data. Most of this prior work specialized to the Exponential or Weibull families.
The present paper does not claim the C1–C2–C3 conditions or the basic likelihood derivation as novel contributions. Rather, its role is to serve as a foundational reference that provides a unified treatment of the framework in a single self-contained document. The value lies in the unified presentation, the extension of the likelihood to all four censoring types (exact, right, left, and interval), the identifiability analysis, and the five-family instantiation table (Section 7). Companion papers and software packages can cite this work for the general theory while focusing on distribution-specific derivations and simulation studies.
Non-parametric competing risks methods—including the Kaplan–Meier estimator, the Nelson–Aalen cumulative hazard estimator, and the Cox proportional hazards partial likelihood—require observing which event type occurred in order to separate cause-specific hazard contributions. Masking eliminates precisely this information: the candidate set does not identify the cause, so partial likelihood and non-parametric estimators cannot be applied directly. The parametric structure of is what makes the problem tractable under masking—it provides enough structure for the likelihood to separate component hazards even when the data cannot. Our framework is therefore parametric by necessity (for identifiability under masking), not merely by convenience.
The remainder of this paper is organized as follows. Section 2 establishes the series system model and derives the system reliability, density, and hazard functions. Section 3 derives the distribution of the component cause of failure. Section 4 defines the observational model, including the masked data notation and a taxonomy of observation types. Section 5 presents the three conditions and derives the core likelihood contribution—the central result of this paper. Section 6 discusses maximum likelihood estimation in the general framework. Section 7 provides hazard function specifications for common parametric families. Sections 8 and 9 discuss extensions, relaxations, and concluding remarks. Table 1 summarizes the principal notation used throughout the paper.
| Symbol | Description |
|---|---|
| Number of components in the series system | |
| System lifetime (random variable) for the th system | |
| Lifetime of component in system | |
| Component cause of system failure | |
| Hazard function of component | |
| Reliability function of component | |
| Density function of component | |
| Full parameter vector | |
| Observation tuple for system | |
| Observed time (, , or interval ) | |
| Observation type label | |
| Candidate set (possible failure causes) | |
| Masking probability for observation | |
| Index sets by observation type | |
| Log-likelihood function |
2 Series System Model
Consider a system composed of components arranged in a series configuration. Each component and system has two possible states: functioning or failed. We assume throughout that all component lifetime distributions are absolutely continuous with respect to Lebesgue measure (i.e., each has a density). We observe independent systems (which need not be identically distributed, since systems may differ by covariates). The lifetime of the th system is denoted by the random variable and the lifetime of its th component by . We assume that the component lifetimes within a single system are statistically independent but not necessarily identically distributed.
Remark 1 (Component independence).
Component independence is a standard assumption in the series system and competing risks literature, but it rules out common-cause failures, load-sharing, and environmental coupling. The entire framework developed below—system reliability as a product of component reliabilities (Theorem 1), system hazard as a sum of component hazards (Theorem 3), and the cause-of-failure distribution (Theorem 6)—depends critically on this assumption. Extensions to dependent competing risks exist but require copula models or other dependence structures and face fundamental identifiability challenges: Tsiatis (1975) showed that, without independence, marginal component distributions are not identifiable from system lifetime data alone. Such extensions are beyond the scope of this paper.
A series system fails when any component fails, so the system lifetime is
| (1) |
The reliability function of the th system is , the probability that the system survives beyond time . The probability density function (pdf) of is , and the hazard function is
| (2) |
representing the instantaneous failure rate at time given survival to time .
Each component’s lifetime distribution is specified by its hazard function , where is a finite-dimensional parameter vector and is an optional covariate vector for the th system. The cumulative hazard, reliability, and density follow:
| (3) | ||||
| (4) | ||||
| (5) |
This specification subsumes standard parametric families and accommodates non-standard hazard shapes (bathtub curves, piecewise-constant rates) and covariate-dependent hazards such as proportional hazards models. When covariates are absent, we suppress and write .
The overall parameter vector is
| (6) |
belonging to a parameter space .
Theorem 1 (System reliability).
The series system has a reliability function given by
| (7) |
Proof.
Since the system fails when any component fails, . By the independence of component lifetimes,
Theorem 2 (System pdf).
The series system has a pdf given by
| (8) |
Proof.
Differentiating the system reliability function,
By the product rule applied recursively,
Theorem 3 (System hazard).
The series system has a hazard function given by
| (9) |
Remark 2.
Combining the hazard and reliability representations, the system pdf admits the convenient form
| (10) |
which we use extensively in subsequent derivations.
3 Component Cause of Failure
Whenever a series system fails, precisely one component is the cause (almost surely, since the component lifetime distributions are assumed to be absolutely continuous with respect to Lebesgue measure). We denote the component cause of failure of the th system by the discrete random variable , with support . The event means that component had the shortest lifetime among all components of the th system.
Theorem 4 (Joint distribution of and ).
The joint pdf of the component cause of failure and the system lifetime is
| (11) |
Proof.
The event requires that component fails at time while all other components survive past time . By the independence of component lifetimes,
Since , substituting gives
Theorem 5 (Marginal probability of cause).
The probability that the th component is the cause of failure is
| (12) |
Proof.
Marginalizing the joint pdf over the system lifetime,
Since , we can rewrite this as
Theorem 6 (Conditional probability of cause given failure time).
Given that the system fails at time , the probability that the th component is the cause is
| (13) |
Proof.
By the definition of conditional probability,
Theorem 6 is a conditional version of Theorem 5; it is also the operationally important result, since the likelihood (Section 5) depends on the conditional cause probability given the observed failure time rather than on the marginal cause probability of Theorem 5. The conditional probability of cause is determined entirely by the ratio of the component’s hazard to the total system hazard at the observed failure time. In a well-designed series system, the designer aims to balance failure rates so that no single component dominates, resulting in roughly balanced cause-of-failure probabilities across components, though this ideal is not always achieved in practice.
4 Observational Model
In practice, we do not observe the component lifetimes directly. Instead, we observe system-level data that may be subject to two forms of masking: censoring of the failure time and masking of the failure cause.
4.1 Observation Types
We consider four types of observations that may arise in reliability studies of series systems (see Klein and Moeschberger, 2005, for background on censored data in survival analysis):
-
1.
Exact failure with candidate set. The system is observed to fail at time , and a candidate set is observed that is indicative of the component cause of failure. When , the cause is exactly identified; when , the cause is masked.
-
2.
Right-censored. The system is still functioning at observation time ; we know only that .
-
3.
Left-censored. The system has already failed before observation time ; we know only that . A candidate set may also be observed if a diagnostic is performed at inspection time to partially identify the failure cause.
-
4.
Interval-censored. The system failure occurred in the interval ; we know only that . A candidate set may accompany the observation when a diagnostic is performed at one of the inspection times.
4.2 Masked Data Notation
Each observation is represented as a tuple , where:
-
•
encodes the observed time information: a failure time , a censoring time , or an interval ;
-
•
is a label indicating the observation type: exact failure (), right-censored (), left-censored (), or interval-censored ();
-
•
is the candidate set, relevant whenever a failure is known to have occurred (exact, left-censored, or interval-censored observations). When no cause information is available, . For right-censored observations (no failure observed), (no component has failed); and
-
•
is the covariate vector for the th system (possibly empty when no covariates are recorded).
The complete data set is , assumed to be independent draws from the observational model.
4.3 Likelihood Contributions by Observation Type
Each observation type contributes differently to the likelihood function . The likelihood contributions depend on three conditions developed in Section 5; Table 2 previews the final results for reference.
| Observation type | Likelihood contribution |
|---|---|
| Exact failure + candidate set | |
| Right-censored | |
| Left-censored + candidate set | |
| Interval-censored + candidate set |
When no cause information is available (), the left-censored and interval-censored contributions reduce to the familiar forms and , respectively, since by Equation (10).
4.4 Dependency Structure
Figure 1 depicts the dependency structure of the data generating process. Observed quantities are shown as shaded nodes; latent quantities (component lifetimes , the system lifetime , and the component cause of failure ) are shown as open nodes. Solid arrows denote structural dependencies; dashed arrows denote influences that the C1–C2–C3 conditions allow us to ignore in the likelihood. The covariate vector influences both the component lifetimes (through the hazard functions ) and potentially the candidate set (e.g., when the operating environment affects diagnostic quality). The conditions in Section 5 allow us to construct a likelihood that does not require modeling the distribution of .
4.5 Example: Masked Data
Table 3 shows an example of masked data for a 3-component series system with a right-censoring time . Systems 5 and 6 are right-censored (their failures were not observed before time ). System 2 has a singleton candidate set, so its cause of failure is exactly identified. The remaining failed systems have candidate sets of size 2, representing partial masking.
| System | |||
|---|---|---|---|
| 1 | 1.1 | ||
| 2 | 1.3 | ||
| 3 | 2.6 | ||
| 4 | 3.7 | ||
| 5 | 5.0 | ||
| 6 | 5.0 |
5 The C1–C2–C3 Likelihood
We now derive the likelihood contribution for an observation where the system failure time is known but the component cause of failure is masked by a candidate set. The derivation below recapitulates and extends the classical argument (Miyakawa, 1984; Usher and Hodgson, 1988) in a unified, general form.
5.1 Joint Distribution of , , and
Our goal is to estimate from observed data . When the system failure is observed (), we observe the system failure time and a candidate set . The joint pdf of the system lifetime and the candidate set is
| (14) |
While we assume the system lifetime pdf is known (up to parameters), the conditional distribution is generally unknown—it depends on the diagnostic procedure, which we do not model.
Since and are statistically dependent, we can introduce into the analysis. By Theorem 4, the joint pdf of and is . The joint pdf of , , and is therefore
| (15) |
Marginalizing over ,
| (16) |
The unknown conditional probability prevents direct use of this expression for likelihood-based inference. We now introduce three conditions that successively simplify Equation (16) until the unknown masking distribution drops out entirely.
5.2 Condition 1: Candidate Set Contains the True Cause
Condition 1 (C1).
The candidate set contains the index of the failed component:
| (17) |
Condition 1 is the minimal requirement for the candidate set to carry useful information about the failure cause: the true cause must not be excluded. In practice, real diagnostics work by narrowing down from the full component set—eliminating candidates that pass functional checks—rather than constructing the candidate set from scratch. Because exclusion of the true cause would require the diagnostic to affirmatively misidentify a functioning component as the sole failure site, C1 holds whenever the diagnostic is competent in this limited sense.
Two common diagnostic architectures illustrate the point. First, automotive on-board diagnostics (OBD) fault codes are generated by the failing module itself: a voltage exceedance or communication timeout triggers the code, so the candidate set inherently includes the true cause. Second, hierarchical troubleshooting trees prune branches based on pass/fail tests at each level; the true cause remains in the surviving subtree at every step unless a test yields a false negative, which is a calibration failure rather than a structural feature of the diagnostic.
In short, C1 asks for diagnostic competence—not actively wrong—rather than diagnostic precision—exactly right. Violating C1 is a pathological scenario (the diagnostic positively excludes the failed component) that would undermine any analysis, masked or otherwise.
What C1 buys.
Under C1, if then . The summation in Equation (16) therefore reduces from to :
| (18) |
What breaks without C1.
Without C1, the summation must range over all components, and the likelihood depends on the masking probabilities for components outside the candidate set. This means the likelihood cannot be simplified without modeling the full masking mechanism.
5.3 Condition 2: Symmetric Masking Within the Candidate Set
Condition 2 (C2).
Given an observed system failure time and candidate set , the masking probability is the same regardless of which component in is the true cause:
| (19) |
Condition 2 is a requirement on the masking mechanism, not merely on the observed data: it must hold for all candidate sets that the mechanism can produce, not just the particular sets realized in the sample. When (a singleton candidate set), C2 is satisfied vacuously—the condition is non-trivial only when . In words, C2 requires that the diagnostic does not discriminate between components within the candidate set: given that a particular candidate set is reported, no member of that set is favored over another. Symmetry arises naturally whenever masking is determined by structural grouping—subsystem, module, or physical region—rather than by component-specific properties. Components that share a group are indistinguishable to the diagnostic precisely because the diagnostic operates at the group level.
Two additional examples reinforce this pattern. In avionics maintenance, field technicians replace line-replaceable units (LRUs): every component inside the unit is equally suspect because the diagnostic identified the unit, not the component. In industrial settings, SCADA monitoring systems report alarms at the subsystem level—e.g., “pump station fault”—without distinguishing which element (motor, valve, seal) triggered the alarm; the candidate set is the group, and all members are symmetric.
We acknowledge that C2 is the condition most likely to be violated in practice, since partial diagnostic information can make one candidate more plausible than another. When asymmetry is suspected, a practical mitigation is to redefine the candidate set at the finest resolution where symmetry still holds, effectively trading a smaller candidate set for a valid application of C2.
What C2 buys.
Under C1 and C2, the masking probability is constant for all and can be factored out of the summation in Equation (18):
| (20) |
where is any element of .
What breaks without C2.
Without C2, the masking probabilities remain inside the summation, coupling the hazard contributions with component-specific masking weights. The MLE then depends on the unknown masking probabilities, which must be jointly estimated or modeled.
5.4 Condition 3: Masking Independent of
Condition 3 (C3).
The masking probabilities, conditioned on the failure time and the component cause of failure, do not depend on the system parameter :
| (21) |
is not a function of .
Condition 3 states that the diagnostic procedure’s behavior is determined by factors external to the component lifetime parameters. The masking probability may depend on the failure time (for exact failures, where the diagnostic is performed at or near the observed failure time), the diagnostician, the testing equipment, or other covariates—but not on . For censored observations, the diagnostic is performed at the inspection time rather than at the unknown failure time; see Remark 3 below. The justification is fundamentally causal: the diagnostic tool was designed and calibrated before any failures were observed, so its behavior cannot depend on the unknown we are estimating. For instance, OBD voltage thresholds are hard-coded at manufacture; a vibration sensor’s frequency band is set during installation. Neither adapts to the lifetime parameters of the components it monitors.
Condition 3 is closely related to the concept of ignorability in the missing-data framework of Little and Rubin (2002). In our setting, the “missingness” is the loss of exact cause information through masking; C3 ensures that the masking mechanism is ignorable for likelihood-based inference, in the sense that the conditional distribution of the candidate set need not be modeled when maximizing the likelihood over (see Little and Rubin, 2002, Ch. 6).
When covariates are present, C3 requires that the masking probabilities do not depend on given the covariates and failure time. If the same covariates influence both the failure rate and the diagnostic quality (e.g., an extreme operating environment that accelerates failures and also degrades sensor accuracy), the practitioner should verify that the masking mechanism remains ignorable after conditioning on .
What C3 buys.
Under C1, C2, and C3, the joint pdf becomes
| (22) |
When we view this as a function of (with the data fixed), is a constant multiplier that does not affect the location of the maximum. The likelihood contribution is therefore
| (23) |
What breaks without C3.
Without C3, the factor depends on and cannot be dropped. The practitioner would need to model the dependence of the masking mechanism on —a substantially harder problem that requires additional data or assumptions about the diagnostic process.
5.5 Real-World Example
To illustrate how the three conditions arise in practice, consider an electronic device with three components arranged in a series configuration. Components 1 and 2 are on a shared circuit board, while component 3 is separate. A diagnostic tool isolates the failure to either the shared circuit board or the individual component. A more detailed board-level inspection sometimes pinpoints the specific failed component; let denote the probability that this inspection succeeds. The conditional probabilities for candidate sets are:
This diagnostic tool satisfies all three conditions:
-
•
C1: The candidate set always contains the failed component—the tool correctly isolates failures to the correct subsystem (each possible candidate set contains the true cause).
-
•
C2: For the candidate set , the masking probability is the same whether component 1 or component 2 failed (both equal ). For singleton candidate sets or , Condition 2 is satisfied trivially.
-
•
C3: The masking probabilities depend only on the diagnostic tool (through ) and not on the component lifetime parameters .
The parameter is essential for identifiability: when the board-level inspection never succeeds (), components 1 and 2 always appear together in every candidate set and their individual parameters cannot be separated—see Section 5.8 for a detailed discussion. Any ensures that some observations produce singleton candidate sets or , providing the information needed to distinguish the two components.
According to Guess et al. (1991), many industrial diagnostic scenarios naturally satisfy these conditions, reinforcing the practical applicability of the framework.
5.6 Censored Observations with Candidate Sets
The C1–C2–C3 derivation in Sections 5.2–5.4 applies to any observation where a failure is known to have occurred. Equation (22) gives the joint density of the failure time and candidate set at a specific time . For left-censored and interval-censored observations, the failure time is not known exactly, so we integrate over the admissible range.
Remark 3 (Diagnostic timing and the masking probability).
For an exact failure, the diagnostic may be performed at or near the observed failure time , so is evaluated at a known point and is a scalar constant in the likelihood. For censored observations, the diagnostic is performed at the inspection time—when the system’s failed state is discovered—not at the unknown failure time . (For interval-censored observations, this is typically the later inspection time , at which the failure is first detected.) The masking probability is therefore , which does not depend on the integration variable . This is the natural model: inspectors diagnose the current state at inspection time and do not have access to the unknown failure time. Consequently, factors out of the integrals in the left-censored and interval-censored likelihood contributions below, and by Condition 3 it does not depend on , so it may be dropped.
Theorem 7 (Left-censored likelihood contribution under C1–C2–C3).
Proof.
By Equation (22), the joint density of the failure time and candidate set at time is . Since the failure time is known only to satisfy , we integrate over . As noted in Remark 3, the diagnostic is performed at inspection time, so does not depend on the integration variable and factors out of the integral. By Condition 3, does not depend on either, so it is a constant factor that may be dropped. ∎
Theorem 8 (Interval-censored likelihood contribution under C1–C2–C3).
Proof.
Remark 4.
When no cause information is available (), these contributions reduce to the standard censored-data forms. For the left-censored case,
by Equation (10). Similarly, the interval-censored contribution reduces to
5.7 Combined Likelihood
We now assemble the full likelihood from the individual contributions (Equation (23), Theorems 7–8, and the right-censored case from Theorem 1). Let , , , and denote the index sets of observations that are exact failures, right-censored, left-censored, and interval-censored, respectively.
Theorem 9 (Likelihood under C1–C2–C3).
Proof.
Each factor follows from the corresponding individual result: the exact-failure contribution from Equation (23), the right-censored contribution from Theorem 1, the left-censored contribution from Theorem 7, and the interval-censored contribution from Theorem 8. Independence across systems gives the product. ∎
5.8 Identifiability
Before addressing identifiability within our framework, we note the classical competing risks result of Tsiatis (1975): without the component independence assumption (Remark 1), the marginal component lifetime distributions are not identifiable from system lifetime data alone—even with complete observation of the failure cause. Our framework inherits identifiability from the independence assumption, which allows the system reliability to factor as a product of component reliabilities (Theorem 1). Given independence, the identifiability question reduces to whether the observed candidate sets provide enough information to separate the component parameters.
Definition 1 (Candidate-set separability).
Let denote the collection of candidate sets that occur with positive probability under the masking mechanism. We say the masking mechanism is separating if, for every pair of distinct components , there exists a candidate set such that and . When this condition fails for some pair —that is, for every —we say and are diagnostically confounded.
The following theorem shows that separability is the key condition governing identifiability under C1–C2–C3.
Theorem 10 (Identifiability under C1–C2–C3).
Let be the hazard function for component , where is the full parameter vector, and suppose each component family is individually identifiable: for all implies . Under C1–C2–C3:
-
(a)
Necessary condition. If components and are diagnostically confounded (Definition 1), then and are not separately identifiable from the observed-data likelihood. Specifically, any reparametrization preserving for all yields the same likelihood value.
-
(b)
Sufficient condition. If the masking mechanism is separating (Definition 1), and at least some observations are exact failures (type ), then the parameter vector is identifiable for any parametric family whose hazard functions are linearly independent over .
Proof.
Part (a). Suppose for every . Then in every exact-failure likelihood contribution (Equation 23), the hazards and appear as a sum wherever they appear at all. The survival factor likewise depends on only through their sum. The same holds for right-, left-, and interval-censored contributions (Table 2), since all involve and . Hence the likelihood depends on and only through the sum , and any reallocation preserving this sum leaves unchanged.
Part (b). Suppose the masking mechanism is separating and let with for all data sets. Then in particular, for a data set consisting of a single exact failure at time with candidate set , the log-likelihood equality gives
| (27) |
for all and all , where and . Differentiating (27) with respect to for two candidate sets with , (which exist by separability) and subtracting yields a relation that isolates the contribution of component . Specifically, from the survival terms we obtain , and from candidate sets separating we obtain that for all whenever are linearly independent. By the individual identifiability assumption, for each . ∎
Part (a) states that diagnostically confounded components are fundamentally unresolvable: the likelihood surface has a ridge along all reparametrizations of the confounded pair, producing an infinite family of equivalent maximizers rather than a unique MLE. Part (b) provides a checkable condition: the analyst needs only inspect the candidate-set structure to verify separability.
Remark 5 (Linear independence of hazard functions).
The linear independence condition in Theorem 10(b) holds generically for all families in Table 4. The Weibull hazard with distinct shapes yields linearly independent power functions. For the exponential family ( for all ), the hazards are constant and hence linearly dependent; in this case, separability of the candidate sets carries the full burden of identifiability. For families with identical functional form across components (e.g., exponential or homogeneous Weibull), identifiability requires that the candidate-set matrix —whose rows are the indicator vectors of the candidate sets in —has full column rank . This is strictly stronger than separability (which requires only that no two columns of are identical) but is guaranteed when includes a singleton for each component .
Remediation of non-identifiability.
When components and are diagnostically confounded, three strategies are available:
-
1.
Collapse into a super-component. Replace the always-grouped components with a single composite component whose hazard is . This reduces the parameter count and restores identifiability at the cost of losing individual component resolution.
-
2.
Impose equality constraints. Assume the grouped components share identical parameter values (), splitting the combined hazard equally. This is appropriate when the components are physically interchangeable (e.g., identical capacitors on the same circuit board).
-
3.
Bayesian regularization. Place informative priors on and to obtain a proper posterior even when the likelihood surface is flat. The posterior concentrates as additional diagnostic information becomes available, so the prior penalty vanishes asymptotically.
In each case, the practitioner should examine the candidate-set structure of the data before fitting the model. Convergence difficulties in numerical optimization (e.g., failure to converge within a maximum number of iterations) may signal that the separability condition of Definition 1 is violated or nearly violated.
6 Maximum Likelihood Estimation
Maximum likelihood estimation (MLE) finds the parameter values that maximize the likelihood of the observed data (Bain and Engelhardt, 1992; Casella and Berger, 2002). A maximum likelihood estimate satisfies
| (28) |
For computational efficiency, we work with the log-likelihood.
Theorem 11 (Log-likelihood).
Proof.
Taking logarithms in Theorem 9 and using gives the result directly. ∎
Remark 6 (Common special case).
6.1 Score Equations
The MLE is found by solving the score equations
| (34) |
for each parameter (the th element of the th component’s parameter vector). In general, these equations do not admit closed-form solutions and must be solved numerically using methods such as Newton–Raphson or quasi-Newton algorithms (Nocedal and Wright, 2006; Byrd et al., 1995).
6.2 Asymptotic Properties
Under standard regularity conditions (which must be verified for each specific distribution family)—including identifiability, smoothness of the log-likelihood, and the true parameter lying in the interior of the parameter space—the MLE is consistent, asymptotically normal, and asymptotically efficient (Casella and Berger, 2002; Lehmann and Casella, 1998). That is, as , (the true parameter) and , where is the per-observation Fisher information matrix. For finite samples, these asymptotic approximations may be inaccurate, and bootstrap methods (e.g., the bias-corrected and accelerated method (Efron, 1987; Efron and Tibshirani, 1994)) provide a nonparametric alternative for constructing confidence intervals.
6.3 General Recipe for Practitioners
Given a hazard function for each component, the practitioner applies the framework as follows:
-
1.
Specify the component hazard functions and compute (analytically if possible, numerically otherwise).
-
2.
Substitute the component-specific and into the log-likelihood (Equation (29)).
-
3.
Differentiate the log-likelihood with respect to each parameter to obtain the score equations.
-
4.
Solve the score equations numerically (e.g., using L-BFGS-B (Byrd et al., 1995) or Newton–Raphson) to obtain .
-
5.
Construct confidence intervals using the observed Fisher information or bootstrap resampling.
Section 7 provides hazard functions for five common parametric families, enabling immediate application of this recipe.
6.4 Worked Example: Exponential Components
We illustrate the recipe using the data in Table 3 ( components, observations) with Exponential component lifetimes. For Exponential components, and , where is the failure rate of the th component.
Since the data contain only exact failures and right-censored observations, the log-likelihood is (Remark 6). For Exponential components, and , so the two contributions are:
| (35) | ||||
| (36) |
Substituting the data from Table 3 (with ), the combined log-likelihood is
| (37) |
The score equations are
| (38) | ||||
| (39) | ||||
| (40) |
Subtracting (40) from (38) gives , so . Setting and , the system reduces to two equations whose solution is
| (41) |
The total system failure rate is , which equals the number of observed failures divided by the total exposure time—a general property of the Exponential MLE. The masking affects only the allocation of hazard across components, not the total.
Component 2 has the highest estimated failure rate, consistent with its appearance in three of the four failure candidate sets (including a singleton). Components 1 and 3 have equal rates by symmetry: each appears in exactly two candidate sets, with identical structure after relabeling.
7 Common Hazard Function Specifications
Table 4 lists hazard functions for five standard parametric families, illustrating how the general framework specializes. Any hazard function satisfying the regularity conditions of Section 6.2 can be used; the named families below are common starting points.
| Family | Parameters | |||
|---|---|---|---|---|
| Exponential | ||||
| Weibull | ||||
| Pareto | ; | |||
| Log-normal | , | |||
| Gamma111The Gamma hazard function has no elementary closed form; it is expressed here via its definition . |
In Table 4, and denote the standard normal CDF and pdf, respectively; is the lower incomplete gamma function; and is the gamma function. For distributions with parameter-dependent support (e.g., the Pareto distribution with ), the system support is , and the product formula for system reliability (Theorem 1) must be applied only on this common support.
Remark 7 (Covariate-dependent hazards).
The hazard function may incorporate observation-level covariates . A common special case is the proportional hazards specification , where is a baseline hazard from any family in Table 4 and is a component-specific regression coefficient vector. The likelihood expressions of Section 5 remain valid; only the functional form of changes.
Remark 8 (Nested models within a family).
Several families in Table 4 admit natural hierarchies of nested sub-models. For instance, the Weibull family contains a common-shape reduction ( parameters) in which the system lifetime is itself Weibull, and a further exponential specialization ( parameters) with fully analytical inference. These nestings enable formal model selection via likelihood ratio tests, AIC, or BIC. A detailed treatment—including simulation studies quantifying when a reduced model is appropriate—is given in Towell (2025f).
8 Discussion
8.1 What the Framework Enables
The hazard-function-based likelihood framework developed in this paper provides a foundation for a family of distribution-specific companion papers. Each companion paper can focus on a specific parametric family—for example, Weibull (Towell, 2025e) or Exponential (Towell, 2025d)—deriving closed-form score equations and Fisher information matrices, conducting simulation studies to assess MLE performance under varying masking and censoring scenarios, and developing specialized software packages, all while citing the present work for the general theory. The layered software stack described in Section 8.4—from component specification (dfr.dist) through series composition (dfr.dist.series) to masked-data likelihood (dfr.lik.series.md)—demonstrates this modularity: new component distributions inherit the full estimation pipeline without distribution-specific likelihood code.
8.2 Relaxation of Conditions
The three conditions (C1–C2–C3) provide a clean mathematical framework, but practitioners may encounter situations where one or more conditions are violated. We briefly sketch what happens in each case:
-
•
Relaxing C1 (candidate set may not contain the true cause): The summation in the likelihood cannot be restricted to ; the full component set must be considered, along with a model for the probability that the true cause is excluded from the candidate set. This introduces additional nuisance parameters.
-
•
Relaxing C2 (asymmetric masking): The masking probabilities remain inside the summation and couple with the component hazards. The MLE depends on the relative masking probabilities, which must be estimated or modeled.
-
•
Relaxing C3 (masking depends on ): The factor cannot be dropped from the likelihood, and the MLE must account for the dependence of the masking mechanism on the lifetime parameters. This leads to a more complex joint estimation problem.
Detailed treatment of these relaxations is beyond the scope of this paper and is deferred to future work.
8.3 Computational Considerations
Several practical issues arise when applying the framework:
-
•
Local optima. The log-likelihood surface may be multimodal, particularly under heavy masking or censoring. Multiple random starting points for the numerical optimizer are recommended.
-
•
Convergence. Failure to converge within a reasonable number of iterations may indicate identifiability issues for the given data set. Such cases should be flagged rather than silently discarded.
-
•
Boundary constraints. Many lifetime distributions have positivity constraints on parameters. Constrained optimization methods such as L-BFGS-B (Byrd et al., 1995) or reparameterization (e.g., optimizing over log-scale parameters) can enforce these constraints.
8.4 Software Ecosystem
The framework developed in this paper is implemented as a layered software stack in the R statistical computing environment (Towell, 2025a, b, c). At the base, the dfr.dist package (Towell, 2025a) provides a hazard-function-first abstraction for individual component lifetime distributions: each distribution is specified by its hazard function, from which the cumulative hazard, reliability, density, and random sampling follow (cf. Equations (3)–(5)). The dfr.dist.series package (Towell, 2025b) composes components into series system distributions using hazard additivity (Theorem 3), with a parameter layout that maps the global vector to component-specific subvectors . The dfr.lik.series.md package (Towell, 2025c) implements the C1–C2–C3 log-likelihood (Theorem 11) for all four observation types (exact, right-censored, left-censored, and interval-censored), taking arbitrary component hazard closures as input. Together these packages enable practitioners to apply the recipe of Section 6.3 with any component distribution in the dfr.dist ecosystem, without writing distribution-specific likelihood code.
9 Conclusion
We have developed a general likelihood framework, expressed in terms of component hazard functions, for estimating component reliability from masked series system data. The framework rests on three conditions (C1–C2–C3) that capture natural properties of diagnostic procedures and allow the unknown masking distribution to be eliminated from the likelihood.
The key results are:
-
•
The joint distribution of the system lifetime, component cause of failure, and candidate set (Section 5.1);
- •
-
•
The combined likelihood contribution under C1–C2–C3 (Theorem 9);
-
•
The general log-likelihood and a recipe for applying the framework to any parametric hazard specification (Section 6);
-
•
Hazard function specifications for five common families (Table 4).
This framework provides a rigorous foundation for distribution-specific companion papers that can focus on deriving score equations, conducting simulation studies, and developing specialized inference tools for particular lifetime distribution families.
References
- Systems in series. In Wiley Encyclopedia of Operations Research and Management Science, External Links: ISBN 9780470400531, Document, Link Cited by: §1.
- Introduction to probability and mathematical statistics. second edition, Duxbury Press. External Links: ISBN 9780534380205 Cited by: §6.
- A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing 16 (5), pp. 1190–1208. Cited by: item 4, §6.1, 3rd item.
- Statistical inference. Duxbury Advanced Series. Cited by: §6.2, §6.
- Better bootstrap confidence intervals. Journal of the American Statistical Association 82 (397), pp. 171–185. Cited by: §6.2.
- An introduction to the bootstrap. CRC press. Cited by: §6.2.
- Estimating system and component reliabilities under partial information on cause of failure. Journal of Statistical Planning and Inference 29, pp. 75–85. External Links: Document Cited by: §1.1, §5.5.
- Survival analysis: techniques for censored and truncated data. Springer Science & Business Media. Cited by: §4.1.
- Theory of point estimation. Springer Science & Business Media. Cited by: §6.2.
- Statistical analysis with missing data. second edition, Wiley-Interscience, Hoboken, NJ. Cited by: §5.4.
- Analysis of incomplete data in competing risks model. IEEE Transactions on Reliability R-33 (4), pp. 293–296. Cited by: §1.1, §5.
- Numerical optimization. Springer. Cited by: §6.1.
- Reliability estimations of components from masked system life data. Reliability Engineering & System Safety 74 (1), pp. 107–113. External Links: Document Cited by: §1.1.
- Algebraic maximum likelihood estimators. Note: R package version 0.9.0. Available: https://github.com/queelius/algebraic.mle/ External Links: Link Cited by: §8.4.
- Reliability estimation in series systems: maximum likelihood techniques for right-censored and masked failure data. Master’s project, Southern Illinois University Edwardsville. External Links: Document, Link Cited by: §1.
- Wei.series.md.c1.c2.c3: estimating reliability of weibull components in series from masked data. Note: R package version 0.9.0 External Links: Document, Link Cited by: §8.4.
- Dfr.dist: distribution-free reliability distributions. Note: R package. Available: https://github.com/queelius/dfr.dist External Links: Link Cited by: §8.4.
- Dfr.dist.series: series system distributions from dynamic failure rate components. Note: R package version 0.1.0 External Links: Link Cited by: §8.4.
- Dfr.lik.series.md: masked-cause likelihood models for series systems. Note: R package version 0.1.0 External Links: Link Cited by: §8.4.
- Exponential series systems with masked failure causes: closed-form MLE and simulation studies. Note: In preparation Cited by: §8.1.
- Maximum likelihood estimation for weibull series systems with masked failure causes. Note: In preparation Cited by: §8.1.
- Model selection for reliability estimation in series systems with masked failure causes. Note: In preparation Cited by: Remark 8.
- A nonidentifiability aspect of the problem of competing risks. Proceedings of the National Academy of Sciences 72 (1), pp. 20–22. Cited by: §5.8, Remark 1.
- Maximum likelihood analysis of component reliability using masked system life-test data. IEEE Transactions on Reliability 37 (5), pp. 550–555. External Links: Document Cited by: §1.1, §5.
- Exact maximum likelihood estimation using masked system data. IEEE Transactions on Reliability 42 (4), pp. 631–635. External Links: Document Cited by: §1.1.