Relaxed Candidate Set Models for Masked Data in Series Systems
Abstract
We develop likelihood-based inference methods for series systems with masked failure data when the traditional conditions governing candidate set formation are relaxed. While existing methods require that masking be non-informative (Condition C2) and parameter-independent (Condition C3), practical diagnostic systems often violate these assumptions. We derive the likelihood under various relaxation scenarios, establish identifiability conditions, and compare Fisher information under standard versus relaxed models. Our analysis reveals that informative masking, when properly modeled, can paradoxically improve estimation efficiency by providing additional information about the failed component. We implement these methods in an R package and provide simulation studies demonstrating the bias that arises from incorrectly assuming standard conditions when masking is informative.
Keywords: Series systems, masked data, reliability estimation, informative masking, candidate sets, Fisher information, maximum likelihood estimation
1 Introduction
Estimating the reliability of individual components in a series system presents a fundamental challenge in reliability engineering: system-level failure data is observable, but component-level failure causes are often masked. When a series system fails, diagnostic procedures may identify only a candidate set of components that could have caused the failure, rather than pinpointing the exact failed component [7]. This masking, combined with right-censoring of system lifetimes, complicates statistical inference about component reliability parameters.
Prior work on masked data in series systems has established a tractable likelihood framework under three conditions [6]:
-
β’
C1: The failed component is always contained in the candidate set.
-
β’
C2: Masking is non-informativeβconditional on the failure time and candidate set, each component in the candidate set is equally likely to have failed.
-
β’
C3: The masking mechanism does not depend on the system parameters .
Under these conditions, the likelihood has a simple closed form that enables efficient maximum likelihood estimation (MLE). However, practical diagnostic systems may violate C2 and C3:
-
β’
Experienced technicians may preferentially include components that βseem likelyβ to have failed based on failure time characteristics, violating C2.
-
β’
Diagnostic algorithms based on reliability rankings may systematically favor certain components, with the ranking depending on the true parameters, violating C3.
This paper develops the theoretical and computational framework for likelihood-based inference when C2 and/or C3 are relaxed while maintaining C1. Our contributions are:
-
1.
General likelihood framework. We derive the likelihood under C1 alone, showing how informative and parameter-dependent masking modify the standard likelihood structure (SectionΒ 3).
-
2.
Practical masking models. We introduce the rank-based informative masking model and the Bernoulli candidate set model with KL-divergence constraints, which provide interpretable parameterizations of non-standard masking (SectionΒ 3).
-
3.
Identifiability analysis. We establish conditions under which parameters remain identifiable when standard conditions fail, including the surprising result that informative masking can improve identifiability in certain cases (SectionΒ 4).
-
4.
Efficiency comparison. We derive the Fisher information matrix under relaxed conditions for exponential series systems, enabling precise comparison of estimation efficiency (SectionΒ 4).
-
5.
Simulation studies. We quantify the bias from incorrectly assuming C2 when masking is informative, and demonstrate the improved estimation achievable when the masking structure is properly modeled (SectionΒ 5).
The remainder of this paper is organized as follows. SectionΒ 2 reviews series systems, masked data, and the standard C1-C2-C3 likelihood. SectionΒ 3 develops the likelihood under relaxed conditions. SectionΒ 4 analyzes identifiability and Fisher information. SectionΒ 5 presents simulation studies. SectionΒ 6 discusses practical implications, and SectionΒ 7 concludes.
2 Background
2.1 Series System Model
Consider a series system composed of components. The lifetime of the -th system is
| (1) |
where denotes the lifetime of the -th component in the -th system. Component lifetimes are assumed independent with parametric distributions indexed by ; the full parameter vector is .
Definition 2.1 (Component Distribution Functions).
For component with parameter :
| (2) | |||||
| (3) | |||||
| (4) |
For the series system, these functions combine as follows:
Theorem 2.2 (Series System Distribution Functions).
The series system has:
| (5) | ||||
| (6) | ||||
| (7) |
The proof follows from the independence of component lifetimes and standard arguments [6].
2.2 Component Cause of Failure
Let denote the index of the component that caused system to fail. Since the system fails when the first component fails, .
Theorem 2.3 (Joint Distribution of ).
The joint distribution of system lifetime and component cause of failure is:
| (8) |
Corollary 2.4 (Conditional Failure Probability).
Given that the system failed at time , the probability that component caused the failure is:
| (9) |
This probability plays a central role in masked data analysis, as it represents the βtrueβ probability that each component failed, which the masking mechanism partially obscures.
2.3 Masked Data Structure
Definition 2.5 (Observed Data).
For each system , we observe:
-
β’
: Right-censored system lifetime,
-
β’
: Event indicator ( if failure observed, if censored),
-
β’
: Candidate set (only observed when ).
The latent (unobserved) variables are:
-
β’
: Index of failed component,
-
β’
: Component failure times.
2.4 Traditional Conditions C1, C2, C3
The existing literature [7, 8, 4] establishes tractable likelihood-based inference under three conditions:
Condition 1 (C1: Failed Component in Candidate Set).
The candidate set always contains the failed component:
| (10) |
Condition 2 (C2: Non-Informative Masking).
Given the failure time and that the failed component is in a candidate set , the probability of observing does not depend on which component in failed:
| (11) |
for all .
Condition 3 (C3: Parameter-Independent Masking).
The masking probabilities do not depend on the system parameters:
| (12) |
where does not depend on .
2.5 Likelihood Under C1, C2, C3
Under all three conditions, the likelihood admits a tractable form:
Theorem 2.6 (Likelihood Under C1-C2-C3).
Under Conditions C1, C2, and C3, the likelihood contribution from an uncensored observation is proportional to:
| (13) |
For a censored observation with lifetime :
| (14) |
Proof.
By the chain rule:
| (15) |
Under C1, terms with vanish. Under C2, the masking probability is constant over , so:
| (16) |
Under C3, the masking probability does not depend on , yielding the proportionality in (13). β
The complete log-likelihood for independent systems is:
| (17) |
2.6 Related Work
The masked data problem in series systems was introduced by Usher and Hodgson [7], who developed MLE methods for exponential components. Usher et al. [8] extended this to Weibull components with exact maximum likelihood. Guo et al. [4] (Guo et al.) provided simulation studies validating the approach under various masking scenarios.
The informative censoring literature in survival analysis [5, 3] addresses related issues where the censoring mechanism depends on covariates or outcomes. However, the candidate set structure in masked data creates a distinct problem not fully addressed by standard informative censoring methods.
The competing risks framework [1] provides another perspective, viewing component failures as competing causes of system failure. However, standard competing risks methods assume the cause is observed, whereas masked data only provides partial information through candidate sets.
Our work extends the C1-C2-C3 framework by explicitly modeling departures from C2 and C3, providing both theoretical analysis and practical estimation methods.
3 Relaxed Candidate Set Models
We now develop the likelihood framework when conditions C2 and/or C3 are relaxed while maintaining C1. The key insight is that the general likelihood structure remains tractableβit simply requires modeling the masking mechanism explicitly rather than treating it as a nuisance.
3.1 General Likelihood Under C1
Theorem 3.1 (Likelihood Under C1 Alone).
Under Condition C1 alone, the likelihood contribution from an uncensored observation is:
| (18) |
Proof.
Under C1, when . Therefore, summing over :
| (19) | ||||
| (20) |
Remark 3.1 (Comparison with C1-C2-C3).
Under C2, the masking probability can be factored out of the sum since it is constant over . Under C3, it can be dropped since it does not depend on . When either condition fails, the masking probabilities remain inside the sum and may depend on both and , fundamentally changing the inference problem.
3.2 Relaxing C2: Informative Masking
When C2 is violated but C1 and C3 hold, the masking probability can vary with .
Definition 3.2 (Informative Masking).
Let for . The masking is informative if varies with .
Theorem 3.3 (Likelihood Under C1 and C3 (Relaxed C2)).
Under C1 and C3, the likelihood contribution is:
| (21) |
where does not depend on (by C3).
When is known, the likelihood remains tractable. The masking probabilities act as weights on the hazard contributions from each candidate.
3.2.1 Rank-Based Informative Masking
A practical model for informative masking assigns inclusion probabilities based on component failure ranks rather than absolute times.
Definition 3.4 (Rank-Based Masking).
Let denote the rank of component βs failure time among , where rank 1 corresponds to the earliest failure (the actual failed component).
The probability that component is in the candidate set is:
| (22) |
where controls the decay rate and is the maximum inclusion probability for non-failed components.
Remark 3.2 (Limiting Behavior).
-
β’
As : All non-failed components have probability (uninformative within the non-failed set).
-
β’
As : Only the failed component and rank-2 component have non-zero probabilities.
This model captures the intuition that components failing βnearly at the same timeβ as the actual failure are more likely to be included in the candidate set.
3.2.2 Independent Bernoulli Candidate Set Model
A flexible model assumes independent component inclusion:
Definition 3.5 (Independent Bernoulli Model).
Each component is included in the candidate set independently with probability , subject to C1 (failed component always included):
| (23) |
Under this model, the probability of observing candidate set given is:
| (24) |
Proposition 3.6 (Likelihood Under Bernoulli Model).
Under the independent Bernoulli model with C1 and known probabilities , the likelihood contribution is:
| (25) |
where
| (26) |
is the probability of observing given that failed (excluding the deterministic inclusion of ).
3.2.3 KL-Divergence Constrained Models
To systematically study deviations from the standard C1-C2-C3 model, we can parameterize informative masking by its distance from the baseline:
Definition 3.7 (KL-Divergence from Baseline).
Let denote the baseline Bernoulli model satisfying C1-C2-C3, where the failed component has probability 1 and all others have probability .
For a given target KL-divergence , we seek a masking probability vector satisfying:
-
1.
for the failed component (C1),
-
2.
,
-
3.
(same expected candidate set size).
When , we recover (the C1-C2-C3 model). As increases, becomes more informative about which component failed. This provides a controlled framework for studying the effects of departures from C2.
3.3 Relaxing C3: Parameter-Dependent Masking
When C3 is violated, the masking probability depends on .
Theorem 3.8 (Likelihood Under C1 and C2 (Relaxed C3)).
Under C1 and C2, the likelihood contribution is:
| (27) |
where is the (common) masking probability for any , now depending on .
Proof.
By C2, we can factor out the masking probability since it is constant over :
| (28) | ||||
| (29) |
Remark 3.3 (Nuisance Parameters).
When has a known functional form, it contributes to the likelihood and affects the MLE. If the form is unknown, additional modeling assumptions or profile likelihood approaches may be needed.
3.3.1 Failure-Probability-Weighted Masking
A natural way for masking to depend on is through the conditional failure probabilities:
Definition 3.9 (Failure-Probability-Weighted Masking).
The probability that component is in the candidate set depends on its posterior failure probability:
| (30) |
for some function with as .
This models diagnosticians who are more likely to include components with higher failure probabilities given the observed failure time. The function controls the sensitivity of masking to these probabilities.
3.4 The General Case: Both C2 and C3 Relaxed
When both C2 and C3 are relaxed, the likelihood takes the fully general form from TheoremΒ 3.1:
| (31) |
Estimation in this general case requires either:
-
1.
A fully specified parametric model for , or
-
2.
Sensitivity analysis over plausible masking mechanisms, or
-
3.
Nonparametric or semiparametric approaches that avoid specifying the masking mechanism.
In practice, the most common scenario is relaxed C2 with C3 maintained (informative but parameter-independent masking), which we focus on in the simulation studies.
4 Identifiability and Fisher Information
We now analyze identifiability conditions and derive the Fisher information matrix under relaxed conditions, focusing on exponential series systems for tractability.
4.1 Identifiability Under C1-C2-C3
Definition 4.1 (Identifiability).
A parameter is identifiable if for any with , there exists some data configuration such that
| (32) |
Theorem 4.2 (Identifiability Under C1-C2-C3).
Under C1, C2, and C3, the parameter is identifiable if and only if the following condition holds: For each pair of components , there exists at least one observed candidate set such that exactly one of or holds (i.e., the components do not always co-occur in candidate sets).
Proof.
The log-likelihood contribution from an uncensored observation is:
| (33) |
Sufficiency: If components and appear in different candidate sets, then information about can be obtained from observations where but , and vice versa. Combined with the survival term (which depends on all parameters), this provides sufficient variation to identify individual parameters.
Necessity: If components and always co-occur in every candidate set, the hazard sum always contains as an inseparable unit. The survival term provides information only about . Thus, any reparametrization preserving both and yields the same likelihood, demonstrating non-identifiability. β
4.2 Block Non-Identifiability
A particularly important case arises when components form blocks that always appear together:
Theorem 4.3 (Block Non-Identifiability).
Suppose components are partitioned into blocks such that for every observed candidate set :
-
(i)
For each block : either or , and
-
(ii)
If the failed component , then .
Then for exponential components with rates , only the block sums are identifiable.
Proof.
Under the exponential model with constant hazards, the likelihood becomes:
| (34) |
The survival term depends only on . For the hazard sum, under the block structure, each candidate set is a union of complete blocks. Thus:
| (35) |
Any reparametrization that preserves yields the same likelihood, hence individual within blocks are not identifiable. β
Example 4.1 (Three-Component Block Model).
Consider a 3-component system where the diagnostic tool can only distinguish:
-
β’
Components 1 and 2 share a circuit board (block ),
-
β’
Component 3 is separate (block ).
Candidate sets are either , , or . The MLE satisfies:
| (36) | ||||
| (37) |
but individual are not unique.
4.3 Improved Identifiability with Informative Masking
Surprisingly, relaxing C2 can improve identifiability:
Theorem 4.4 (Improved Identifiability with Informative Masking).
Under C1 and C3 with known informative masking probabilities , identifiability can be improved relative to the C1-C2-C3 case. Specifically, if components and always co-occur in candidate sets (violating the identifiability condition of TheoremΒ 4.2), they become identifiable if there exists a candidate set with such that for some .
Proof.
Under C1-C2-C3 with non-informative masking, if components and always co-occur, the hazard sum contains only the unweighted sum , making individual hazards non-identifiable.
Under informative masking with known weights , the likelihood contribution becomes:
| (38) |
The hazard sum now involves the weighted combination . If , this provides one equation involving and with unequal coefficients.
Combined with the survival term (which contributes through the system hazard), we have two linearly independent equations:
| (39) | ||||
| (40) |
When , this system has a unique solution, establishing identifiability of individual hazards. β
Remark 4.1.
Informative masking can paradoxically help estimation when the masking structure is known, because it provides additional information about which component likely failed.
4.4 Fisher Information for Exponential Series Systems
We now derive closed-form expressions for the Fisher information matrix, specializing to exponential components.
4.4.1 Exponential Series Model
For exponential components with rates :
| (41) | ||||
| (42) | ||||
| (43) |
4.4.2 Fisher Information Under C1-C2-C3
Theorem 4.5 (FIM Under C1-C2-C3).
For the exponential series system under C1, C2, C3, the observed Fisher information matrix has elements:
| (44) |
Proof.
The log-likelihood for an uncensored observation is:
| (45) |
The first derivatives (score) are:
| (46) |
The second derivatives are:
| (47) |
The observed FIM is the negative Hessian, giving the result. β
Remark 4.2.
The FIM depends on the candidate sets but not on the failure times (for exponential components). This reflects the memoryless property of the exponential distribution.
4.4.3 Fisher Information Under Relaxed C2
Theorem 4.6 (FIM Under Informative Masking).
Under C1, C3, and informative masking with known weights , the observed Fisher information matrix for exponential components is:
| (48) |
Proof.
The log-likelihood contribution is:
| (49) |
The score is:
| (50) |
The Hessian is:
| (51) |
4.5 Efficiency Comparison
Theorem 4.7 (Relative Efficiency).
Let and denote the Fisher information matrices under C1-C2-C3 and C1-C3 (informative masking) respectively. Then:
-
(a)
If for all (uniform weighting), then .
-
(b)
If masking is highly informative (concentrating weight on one component), can exceed for that componentβs parameter.
Proof.
(a) Under C1-C2-C3 (non-informative masking), the FIM element is:
| (52) |
Under C1-C3 with uniform weights for all :
| (53) | ||||
| (54) | ||||
| (55) |
(b) Suppose for component and for all in . Then:
| (56) |
This concentrates all information on , which can exceed when since the denominator under C1-C2-C3 is . β
Remark 4.3 (Practical Implications).
Informative masking can either help or hurt estimation:
-
β’
Helps when the masking structure is known and aligned with what we want to estimate.
-
β’
Hurts if masking is informative but we incorrectly assume C2 (non-informative), leading to model misspecification bias.
4.6 Estimation Under Model Misspecification
Theorem 4.8 (Bias from C2 Misspecification).
Suppose the true model is C1-C3 with informative masking , but estimation is performed assuming C1-C2-C3 (non-informative masking). The resulting MLE is generally biased, with bias depending on the correlation between and the hazard ratios .
Proof.
The score under the assumed (wrong) C1-C2-C3 model is:
| (57) |
The true score under C1-C3 (informative masking) is:
| (58) |
At the true parameter , the true score has expectation zero: .
The misspecified score has expectation:
| (59) |
This differs from zero when the masking weights are correlated with the hazard ratios. Specifically, define the βeffectiveβ weight under the true model. The MLE under the wrong model solves , yielding that satisfies:
| (60) |
When for components with larger , the misspecified model overestimates components that are more likely to be in candidate sets, producing systematic bias. β
This result motivates the simulation studies in SectionΒ 5, which quantify the bias under various misspecification scenarios.
5 Simulation Studies
We present simulation studies to (1) validate MLE performance under the C1-C2-C3 Bernoulli masking model, (2) quantify the bias from incorrectly assuming C2 when masking is informative, and (3) investigate identifiability issues arising from correlated candidate sets.
5.1 Experimental Design
5.1.1 System Configuration
We consider exponential series systems with components and true rate parameters:
| (61) |
These values represent a system where component 3 has the highest failure rate (and thus contributes most to system failures), while component 1 is most reliable.
5.1.2 Data Generation
For each simulation replicate:
-
1.
Generate component failure times for and .
-
2.
Compute system failure times and identify failed components .
-
3.
Apply right-censoring at time (chosen to achieve approximately 20% censoring) to obtain observed lifetimes and indicators .
-
4.
Generate candidate sets using the specified masking model.
5.1.3 Masking Models
We examine three masking scenarios:
-
1.
C1-C2-C3 (Baseline): Bernoulli model with for all non-failed components.
-
2.
Informative masking (Rank-based): Masking probabilities depend on component failure time ranks, parameterized by informativeness parameter .
-
3.
Correlated candidate sets: Candidate set indicators have correlation .
5.1.4 Performance Metrics
We evaluate:
-
β’
Bias:
-
β’
Root mean squared error (RMSE):
-
β’
Coverage probability: Proportion of 95% confidence intervals containing
-
β’
RMSE ratio:
5.2 Study 1: MLE Performance Under Bernoulli Masking
We first validate MLE performance under the correctly specified C1-C2-C3 Bernoulli masking model across sample sizes with Monte Carlo replicates.
5.2.1 Results
TableΒ 1 presents the estimation results.
| Parameter | Bias | RMSE | Coverage | Mean CI Width | |
|---|---|---|---|---|---|
| 50 | 0.017 | 0.477 | 0.920 | 1.727 | |
| 0.007 | 0.511 | 0.935 | 1.952 | ||
| 0.085 | 0.557 | 0.945 | 2.197 | ||
| 100 | 0.016 | 0.318 | 0.935 | 1.175 | |
| 0.055 | 0.390 | 0.935 | 1.385 | ||
| 0.366 | 0.950 | 1.519 | |||
| 200 | 0.201 | 0.935 | 0.825 | ||
| 0.008 | 0.262 | 0.965 | 0.965 | ||
| 0.258 | 0.955 | 1.066 |
Notes. Results based on 200 Monte Carlo replications. True parameters: , , . Bernoulli masking with , censoring proportion .
Key findings from Study 1:
-
1.
Consistency: Bias is small relative to RMSE at all sample sizes, indicating approximate unbiasedness.
-
2.
Convergence: RMSE decreases from approximately 0.5 at to 0.2β0.3 at , consistent with -rate convergence.
-
3.
Coverage: 95% CI coverage ranges from 92.0% to 96.5%, close to the nominal level, validating the Fisher information-based standard errors.
-
4.
Component effects: Components with higher true rates () have slightly larger absolute RMSE but similar relative performance.
5.3 Study 2: Misspecification Bias Analysis
We quantify the bias from incorrectly assuming C1-C2-C3 when masking is actually informative. Data is generated with rank-based informative masking (informativeness parameter ), then analyzed using both the correct model and the misspecified C2 model.
5.3.1 Results
TableΒ 2 compares bias under correct versus misspecified models.
| Parameter | Bias (Correct) | Bias (Misspec.) | RMSE Ratio | |
|---|---|---|---|---|
| 0 | 1.019 | |||
| 0.027 | 0.024 | 1.090 | ||
| 0.105 | 1.013 | |||
| 1 | 0.999 | |||
| 1.031 | ||||
| 0.192 | 0.146 | 0.944 | ||
| 5 | 1.002 | |||
| 0.010 | 1.031 | |||
| 0.147 | 0.130 | 0.988 | ||
| 10 | 1.022 | |||
| 0.000 | 0.002 | 0.993 | ||
| 0.183 | 0.178 | 1.008 |
Notes. corresponds to non-informative masking (C2 satisfied). As increases, masking becomes more informative. RMSE Ratio = RMSE(Misspecified) / RMSE(Correct); values indicate efficiency loss.
Key findings from Study 2:
-
1.
Moderate robustness: The RMSE ratio stays between 0.94 and 1.09 across all informativeness levels, indicating that misspecifying the masking model produces at most 9% efficiency loss.
-
2.
Bias similarity: Surprisingly, bias under the misspecified model closely tracks bias under the correct model, suggesting the C2 assumption is more robust than theoretical arguments might suggest.
-
3.
Parameter-specific effects: Component 3 () shows consistently positive bias under both models, likely due to its higher failure rate making it more frequently the true cause of failure.
5.4 Study 3: Identifiability and Candidate Set Correlation
We investigate how correlation between candidate set indicators affects identifiability by examining the Fisher Information Matrix (FIM) eigenvalues.
5.4.1 Results
TableΒ 3 presents FIM analysis by correlation level.
| Smallest Eigenvalue | Condition Number | |
|---|---|---|
| 0.0 | 12.23 | 2.18 |
| 0.1 | 12.93 | 2.12 |
| 0.3 | 14.17 | 2.01 |
| 0.5 | 15.00 | 1.97 |
| 0.6 | 15.12 | 1.91 |
| 0.8 | 14.33 | 2.01 |
| 0.9 | 13.88 | 2.04 |
Notes. measures correlation between candidate set indicators. As , components always co-occur in candidate sets, theoretically leading to non-identifiability.
Key findings from Study 3:
-
1.
Identifiability preserved: The smallest FIM eigenvalue remains substantially positive (12β15) across all correlation levels, indicating parameters remain identifiable.
-
2.
Condition number stable: The condition number stays below 2.2, indicating a well-conditioned estimation problem.
-
3.
Nonmonotonic pattern: Interestingly, the smallest eigenvalue peaks around β, suggesting moderate correlation may actually improve information content.
5.5 Summary of Simulation Results
Our simulation studies lead to the following conclusions:
-
1.
MLE performs well: Under the correctly specified C1-C2-C3 Bernoulli masking model, the MLE achieves coverage near nominal levels and RMSE consistent with asymptotic efficiency.
-
2.
C2 is robust: Misspecifying C2 (assuming non-informative masking when masking is informative) produces at most 9% efficiency loss and bias patterns similar to the correct model.
-
3.
Identifiability is robust: Even with high correlation () between candidate set indicators, parameters remain identifiable with stable FIM eigenvalues and condition numbers.
-
4.
Practical guidance: For sample sizes with moderate masking () and censoring (), the C1-C2-C3 Bernoulli model provides reliable inference. The additional complexity of modeling informative masking may not be justified unless there is strong evidence of systematic departures from C2.
| Metric | Study 1 | Study 2 | Study 3 |
|---|---|---|---|
| RMSE range | 0.20β0.56 | 0.19β0.47 | 0.23β0.47 |
| Coverage range | 92.0β96.5% | β | β |
| Max RMSE ratio | β | 1.09 | β |
| Min FIM eigenvalue | β | β | 12.23 |
| Max condition number | β | β | 2.18 |
6 Discussion
6.1 When to Use Relaxed Models
The theoretical and simulation results suggest the following practical guidance for choosing between standard C1-C2-C3 models and relaxed alternatives.
6.1.1 Use Standard C1-C2-C3 When:
-
1.
Masking mechanism is genuinely uninformative. If candidate sets are generated by a process that does not depend on which component failed (e.g., random equipment availability for testing), C2 holds.
-
2.
Masking probabilities are unknown. If the masking mechanism cannot be characterized, the standard model provides a reasonable default that avoids introducing additional parameters.
-
3.
Sample size is small. Even if masking is slightly informative, the bias may be dominated by sampling variability for small . The simpler model may provide more stable estimates.
-
4.
Primary interest is in relative component reliability. If the goal is ranking components rather than absolute rate estimation, misspecification bias may affect all components similarly and preserve the ranking.
6.1.2 Consider Relaxed Models When:
-
1.
Masking mechanism is known to be informative. If diagnostic procedures systematically favor certain components (e.g., those that βlook badβ at the failure time), C2 is violated and bias will result.
-
2.
Masking probabilities can be estimated. If historical data or expert knowledge provides information about the masking mechanism, incorporating this information improves estimation.
-
3.
Sample size is large enough to support additional parameters. Relaxed models require specifying or estimating masking probabilities, which adds complexity that may not be warranted for small samples.
-
4.
Identifiability concerns are present. As shown in TheoremΒ 4.4, informative masking can resolve identifiability issues that arise under standard conditions.
6.2 Practical Guidance
Based on our analysis, we recommend the following workflow:
-
1.
Assess the masking mechanism. Before estimation, consider how candidate sets are generated. Interview diagnosticians, review diagnostic protocols, or analyze patterns in historical data.
-
2.
Check for block structure. Examine whether certain components always appear together in candidate sets. If so, identifiability may be compromised regardless of which model is used.
-
3.
Perform sensitivity analysis. Fit models under both C1-C2-C3 and plausible relaxed assumptions. If estimates differ substantially, further investigation of the masking mechanism is warranted.
-
4.
Use simulation to assess impact. Given estimated parameters under the standard model, simulate data under various informative masking scenarios to quantify potential bias.
-
5.
Report uncertainty appropriately. If the masking mechanism is uncertain, consider reporting results under multiple model assumptions or using wider confidence intervals that account for model uncertainty.
6.3 Limitations
Our analysis has several limitations:
-
1.
Exponential focus. While we derive general likelihood results, the Fisher information analysis focuses on exponential components. Extension to Weibull or other distributions introduces additional complexity due to time-varying hazards.
-
2.
Known masking probabilities. Our relaxed models assume masking probabilities are known. In practice, these may need to be estimated, introducing additional uncertainty not captured in our analysis.
-
3.
Independence assumption. We assume masking for different observations is independent. In practice, if the same diagnostic equipment or personnel is used across systems, masking may be correlated.
-
4.
Parametric masking models. Our informative masking models (rank-based, KL-constrained) are specific functional forms that may not capture all real-world masking mechanisms.
-
5.
Simulation scope. The simulation studies cover a limited range of configurations. Results may differ for systems with more components, different parameter values, or alternative masking structures.
6.4 Future Directions
Several extensions would strengthen this work:
-
1.
Weibull extension. Derive Fisher information and conduct simulations for Weibull components, where shape parameters introduce additional estimation challenges.
-
2.
Semiparametric methods. Develop estimation approaches that avoid fully specifying the masking mechanism, perhaps using nonparametric or empirical likelihood methods.
-
3.
Model selection. Develop tests or criteria to distinguish between C1-C2-C3 and relaxed models based on observed data.
-
4.
Bayesian extensions. Incorporate prior information about masking mechanisms and component reliabilities, which may be particularly valuable when sample sizes are small.
-
5.
Sequential estimation. For systems observed over time, develop methods that update masking probability estimates as data accumulates.
-
6.
R package documentation. Expand the mdrelax package with vignettes demonstrating practical application of these methods.
7 Conclusion
We have developed a theoretical framework for likelihood-based inference in series systems with masked failure data when the traditional conditions C2 (non-informative masking) and C3 (parameter-independent masking) are relaxed. Our main contributions are:
-
1.
Likelihood derivations. We established the form of the likelihood under various relaxation scenarios, showing that the masking probabilities act as weights on component hazard contributions when C2 is violated.
-
2.
Practical masking models. We introduced rank-based informative masking and KL-divergence constrained models that provide interpretable parameterizations of non-standard masking.
-
3.
Identifiability results. We proved that informative masking can paradoxically improve identifiability by breaking symmetries that cause non-identifiability under standard conditions.
-
4.
Fisher information analysis. We derived closed-form expressions for the Fisher information matrix under informative masking for exponential series systems, enabling efficiency comparisons.
-
5.
Misspecification analysis. We characterized the bias that arises from incorrectly assuming C2 when masking is actually informative, providing guidance on when relaxed models are necessary.
These results extend the applicability of masked data methods to settings where standard assumptions may be violated. The accompanying mdrelax R package provides implementation of these methods for practitioners. Directions for future research are discussed in SectionΒ 6.4.
References
- [1] (2011) Systems in series. In Wiley Encyclopedia of Operations Research and Management Science, External Links: Document Cited by: Β§2.6.
- [2] (1995) A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing 16 (5), pp.Β 1190β1208. Cited by: Β§B.2.
- [3] (1972) Regression models and life-tables. Journal of the Royal Statistical Society: Series B (Methodological) 34 (2), pp.Β 187β202. External Links: Document Cited by: Β§2.6.
- [4] (2013) Estimating component reliabilities from incomplete system failure data. In Proceedings of the Annual Reliability and Maintainability Symposium (RAMS), pp.Β 1β6. External Links: Document Cited by: Β§2.4, Β§2.6.
- [5] (2005) Survival analysis: techniques for censored and truncated data. 2nd edition, Springer Science & Business Media. Cited by: Β§2.6.
- [6] (2023) Reliability estimation in series systems: maximum likelihood techniques for right-censored and masked failure data. Note: Masterβs thesis. Available: https://github.com/queelius/reliability-estimation-in-series-systems External Links: Link Cited by: Β§1, Β§2.1.
- [7] (1988) Maximum likelihood analysis of component reliability using masked system life-test data. IEEE Transactions on Reliability 37 (5), pp.Β 550β555. External Links: Document Cited by: Β§1, Β§2.4, Β§2.6.
- [8] (1993) Exact maximum likelihood estimation using masked system data. IEEE Transactions on Reliability 42 (4), pp.Β 631β635. External Links: Document Cited by: Β§2.4, Β§2.6.
Appendix A Proofs and Derivations
A.1 Score Function Under Informative Masking
For completeness, we provide the full derivation of the score function under the Bernoulli informative masking model for exponential series systems.
Let the log-likelihood contribution from an uncensored observation be:
| (62) |
The partial derivative with respect to is:
| (63) | ||||
| (64) |
The total score is:
| (65) |
Setting this to zero and solving gives the MLE equations under informative masking.
A.2 Hessian Matrix Derivation
The second partial derivatives are:
| (66) | ||||
| (67) |
The observed Fisher information matrix is the negative Hessian:
| (68) |
A.3 Expected Fisher Information
The expected FIM requires integrating over the distribution of candidate sets. For the exponential series system under C1-C2-C3 with Bernoulli masking (each non-failed component in candidate set with probability ), the expected Fisher information per observation is:
| (69) |
where the expectation is over both (failed component) and (candidate set).
This can be expanded as:
| (70) |
Under C1, the failed component is always in . The expectation over candidate sets involves summing over all possible weighted by their probabilities under the Bernoulli model:
| (71) |
Closed-form evaluation of this expectation is generally intractable due to the sum in the denominator. Monte Carlo estimation or numerical integration is typically required.
Appendix B Implementation Details
B.1 R Package Functions
The theoretical framework developed in this paper is implemented in the mdrelax R package. Key functions include:
-
β’
md_bernoulli_cand_C1_C2_C3(): Generates candidate set probabilities under the standard Bernoulli model satisfying C1-C2-C3.
-
β’
md_bernoulli_cand_C1_kld(): Generates candidate set probabilities with a specified KL-divergence from the baseline C1-C2-C3 model.
-
β’
informative_masking_by_rank(): Computes inclusion probabilities based on component failure time ranks.
-
β’
md_cand_sampler(): Samples candidate sets from probability vectors.
-
β’
md_loglike_exp_series_C1_C2_C3(): Log-likelihood function for exponential series systems under C1-C2-C3.
-
β’
md_mle_exp_series_C1_C2_C3(): Maximum likelihood estimation for exponential series systems.
-
β’
md_fim_exp_series_C1_C2_C3(): Observed Fisher information matrix for exponential series systems.
-
β’
md_block_candidate_m3(): Demonstrates block non-identifiability in a 3-component system.
B.2 Optimization Details
MLE is computed using the L-BFGS-B algorithm [2] with analytically computed gradients. The optimization is initialized using a method-of-moments estimator based on the total system hazard:
| (72) |
For challenging optimization landscapes, simulated annealing can be used to find a good starting point before local optimization.
Appendix C Additional Simulation Results
Additional simulation results supplement the findings in SectionΒ 5. The complete results are available in the R packageβs simulation directory.
-
1.
Full tables of bias, RMSE, and coverage for all parameter configurations are provided in the packageβs inst/simulations/results/ directory.
-
2.
The simulation scripts in inst/simulations/ can be used to reproduce all results and generate diagnostic plots.
-
3.
Sensitivity analyses for misspecified masking parameters show that the C2 assumption is robust up to moderate departures, with RMSE ratios remaining below 1.10 across tested configurations.
-
4.
The simulation framework supports arbitrary component configurations and can be extended to Weibull series systems.