Relaxed Candidate Set Models for Masked Data in Series Systems

Alex Towell
lex@metafunctor.com
(December 3, 2025)
Abstract

We develop likelihood-based inference methods for series systems with masked failure data when the traditional conditions governing candidate set formation are relaxed. While existing methods require that masking be non-informative (Condition C2) and parameter-independent (Condition C3), practical diagnostic systems often violate these assumptions. We derive the likelihood under various relaxation scenarios, establish identifiability conditions, and compare Fisher information under standard versus relaxed models. Our analysis reveals that informative masking, when properly modeled, can paradoxically improve estimation efficiency by providing additional information about the failed component. We implement these methods in an R package and provide simulation studies demonstrating the bias that arises from incorrectly assuming standard conditions when masking is informative.

Keywords: Series systems, masked data, reliability estimation, informative masking, candidate sets, Fisher information, maximum likelihood estimation

1 Introduction

Estimating the reliability of individual components in a series system presents a fundamental challenge in reliability engineering: system-level failure data is observable, but component-level failure causes are often masked. When a series system fails, diagnostic procedures may identify only a candidate set of components that could have caused the failure, rather than pinpointing the exact failed component [7]. This masking, combined with right-censoring of system lifetimes, complicates statistical inference about component reliability parameters.

Prior work on masked data in series systems has established a tractable likelihood framework under three conditions [6]:

  • β€’

    C1: The failed component is always contained in the candidate set.

  • β€’

    C2: Masking is non-informativeβ€”conditional on the failure time and candidate set, each component in the candidate set is equally likely to have failed.

  • β€’

    C3: The masking mechanism does not depend on the system parameters 𝜽.

Under these conditions, the likelihood has a simple closed form that enables efficient maximum likelihood estimation (MLE). However, practical diagnostic systems may violate C2 and C3:

  • β€’

    Experienced technicians may preferentially include components that β€œseem likely” to have failed based on failure time characteristics, violating C2.

  • β€’

    Diagnostic algorithms based on reliability rankings may systematically favor certain components, with the ranking depending on the true parameters, violating C3.

This paper develops the theoretical and computational framework for likelihood-based inference when C2 and/or C3 are relaxed while maintaining C1. Our contributions are:

  1. 1.

    General likelihood framework. We derive the likelihood under C1 alone, showing how informative and parameter-dependent masking modify the standard likelihood structure (SectionΒ 3).

  2. 2.

    Practical masking models. We introduce the rank-based informative masking model and the Bernoulli candidate set model with KL-divergence constraints, which provide interpretable parameterizations of non-standard masking (SectionΒ 3).

  3. 3.

    Identifiability analysis. We establish conditions under which parameters remain identifiable when standard conditions fail, including the surprising result that informative masking can improve identifiability in certain cases (SectionΒ 4).

  4. 4.

    Efficiency comparison. We derive the Fisher information matrix under relaxed conditions for exponential series systems, enabling precise comparison of estimation efficiency (SectionΒ 4).

  5. 5.

    Simulation studies. We quantify the bias from incorrectly assuming C2 when masking is informative, and demonstrate the improved estimation achievable when the masking structure is properly modeled (SectionΒ 5).

The remainder of this paper is organized as follows. SectionΒ 2 reviews series systems, masked data, and the standard C1-C2-C3 likelihood. SectionΒ 3 develops the likelihood under relaxed conditions. SectionΒ 4 analyzes identifiability and Fisher information. SectionΒ 5 presents simulation studies. SectionΒ 6 discusses practical implications, and SectionΒ 7 concludes.

2 Background

2.1 Series System Model

Consider a series system composed of m components. The lifetime of the i-th system is

Ti=min⁑{Ti⁒1,Ti⁒2,…,Ti⁒m}, (1)

where Ti⁒j denotes the lifetime of the j-th component in the i-th system. Component lifetimes are assumed independent with parametric distributions indexed by 𝜽j; the full parameter vector is 𝜽=(𝜽1,…,𝜽m)βˆˆπ›€.

Definition 2.1 (Component Distribution Functions).

For component j with parameter 𝛉j:

Rj⁒(t;𝜽j) =Pr⁒{Ti⁒j>t} (reliability function), (2)
fj⁒(t;𝜽j) =βˆ’dd⁒t⁒Rj⁒(t;𝜽j) (density function), (3)
hj⁒(t;𝜽j) =fj⁒(t;𝜽j)Rj⁒(t;𝜽j) (hazard function). (4)

For the series system, these functions combine as follows:

Theorem 2.2 (Series System Distribution Functions).

The series system has:

RTi⁒(t;𝜽) =∏j=1mRj⁒(t;𝜽j), (5)
hTi⁒(t;𝜽) =βˆ‘j=1mhj⁒(t;𝜽j), (6)
fTi⁒(t;𝜽) =hTi⁒(t;𝜽)β‹…RTi⁒(t;𝜽). (7)

The proof follows from the independence of component lifetimes and standard arguments [6].

2.2 Component Cause of Failure

Let Ki∈{1,…,m} denote the index of the component that caused system i to fail. Since the system fails when the first component fails, Ki=arg⁑minj⁑Ti⁒j.

Theorem 2.3 (Joint Distribution of (Ti,Ki)).

The joint distribution of system lifetime and component cause of failure is:

fTi,Ki⁒(t,k;𝜽)=hk⁒(t;𝜽k)β’βˆβ„“=1mRℓ⁒(t;πœ½β„“). (8)
Corollary 2.4 (Conditional Failure Probability).

Given that the system failed at time t, the probability that component j caused the failure is:

Pr⁒{Ki=j∣Ti=t}=hj⁒(t;𝜽j)βˆ‘β„“=1mhℓ⁒(t;πœ½β„“). (9)

This probability plays a central role in masked data analysis, as it represents the β€œtrue” probability that each component failed, which the masking mechanism partially obscures.

2.3 Masked Data Structure

Definition 2.5 (Observed Data).

For each system i, we observe:

  • β€’

    Si=min⁑{Ti,Ο„i}: Right-censored system lifetime,

  • β€’

    Ξ΄i=𝟏Ti≀τi: Event indicator (1 if failure observed, 0 if censored),

  • β€’

    CiβŠ†{1,…,m}: Candidate set (only observed when Ξ΄i=1).

The latent (unobserved) variables are:

  • β€’

    Ki∈{1,…,m}: Index of failed component,

  • β€’

    (Ti⁒1,…,Ti⁒m): Component failure times.

2.4 Traditional Conditions C1, C2, C3

The existing literature [7, 8, 4] establishes tractable likelihood-based inference under three conditions:

Condition 1 (C1: Failed Component in Candidate Set).

The candidate set always contains the failed component:

Pr⁒{Ki∈Ci}=1. (10)
Condition 2 (C2: Non-Informative Masking).

Given the failure time and that the failed component is in a candidate set c, the probability of observing c does not depend on which component in c failed:

Pr⁒{Ci=c∣Ti=t,Ki=j}=Pr⁒{Ci=c∣Ti=t,Ki=jβ€²} (11)

for all j,jβ€²βˆˆc.

Condition 3 (C3: Parameter-Independent Masking).

The masking probabilities do not depend on the system parameters:

Pr⁒{Ci=c∣Ti=t,Ki=j}=βi⁒(c,t,j), (12)

where Ξ²i does not depend on 𝛉.

2.5 Likelihood Under C1, C2, C3

Under all three conditions, the likelihood admits a tractable form:

Theorem 2.6 (Likelihood Under C1-C2-C3).

Under Conditions C1, C2, and C3, the likelihood contribution from an uncensored observation (si,ci) is proportional to:

Li⁒(𝜽)βˆβˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k). (13)

For a censored observation with lifetime si:

Li⁒(𝜽)=βˆβ„“=1mRℓ⁒(si;πœ½β„“). (14)
Proof.

By the chain rule:

fTi,Ci⁒(t,c;𝜽)=βˆ‘k=1mfTi,Ki⁒(t,k;𝜽)β‹…Pr𝜽⁒{Ci=c∣Ti=t,Ki=k}. (15)

Under C1, terms with kβˆ‰c vanish. Under C2, the masking probability is constant over k∈c, so:

fTi,Ci⁒(t,c;𝜽)=Pr⁒{Ci=c∣Ti=t,Ki∈c}β‹…βˆβ„“=1mRℓ⁒(t;πœ½β„“)β’βˆ‘k∈chk⁒(t;𝜽k). (16)

Under C3, the masking probability does not depend on 𝜽, yielding the proportionality in (13). ∎

The complete log-likelihood for n independent systems is:

ℓ⁒(𝜽)=βˆ‘i=1n[βˆ‘j=1mlog⁑Rj⁒(si;𝜽j)+Ξ΄i⁒log⁑(βˆ‘k∈cihk⁒(si;𝜽k))]. (17)

2.6 Related Work

The masked data problem in series systems was introduced by Usher and Hodgson [7], who developed MLE methods for exponential components. Usher et al. [8] extended this to Weibull components with exact maximum likelihood. Guo et al. [4] (Guo et al.) provided simulation studies validating the approach under various masking scenarios.

The informative censoring literature in survival analysis [5, 3] addresses related issues where the censoring mechanism depends on covariates or outcomes. However, the candidate set structure in masked data creates a distinct problem not fully addressed by standard informative censoring methods.

The competing risks framework [1] provides another perspective, viewing component failures as competing causes of system failure. However, standard competing risks methods assume the cause is observed, whereas masked data only provides partial information through candidate sets.

Our work extends the C1-C2-C3 framework by explicitly modeling departures from C2 and C3, providing both theoretical analysis and practical estimation methods.

3 Relaxed Candidate Set Models

We now develop the likelihood framework when conditions C2 and/or C3 are relaxed while maintaining C1. The key insight is that the general likelihood structure remains tractableβ€”it simply requires modeling the masking mechanism explicitly rather than treating it as a nuisance.

3.1 General Likelihood Under C1

Theorem 3.1 (Likelihood Under C1 Alone).

Under Condition C1 alone, the likelihood contribution from an uncensored observation (si,ci) is:

Li⁒(𝜽)=βˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k)β‹…Pr𝜽⁒{Ci=ci∣Ti=si,Ki=k}. (18)
Proof.

Under C1, Pr𝜽⁒{Ci=c∣Ti=t,Ki=k}=0 when kβˆ‰c. Therefore, summing over Ki:

fTi,Ci⁒(t,c;𝜽) =βˆ‘k=1mhk⁒(t;𝜽k)β’βˆβ„“=1mRℓ⁒(t;πœ½β„“)β‹…Pr𝜽⁒{Ci=c∣Ti=t,Ki=k} (19)
=βˆβ„“=1mRℓ⁒(t;πœ½β„“)β‹…βˆ‘k∈chk⁒(t;𝜽k)⁒Pr𝜽⁒{Ci=c∣Ti=t,Ki=k}.∎ (20)
Remark 3.1 (Comparison with C1-C2-C3).

Under C2, the masking probability can be factored out of the sum since it is constant over k∈c. Under C3, it can be dropped since it does not depend on 𝜽. When either condition fails, the masking probabilities remain inside the sum and may depend on both k and 𝜽, fundamentally changing the inference problem.

3.2 Relaxing C2: Informative Masking

When C2 is violated but C1 and C3 hold, the masking probability Pr⁒{Ci=c∣Ti=t,Ki=k} can vary with k∈c.

Definition 3.2 (Informative Masking).

Let Ο€k⁒c⁒(t)=Pr⁒{Ci=c∣Ti=t,Ki=k} for k∈c. The masking is informative if Ο€k⁒c⁒(t) varies with k.

Theorem 3.3 (Likelihood Under C1 and C3 (Relaxed C2)).

Under C1 and C3, the likelihood contribution is:

Li⁒(𝜽)=βˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k)β‹…Ο€k,ci⁒(si), (21)

where Ο€k,c⁒(t) does not depend on 𝛉 (by C3).

When Ο€k⁒c⁒(t) is known, the likelihood remains tractable. The masking probabilities act as weights on the hazard contributions from each candidate.

3.2.1 Rank-Based Informative Masking

A practical model for informative masking assigns inclusion probabilities based on component failure ranks rather than absolute times.

Definition 3.4 (Rank-Based Masking).

Let rk⁒(𝐭)∈{1,…,m} denote the rank of component k’s failure time among (t1,…,tm), where rank 1 corresponds to the earliest failure (the actual failed component).

The probability that component j is in the candidate set is:

qj={1if ⁒rj=1⁒ (failed component),β⁒exp⁑(βˆ’Ξ±β’(rjβˆ’2))if ⁒rjβ‰₯2, (22)

where Ξ±β‰₯0 controls the decay rate and β∈[0,1] is the maximum inclusion probability for non-failed components.

Remark 3.2 (Limiting Behavior).
  • β€’

    As Ξ±β†’0: All non-failed components have probability Ξ² (uninformative within the non-failed set).

  • β€’

    As Ξ±β†’βˆž: Only the failed component and rank-2 component have non-zero probabilities.

This model captures the intuition that components failing β€œnearly at the same time” as the actual failure are more likely to be included in the candidate set.

3.2.2 Independent Bernoulli Candidate Set Model

A flexible model assumes independent component inclusion:

Definition 3.5 (Independent Bernoulli Model).

Each component j is included in the candidate set independently with probability qj, subject to C1 (failed component always included):

Pr⁒{j∈Ci∣Ti=t,Ki=k}={1if ⁒j=k,qjotherwise. (23)

Under this model, the probability of observing candidate set c given (Ti=t,Ki=k) is:

Pr⁒{Ci=c∣Ti=t,Ki=k}=𝟏k∈c⁒∏j∈cβˆ–{k}qj⁒∏jβˆ‰c(1βˆ’qj). (24)
Proposition 3.6 (Likelihood Under Bernoulli Model).

Under the independent Bernoulli model with C1 and known probabilities (q1,…,qm), the likelihood contribution is:

Li⁒(𝜽)=βˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k)⁒wk⁒(ci), (25)

where

wk⁒(c)=∏j∈cβˆ–{k}qj⁒∏jβˆ‰c(1βˆ’qj) (26)

is the probability of observing c given that k failed (excluding the deterministic inclusion of k).

3.2.3 KL-Divergence Constrained Models

To systematically study deviations from the standard C1-C2-C3 model, we can parameterize informative masking by its distance from the baseline:

Definition 3.7 (KL-Divergence from Baseline).

Let P=(p,…,p,1,p,…,p) denote the baseline Bernoulli model satisfying C1-C2-C3, where the failed component has probability 1 and all others have probability p.

For a given target KL-divergence dβ‰₯0, we seek a masking probability vector Q=(q1,…,qm) satisfying:

  1. 1.

    qk=1 for the failed component (C1),

  2. 2.

    KL⁒(Pβˆ₯Q)β‰ˆd,

  3. 3.

    βˆ‘jqj=βˆ‘jpj (same expected candidate set size).

When d=0, we recover Q=P (the C1-C2-C3 model). As d increases, Q becomes more informative about which component failed. This provides a controlled framework for studying the effects of departures from C2.

3.3 Relaxing C3: Parameter-Dependent Masking

When C3 is violated, the masking probability Pr𝜽⁒{Ci=c∣Ti=t,Ki=k} depends on 𝜽.

Theorem 3.8 (Likelihood Under C1 and C2 (Relaxed C3)).

Under C1 and C2, the likelihood contribution is:

Li⁒(𝜽)=Ο€ci⁒(si;𝜽)β‹…βˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k), (27)

where Ο€c⁒(t;𝛉)=Pr𝛉⁒{Ci=c∣Ti=t,Ki∈c} is the (common) masking probability for any k∈c, now depending on 𝛉.

Proof.

By C2, we can factor out the masking probability since it is constant over k∈c:

fTi,Ci⁒(t,c;𝜽) =βˆβ„“=1mRℓ⁒(t;πœ½β„“)β’βˆ‘k∈chk⁒(t;𝜽k)⁒Pr𝜽⁒{Ci=c∣Ti=t,Ki=k} (28)
=Ο€c⁒(t;𝜽)β’βˆβ„“=1mRℓ⁒(t;πœ½β„“)β’βˆ‘k∈chk⁒(t;𝜽k).∎ (29)
Remark 3.3 (Nuisance Parameters).

When Ο€c⁒(t;𝜽) has a known functional form, it contributes to the likelihood and affects the MLE. If the form is unknown, additional modeling assumptions or profile likelihood approaches may be needed.

3.3.1 Failure-Probability-Weighted Masking

A natural way for masking to depend on 𝜽 is through the conditional failure probabilities:

Definition 3.9 (Failure-Probability-Weighted Masking).

The probability that component j is in the candidate set depends on its posterior failure probability:

Pr𝜽⁒{j∈Ci∣Ti=t}=g⁒(hj⁒(t;𝜽j)βˆ‘β„“=1mhℓ⁒(t;πœ½β„“)) (30)

for some function g:[0,1]β†’[0,1] with g⁒(x)β†’1 as xβ†’1.

This models diagnosticians who are more likely to include components with higher failure probabilities given the observed failure time. The function g controls the sensitivity of masking to these probabilities.

3.4 The General Case: Both C2 and C3 Relaxed

When both C2 and C3 are relaxed, the likelihood takes the fully general form from TheoremΒ 3.1:

Li⁒(𝜽)=βˆβ„“=1mRℓ⁒(si;πœ½β„“)β‹…βˆ‘k∈cihk⁒(si;𝜽k)β‹…Ο€k,ci⁒(si;𝜽). (31)

Estimation in this general case requires either:

  1. 1.

    A fully specified parametric model for Ο€k,c⁒(t;𝜽), or

  2. 2.

    Sensitivity analysis over plausible masking mechanisms, or

  3. 3.

    Nonparametric or semiparametric approaches that avoid specifying the masking mechanism.

In practice, the most common scenario is relaxed C2 with C3 maintained (informative but parameter-independent masking), which we focus on in the simulation studies.

4 Identifiability and Fisher Information

We now analyze identifiability conditions and derive the Fisher information matrix under relaxed conditions, focusing on exponential series systems for tractability.

4.1 Identifiability Under C1-C2-C3

Definition 4.1 (Identifiability).

A parameter 𝛉 is identifiable if for any 𝛉,π›‰β€²βˆˆπ›€ with 𝛉≠𝛉′, there exists some data configuration D such that

L⁒(𝜽;D)β‰ L⁒(πœ½β€²;D). (32)
Theorem 4.2 (Identifiability Under C1-C2-C3).

Under C1, C2, and C3, the parameter 𝛉 is identifiable if and only if the following condition holds: For each pair of components jβ‰ jβ€², there exists at least one observed candidate set c such that exactly one of j∈c or jβ€²βˆˆc holds (i.e., the components do not always co-occur in candidate sets).

Proof.

The log-likelihood contribution from an uncensored observation is:

β„“i⁒(𝜽)=βˆ‘β„“=1mlog⁑Rℓ⁒(si;πœ½β„“)+log⁑(βˆ‘k∈cihk⁒(si;𝜽k)). (33)

Sufficiency: If components j and jβ€² appear in different candidate sets, then information about hj can be obtained from observations where j∈ci but jβ€²βˆ‰ci, and vice versa. Combined with the survival term (which depends on all parameters), this provides sufficient variation to identify individual parameters.

Necessity: If components j and jβ€² always co-occur in every candidate set, the hazard sum always contains hj+hjβ€² as an inseparable unit. The survival term provides information only about βˆ‘β„“Ξ»β„“. Thus, any reparametrization preserving both hj+hjβ€² and βˆ‘β„“hβ„“ yields the same likelihood, demonstrating non-identifiability. ∎

4.2 Block Non-Identifiability

A particularly important case arises when components form blocks that always appear together:

Theorem 4.3 (Block Non-Identifiability).

Suppose components are partitioned into blocks B1,…,Br such that for every observed candidate set ci:

  1. (i)

    For each block Bβ„“: either Bβ„“βŠ†ci or Bβ„“βˆ©ci=βˆ…, and

  2. (ii)

    If the failed component k∈Bβ„“, then Bβ„“βŠ†ci.

Then for exponential components with rates (Ξ»1,…,Ξ»m), only the block sums Ξ›β„“=βˆ‘j∈Bβ„“Ξ»j are identifiable.

Proof.

Under the exponential model with constant hazards, the likelihood becomes:

L⁒(𝜽)∝∏i:Ξ΄i=1exp⁑(βˆ’siβ’βˆ‘j=1mΞ»j)β’βˆ‘k∈ciΞ»k. (34)

The survival term depends only on βˆ‘j=1mΞ»j. For the hazard sum, under the block structure, each candidate set ci is a union of complete blocks. Thus:

βˆ‘k∈ciΞ»k=βˆ‘β„“:Bβ„“βŠ†ciΞ›β„“. (35)

Any reparametrization that preserves (Ξ›1,…,Ξ›r) yields the same likelihood, hence individual Ξ»j within blocks are not identifiable. ∎

Example 4.1 (Three-Component Block Model).

Consider a 3-component system where the diagnostic tool can only distinguish:

  • β€’

    Components 1 and 2 share a circuit board (block B1={1,2}),

  • β€’

    Component 3 is separate (block B2={3}).

Candidate sets are either {1,2}, {3}, or {1,2,3}. The MLE satisfies:

Ξ»^1+Ξ»^2 =Ξ»1+Ξ»2, (36)
Ξ»^3 =Ξ»3, (37)

but individual Ξ»^1,Ξ»^2 are not unique.

4.3 Improved Identifiability with Informative Masking

Surprisingly, relaxing C2 can improve identifiability:

Theorem 4.4 (Improved Identifiability with Informative Masking).

Under C1 and C3 with known informative masking probabilities Ο€k⁒c⁒(t), identifiability can be improved relative to the C1-C2-C3 case. Specifically, if components k and kβ€² always co-occur in candidate sets (violating the identifiability condition of TheoremΒ 4.2), they become identifiable if there exists a candidate set c with k,kβ€²βˆˆc such that Ο€k⁒c⁒(t)β‰ Ο€k′⁒c⁒(t) for some t>0.

Proof.

Under C1-C2-C3 with non-informative masking, if components k and kβ€² always co-occur, the hazard sum contains only the unweighted sum hk+hkβ€², making individual hazards non-identifiable.

Under informative masking with known weights Ο€k⁒c, the likelihood contribution becomes:

Li⁒(𝜽)=R⁒(si;𝜽)β’βˆ‘j∈cihj⁒(si;𝜽j)⁒πj,ci⁒(si). (38)

The hazard sum now involves the weighted combination hk⁒πk⁒c+hk′⁒πk′⁒c. If Ο€k⁒cβ‰ Ο€k′⁒c, this provides one equation involving hk and hkβ€² with unequal coefficients.

Combined with the survival term (which contributes hk+hkβ€² through the system hazard), we have two linearly independent equations:

Ο€k⁒c⁒hk+Ο€k′⁒c⁒hkβ€² =A1(from hazard sum), (39)
hk+hkβ€² =A2(from survival term). (40)

When Ο€k⁒cβ‰ Ο€k′⁒c, this system has a unique solution, establishing identifiability of individual hazards. ∎

Remark 4.1.

Informative masking can paradoxically help estimation when the masking structure is known, because it provides additional information about which component likely failed.

4.4 Fisher Information for Exponential Series Systems

We now derive closed-form expressions for the Fisher information matrix, specializing to exponential components.

4.4.1 Exponential Series Model

For exponential components with rates 𝝀=(Ξ»1,…,Ξ»m):

Rj⁒(t;Ξ»j) =eβˆ’Ξ»j⁒t, (41)
hj⁒(t;λj) =λj, (42)
RTi⁒(t;𝝀) =eβˆ’Ξ›β’t,where ⁒Λ=βˆ‘j=1mΞ»j. (43)

4.4.2 Fisher Information Under C1-C2-C3

Theorem 4.5 (FIM Under C1-C2-C3).

For the exponential series system under C1, C2, C3, the observed Fisher information matrix has elements:

ℐj⁒k⁒(𝝀)=βˆ‘i:Ξ΄i=1𝟏j∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciΞ»β„“)2. (44)
Proof.

The log-likelihood for an uncensored observation i is:

β„“i⁒(𝝀)=βˆ’si⁒Λ+log⁑(βˆ‘k∈ciΞ»k). (45)

The first derivatives (score) are:

βˆ‚β„“iβˆ‚Ξ»j=βˆ’si+𝟏j∈ciβˆ‘k∈ciΞ»k. (46)

The second derivatives are:

βˆ‚2β„“iβˆ‚Ξ»jβ’βˆ‚Ξ»k=βˆ’πŸj∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciΞ»β„“)2. (47)

The observed FIM is the negative Hessian, giving the result. ∎

Remark 4.2.

The FIM depends on the candidate sets but not on the failure times (for exponential components). This reflects the memoryless property of the exponential distribution.

4.4.3 Fisher Information Under Relaxed C2

Theorem 4.6 (FIM Under Informative Masking).

Under C1, C3, and informative masking with known weights Ο€k⁒c, the observed Fisher information matrix for exponential components is:

ℐj⁒k⁒(𝝀)=βˆ‘i:Ξ΄i=1Ο€j,ci⁒πk,ci(βˆ‘β„“βˆˆciλℓ⁒πℓ,ci)2⁒𝟏j∈ci⁒𝟏k∈ci. (48)
Proof.

The log-likelihood contribution is:

β„“i⁒(𝝀)=βˆ’si⁒Λ+log⁑(βˆ‘k∈ciΞ»k⁒πk,ci). (49)

The score is:

βˆ‚β„“iβˆ‚Ξ»j=βˆ’si+Ο€j,ci⁒𝟏j∈ciβˆ‘k∈ciΞ»k⁒πk,ci. (50)

The Hessian is:

βˆ‚2β„“iβˆ‚Ξ»jβ’βˆ‚Ξ»k=βˆ’Ο€j,ci⁒πk,ci⁒𝟏j∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciλℓ⁒πℓ,ci)2.∎ (51)

4.5 Efficiency Comparison

Theorem 4.7 (Relative Efficiency).

Let ℐ(C123) and ℐ(C13) denote the Fisher information matrices under C1-C2-C3 and C1-C3 (informative masking) respectively. Then:

  1. (a)

    If Ο€k⁒c=1/|c| for all k∈c (uniform weighting), then ℐ(C13)=ℐ(C123).

  2. (b)

    If masking is highly informative (concentrating weight on one component), ℐ(C13) can exceed ℐ(C123) for that component’s parameter.

Proof.

(a) Under C1-C2-C3 (non-informative masking), the FIM element is:

ℐj⁒k(C123)=βˆ‘i:Ξ΄i=1𝟏j∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciΞ»β„“)2. (52)

Under C1-C3 with uniform weights πℓ⁒c=1/|c| for all β„“βˆˆc:

ℐj⁒k(C13) =βˆ‘i:Ξ΄i=1(1/|ci|)2⁒𝟏j∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciλℓ⁒(1/|ci|))2 (53)
=βˆ‘i:Ξ΄i=1(1/|ci|2)⁒𝟏j∈ci⁒𝟏k∈ci(1/|ci|)2⁒(βˆ‘β„“βˆˆciΞ»β„“)2 (54)
=βˆ‘i:Ξ΄i=1𝟏j∈ci⁒𝟏k∈ci(βˆ‘β„“βˆˆciΞ»β„“)2=ℐj⁒k(C123). (55)

(b) Suppose Ο€k⁒c=1 for component k and Ο€k′⁒c=0 for all kβ€²β‰ k in c. Then:

ℐk⁒k(C13)=βˆ‘i:Ξ΄i=1,k∈ci1Ξ»k2. (56)

This concentrates all information on Ξ»k, which can exceed ℐk⁒k(C123) when |ci|>1 since the denominator under C1-C2-C3 is (βˆ‘β„“βˆˆciΞ»β„“)2>Ξ»k2. ∎

Remark 4.3 (Practical Implications).

Informative masking can either help or hurt estimation:

  • β€’

    Helps when the masking structure is known and aligned with what we want to estimate.

  • β€’

    Hurts if masking is informative but we incorrectly assume C2 (non-informative), leading to model misspecification bias.

4.6 Estimation Under Model Misspecification

Theorem 4.8 (Bias from C2 Misspecification).

Suppose the true model is C1-C3 with informative masking Ο€k⁒c⁒(t), but estimation is performed assuming C1-C2-C3 (non-informative masking). The resulting MLE 𝛉^ is generally biased, with bias depending on the correlation between Ο€k⁒c and the hazard ratios hk⁒(t;𝛉k)/βˆ‘β„“βˆˆchℓ⁒(t;𝛉ℓ).

Proof.

The score under the assumed (wrong) C1-C2-C3 model is:

βˆ‚β„“iwrongβˆ‚Ξ»j=βˆ’si+𝟏j∈ciβˆ‘k∈ciΞ»k. (57)

The true score under C1-C3 (informative masking) is:

βˆ‚β„“itrueβˆ‚Ξ»j=βˆ’si+Ο€j,ci⁒𝟏j∈ciβˆ‘k∈ciΞ»k⁒πk,ci. (58)

At the true parameter πœ½βˆ—, the true score has expectation zero: 𝔼⁒[βˆ‚β„“itrue/βˆ‚Ξ»j]=0.

The misspecified score has expectation:

𝔼⁒[βˆ‚β„“iwrongβˆ‚Ξ»j] =βˆ’π”Όβ’[si]+𝔼⁒[𝟏j∈ciβˆ‘k∈ciΞ»kβˆ—]. (59)

This differs from zero when the masking weights Ο€k⁒c are correlated with the hazard ratios. Specifically, define the β€œeffective” weight wj=𝔼⁒[𝟏j∈c/βˆ‘k∈cΞ»kβˆ—] under the true model. The MLE under the wrong model solves 𝔼⁒[βˆ‚β„“wrong/βˆ‚Ξ»j]=0, yielding Ξ»^j that satisfies:

Ξ»^j=𝔼⁒[𝟏j∈ci]𝔼⁒[βˆ‘k∈ciΞ»kβˆ—β‹…si/βˆ‘k∈ciΞ»kβˆ—]. (60)

When Ο€j⁒c>Ο€k⁒c for components with larger Ξ»jβˆ—, the misspecified model overestimates components that are more likely to be in candidate sets, producing systematic bias. ∎

This result motivates the simulation studies in SectionΒ 5, which quantify the bias under various misspecification scenarios.

5 Simulation Studies

We present simulation studies to (1) validate MLE performance under the C1-C2-C3 Bernoulli masking model, (2) quantify the bias from incorrectly assuming C2 when masking is informative, and (3) investigate identifiability issues arising from correlated candidate sets.

5.1 Experimental Design

5.1.1 System Configuration

We consider exponential series systems with m=3 components and true rate parameters:

π€βˆ—=(Ξ»1βˆ—,Ξ»2βˆ—,Ξ»3βˆ—)=(1.0,1.5,2.0). (61)

These values represent a system where component 3 has the highest failure rate (and thus contributes most to system failures), while component 1 is most reliable.

5.1.2 Data Generation

For each simulation replicate:

  1. 1.

    Generate component failure times Ti⁒j∼Exp⁒(Ξ»jβˆ—) for i=1,…,n and j=1,…,m.

  2. 2.

    Compute system failure times Ti=minj⁑Ti⁒j and identify failed components Ki=arg⁑minj⁑Ti⁒j.

  3. 3.

    Apply right-censoring at time Ο„ (chosen to achieve approximately 20% censoring) to obtain observed lifetimes Si=min⁑(Ti,Ο„) and indicators Ξ΄i.

  4. 4.

    Generate candidate sets using the specified masking model.

5.1.3 Masking Models

We examine three masking scenarios:

  1. 1.

    C1-C2-C3 (Baseline): Bernoulli model with p=0.3 for all non-failed components.

  2. 2.

    Informative masking (Rank-based): Masking probabilities depend on component failure time ranks, parameterized by informativeness parameter α∈{0,1,2,5,10}.

  3. 3.

    Correlated candidate sets: Candidate set indicators have correlation ρ∈{0,0.1,0.3,0.5,0.6,0.8,0.9}.

5.1.4 Performance Metrics

We evaluate:

  • β€’

    Bias: Bias⁒(Ξ»^j)=𝔼⁒[Ξ»^j]βˆ’Ξ»jβˆ—

  • β€’

    Root mean squared error (RMSE): RMSE⁒(Ξ»^j)=𝔼⁒[(Ξ»^jβˆ’Ξ»jβˆ—)2]

  • β€’

    Coverage probability: Proportion of 95% confidence intervals containing Ξ»jβˆ—

  • β€’

    RMSE ratio: RMSEmisspec/RMSEcorrect

5.2 Study 1: MLE Performance Under Bernoulli Masking

We first validate MLE performance under the correctly specified C1-C2-C3 Bernoulli masking model across sample sizes n∈{50,100,200} with B=200 Monte Carlo replicates.

5.2.1 Results

TableΒ 1 presents the estimation results.

Table 1: Maximum Likelihood Estimation Performance by Sample Size
n Parameter Bias RMSE Coverage Mean CI Width
50 Ξ»1 0.017 0.477 0.920 1.727
Ξ»2 0.007 0.511 0.935 1.952
Ξ»3 0.085 0.557 0.945 2.197
100 Ξ»1 0.016 0.318 0.935 1.175
Ξ»2 0.055 0.390 0.935 1.385
Ξ»3 βˆ’0.037 0.366 0.950 1.519
200 Ξ»1 βˆ’0.005 0.201 0.935 0.825
Ξ»2 0.008 0.262 0.965 0.965
Ξ»3 βˆ’0.033 0.258 0.955 1.066

Notes. Results based on 200 Monte Carlo replications. True parameters: Ξ»1=1.0, Ξ»2=1.5, Ξ»3=2.0. Bernoulli masking with p=0.3, censoring proportion β‰ˆ20%.

Refer to caption
Figure 1: RMSE of MLE by sample size. All three component rate parameters show decreasing RMSE as sample size increases, consistent with n-convergence.
Refer to caption
Figure 2: 95% confidence interval coverage probability by sample size. Coverage is near the nominal 95% level across all parameters and sample sizes, validating the asymptotic normal approximation.

Key findings from Study 1:

  1. 1.

    Consistency: Bias is small relative to RMSE at all sample sizes, indicating approximate unbiasedness.

  2. 2.

    Convergence: RMSE decreases from approximately 0.5 at n=50 to 0.2–0.3 at n=200, consistent with n-rate convergence.

  3. 3.

    Coverage: 95% CI coverage ranges from 92.0% to 96.5%, close to the nominal level, validating the Fisher information-based standard errors.

  4. 4.

    Component effects: Components with higher true rates (Ξ»3=2.0) have slightly larger absolute RMSE but similar relative performance.

5.3 Study 2: Misspecification Bias Analysis

We quantify the bias from incorrectly assuming C1-C2-C3 when masking is actually informative. Data is generated with rank-based informative masking (informativeness parameter Ξ±), then analyzed using both the correct model and the misspecified C2 model.

5.3.1 Results

TableΒ 2 compares bias under correct versus misspecified models.

Table 2: Bias Comparison: Correct vs Misspecified Model
Ξ± Parameter Bias (Correct) Bias (Misspec.) RMSE Ratio
0 Ξ»1 βˆ’0.129 βˆ’0.001 1.019
Ξ»2 0.027 0.024 1.090
Ξ»3 0.105 βˆ’0.020 1.013
1 Ξ»1 βˆ’0.136 βˆ’0.095 0.999
Ξ»2 βˆ’0.009 βˆ’0.004 1.031
Ξ»3 0.192 0.146 0.944
5 Ξ»1 βˆ’0.129 βˆ’0.127 1.002
Ξ»2 βˆ’0.005 0.010 1.031
Ξ»3 0.147 0.130 0.988
10 Ξ»1 βˆ’0.153 βˆ’0.152 1.022
Ξ»2 0.000 0.002 0.993
Ξ»3 0.183 0.178 1.008

Notes. Ξ±=0 corresponds to non-informative masking (C2 satisfied). As Ξ± increases, masking becomes more informative. RMSE Ratio = RMSE(Misspecified) / RMSE(Correct); values >1 indicate efficiency loss.

Refer to caption
Figure 3: Bias comparison between correct and misspecified models as masking informativeness increases. The misspecified model (incorrectly assuming C2) shows similar bias patterns to the correct model for moderate informativeness.
Refer to caption
Figure 4: RMSE ratio (misspecified/correct) by informativeness parameter. Values near 1 indicate minimal efficiency loss from misspecification. The maximum ratio of 1.09 suggests the C2 assumption is reasonably robust.

Key findings from Study 2:

  1. 1.

    Moderate robustness: The RMSE ratio stays between 0.94 and 1.09 across all informativeness levels, indicating that misspecifying the masking model produces at most 9% efficiency loss.

  2. 2.

    Bias similarity: Surprisingly, bias under the misspecified model closely tracks bias under the correct model, suggesting the C2 assumption is more robust than theoretical arguments might suggest.

  3. 3.

    Parameter-specific effects: Component 3 (Ξ»3) shows consistently positive bias under both models, likely due to its higher failure rate making it more frequently the true cause of failure.

5.4 Study 3: Identifiability and Candidate Set Correlation

We investigate how correlation between candidate set indicators affects identifiability by examining the Fisher Information Matrix (FIM) eigenvalues.

5.4.1 Results

TableΒ 3 presents FIM analysis by correlation level.

Table 3: Fisher Information Matrix Analysis by Candidate Set Correlation
ρ Smallest Eigenvalue Condition Number
0.0 12.23 2.18
0.1 12.93 2.12
0.3 14.17 2.01
0.5 15.00 1.97
0.6 15.12 1.91
0.8 14.33 2.01
0.9 13.88 2.04

Notes. ρ measures correlation between candidate set indicators. As ρ→1, components always co-occur in candidate sets, theoretically leading to non-identifiability.

Refer to caption
Figure 5: Smallest FIM eigenvalue versus candidate set correlation. The eigenvalue remains bounded away from zero even at ρ=0.9, indicating identifiability is maintained in our simulation setup.
Refer to caption
Figure 6: RMSE of MLE by candidate set correlation. Performance remains stable across correlation levels, consistent with the FIM eigenvalue analysis.

Key findings from Study 3:

  1. 1.

    Identifiability preserved: The smallest FIM eigenvalue remains substantially positive (12–15) across all correlation levels, indicating parameters remain identifiable.

  2. 2.

    Condition number stable: The condition number stays below 2.2, indicating a well-conditioned estimation problem.

  3. 3.

    Nonmonotonic pattern: Interestingly, the smallest eigenvalue peaks around ρ=0.5–0.6, suggesting moderate correlation may actually improve information content.

5.5 Summary of Simulation Results

Our simulation studies lead to the following conclusions:

  1. 1.

    MLE performs well: Under the correctly specified C1-C2-C3 Bernoulli masking model, the MLE achieves coverage near nominal levels and RMSE consistent with asymptotic efficiency.

  2. 2.

    C2 is robust: Misspecifying C2 (assuming non-informative masking when masking is informative) produces at most 9% efficiency loss and bias patterns similar to the correct model.

  3. 3.

    Identifiability is robust: Even with high correlation (ρ=0.9) between candidate set indicators, parameters remain identifiable with stable FIM eigenvalues and condition numbers.

  4. 4.

    Practical guidance: For sample sizes nβ‰₯100 with moderate masking (pβ‰ˆ0.3) and censoring (β‰ˆ20%), the C1-C2-C3 Bernoulli model provides reliable inference. The additional complexity of modeling informative masking may not be justified unless there is strong evidence of systematic departures from C2.

Table 4: Summary of key simulation findings.
Metric Study 1 Study 2 Study 3
RMSE range 0.20–0.56 0.19–0.47 0.23–0.47
Coverage range 92.0–96.5% β€” β€”
Max RMSE ratio β€” 1.09 β€”
Min FIM eigenvalue β€” β€” 12.23
Max condition number β€” β€” 2.18

6 Discussion

6.1 When to Use Relaxed Models

The theoretical and simulation results suggest the following practical guidance for choosing between standard C1-C2-C3 models and relaxed alternatives.

6.1.1 Use Standard C1-C2-C3 When:

  1. 1.

    Masking mechanism is genuinely uninformative. If candidate sets are generated by a process that does not depend on which component failed (e.g., random equipment availability for testing), C2 holds.

  2. 2.

    Masking probabilities are unknown. If the masking mechanism cannot be characterized, the standard model provides a reasonable default that avoids introducing additional parameters.

  3. 3.

    Sample size is small. Even if masking is slightly informative, the bias may be dominated by sampling variability for small n. The simpler model may provide more stable estimates.

  4. 4.

    Primary interest is in relative component reliability. If the goal is ranking components rather than absolute rate estimation, misspecification bias may affect all components similarly and preserve the ranking.

6.1.2 Consider Relaxed Models When:

  1. 1.

    Masking mechanism is known to be informative. If diagnostic procedures systematically favor certain components (e.g., those that β€œlook bad” at the failure time), C2 is violated and bias will result.

  2. 2.

    Masking probabilities can be estimated. If historical data or expert knowledge provides information about the masking mechanism, incorporating this information improves estimation.

  3. 3.

    Sample size is large enough to support additional parameters. Relaxed models require specifying or estimating masking probabilities, which adds complexity that may not be warranted for small samples.

  4. 4.

    Identifiability concerns are present. As shown in TheoremΒ 4.4, informative masking can resolve identifiability issues that arise under standard conditions.

6.2 Practical Guidance

Based on our analysis, we recommend the following workflow:

  1. 1.

    Assess the masking mechanism. Before estimation, consider how candidate sets are generated. Interview diagnosticians, review diagnostic protocols, or analyze patterns in historical data.

  2. 2.

    Check for block structure. Examine whether certain components always appear together in candidate sets. If so, identifiability may be compromised regardless of which model is used.

  3. 3.

    Perform sensitivity analysis. Fit models under both C1-C2-C3 and plausible relaxed assumptions. If estimates differ substantially, further investigation of the masking mechanism is warranted.

  4. 4.

    Use simulation to assess impact. Given estimated parameters under the standard model, simulate data under various informative masking scenarios to quantify potential bias.

  5. 5.

    Report uncertainty appropriately. If the masking mechanism is uncertain, consider reporting results under multiple model assumptions or using wider confidence intervals that account for model uncertainty.

6.3 Limitations

Our analysis has several limitations:

  1. 1.

    Exponential focus. While we derive general likelihood results, the Fisher information analysis focuses on exponential components. Extension to Weibull or other distributions introduces additional complexity due to time-varying hazards.

  2. 2.

    Known masking probabilities. Our relaxed models assume masking probabilities are known. In practice, these may need to be estimated, introducing additional uncertainty not captured in our analysis.

  3. 3.

    Independence assumption. We assume masking for different observations is independent. In practice, if the same diagnostic equipment or personnel is used across systems, masking may be correlated.

  4. 4.

    Parametric masking models. Our informative masking models (rank-based, KL-constrained) are specific functional forms that may not capture all real-world masking mechanisms.

  5. 5.

    Simulation scope. The simulation studies cover a limited range of configurations. Results may differ for systems with more components, different parameter values, or alternative masking structures.

6.4 Future Directions

Several extensions would strengthen this work:

  1. 1.

    Weibull extension. Derive Fisher information and conduct simulations for Weibull components, where shape parameters introduce additional estimation challenges.

  2. 2.

    Semiparametric methods. Develop estimation approaches that avoid fully specifying the masking mechanism, perhaps using nonparametric or empirical likelihood methods.

  3. 3.

    Model selection. Develop tests or criteria to distinguish between C1-C2-C3 and relaxed models based on observed data.

  4. 4.

    Bayesian extensions. Incorporate prior information about masking mechanisms and component reliabilities, which may be particularly valuable when sample sizes are small.

  5. 5.

    Sequential estimation. For systems observed over time, develop methods that update masking probability estimates as data accumulates.

  6. 6.

    R package documentation. Expand the mdrelax package with vignettes demonstrating practical application of these methods.

7 Conclusion

We have developed a theoretical framework for likelihood-based inference in series systems with masked failure data when the traditional conditions C2 (non-informative masking) and C3 (parameter-independent masking) are relaxed. Our main contributions are:

  1. 1.

    Likelihood derivations. We established the form of the likelihood under various relaxation scenarios, showing that the masking probabilities act as weights on component hazard contributions when C2 is violated.

  2. 2.

    Practical masking models. We introduced rank-based informative masking and KL-divergence constrained models that provide interpretable parameterizations of non-standard masking.

  3. 3.

    Identifiability results. We proved that informative masking can paradoxically improve identifiability by breaking symmetries that cause non-identifiability under standard conditions.

  4. 4.

    Fisher information analysis. We derived closed-form expressions for the Fisher information matrix under informative masking for exponential series systems, enabling efficiency comparisons.

  5. 5.

    Misspecification analysis. We characterized the bias that arises from incorrectly assuming C2 when masking is actually informative, providing guidance on when relaxed models are necessary.

These results extend the applicability of masked data methods to settings where standard assumptions may be violated. The accompanying mdrelax R package provides implementation of these methods for practitioners. Directions for future research are discussed in SectionΒ 6.4.

References

  • [1] M. Agustin (2011) Systems in series. In Wiley Encyclopedia of Operations Research and Management Science, External Links: Document Cited by: Β§2.6.
  • [2] R. H. Byrd, P. Lu, J. Nocedal, and C. Zhu (1995) A limited memory algorithm for bound constrained optimization. SIAM Journal on Scientific Computing 16 (5), pp.Β 1190–1208. Cited by: Β§B.2.
  • [3] D. R. Cox (1972) Regression models and life-tables. Journal of the Royal Statistical Society: Series B (Methodological) 34 (2), pp.Β 187–202. External Links: Document Cited by: Β§2.6.
  • [4] H. Guo, P. Niu, and F. Szidarovszky (2013) Estimating component reliabilities from incomplete system failure data. In Proceedings of the Annual Reliability and Maintainability Symposium (RAMS), pp.Β 1–6. External Links: Document Cited by: Β§2.4, Β§2.6.
  • [5] J. P. Klein and M. L. Moeschberger (2005) Survival analysis: techniques for censored and truncated data. 2nd edition, Springer Science & Business Media. Cited by: Β§2.6.
  • [6] A. Towell (2023) Reliability estimation in series systems: maximum likelihood techniques for right-censored and masked failure data. Note: Master’s thesis. Available: https://github.com/queelius/reliability-estimation-in-series-systems External Links: Link Cited by: Β§1, Β§2.1.
  • [7] J. S. Usher and T. J. Hodgson (1988) Maximum likelihood analysis of component reliability using masked system life-test data. IEEE Transactions on Reliability 37 (5), pp.Β 550–555. External Links: Document Cited by: Β§1, Β§2.4, Β§2.6.
  • [8] J. S. Usher, D. K. J. Lin, and F. M. Guess (1993) Exact maximum likelihood estimation using masked system data. IEEE Transactions on Reliability 42 (4), pp.Β 631–635. External Links: Document Cited by: Β§2.4, Β§2.6.

Appendix A Proofs and Derivations

A.1 Score Function Under Informative Masking

For completeness, we provide the full derivation of the score function under the Bernoulli informative masking model for exponential series systems.

Let the log-likelihood contribution from an uncensored observation be:

β„“i⁒(𝝀)=βˆ’siβ’βˆ‘j=1mΞ»j+log⁑(βˆ‘k∈ciΞ»k⁒πk,ci). (62)

The partial derivative with respect to Ξ»j is:

βˆ‚β„“iβˆ‚Ξ»j =βˆ’si+βˆ‚βˆ‚Ξ»j⁒log⁑(βˆ‘k∈ciΞ»k⁒πk,ci) (63)
=βˆ’si+Ο€j,ci⁒𝟏j∈ciβˆ‘k∈ciΞ»k⁒πk,ci. (64)

The total score is:

βˆ‚β„“βˆ‚Ξ»j=βˆ‘i=1nβˆ‚β„“iβˆ‚Ξ»j=βˆ’βˆ‘i=1nsi+βˆ‘i:Ξ΄i=1Ο€j,ci⁒𝟏j∈ciβˆ‘k∈ciΞ»k⁒πk,ci. (65)

Setting this to zero and solving gives the MLE equations under informative masking.

A.2 Hessian Matrix Derivation

The second partial derivatives are:

βˆ‚2β„“iβˆ‚Ξ»jβ’βˆ‚Ξ»β„“ =βˆ‚βˆ‚Ξ»β„“β’[Ο€j,ci⁒𝟏j∈ciβˆ‘k∈ciΞ»k⁒πk,ci] (66)
=βˆ’Ο€j,ci⁒πℓ,ci⁒𝟏j∈ciβ’πŸβ„“βˆˆci(βˆ‘k∈ciΞ»k⁒πk,ci)2. (67)

The observed Fisher information matrix is the negative Hessian:

ℐj⁒ℓ⁒(𝝀)=βˆ’βˆ‚2β„“βˆ‚Ξ»jβ’βˆ‚Ξ»β„“=βˆ‘i:Ξ΄i=1Ο€j,ci⁒πℓ,ci⁒𝟏j∈ciβ’πŸβ„“βˆˆci(βˆ‘k∈ciΞ»k⁒πk,ci)2. (68)

A.3 Expected Fisher Information

The expected FIM requires integrating over the distribution of candidate sets. For the exponential series system under C1-C2-C3 with Bernoulli masking (each non-failed component in candidate set with probability p), the expected Fisher information per observation is:

𝔼⁒[ℐj⁒k]=𝔼⁒[𝟏j∈C⁒𝟏k∈C(βˆ‘β„“βˆˆCΞ»β„“)2], (69)

where the expectation is over both K (failed component) and C (candidate set).

This can be expanded as:

𝔼⁒[ℐj⁒k] =βˆ‘k0=1mPr⁒{K=k0}⁒𝔼⁒[𝟏j∈C⁒𝟏k∈C(βˆ‘β„“βˆˆCΞ»β„“)2|K=k0]. (70)

Under C1, the failed component k0 is always in C. The expectation over candidate sets involves summing over all possible Cβˆ‹k0 weighted by their probabilities under the Bernoulli model:

Pr⁒{C=c∣K=k0}=∏j∈cβˆ–{k0}p⁒∏jβˆ‰c(1βˆ’p). (71)

Closed-form evaluation of this expectation is generally intractable due to the sum in the denominator. Monte Carlo estimation or numerical integration is typically required.

Appendix B Implementation Details

B.1 R Package Functions

The theoretical framework developed in this paper is implemented in the mdrelax R package. Key functions include:

  • β€’

    md_bernoulli_cand_C1_C2_C3(): Generates candidate set probabilities under the standard Bernoulli model satisfying C1-C2-C3.

  • β€’

    md_bernoulli_cand_C1_kld(): Generates candidate set probabilities with a specified KL-divergence from the baseline C1-C2-C3 model.

  • β€’

    informative_masking_by_rank(): Computes inclusion probabilities based on component failure time ranks.

  • β€’

    md_cand_sampler(): Samples candidate sets from probability vectors.

  • β€’

    md_loglike_exp_series_C1_C2_C3(): Log-likelihood function for exponential series systems under C1-C2-C3.

  • β€’

    md_mle_exp_series_C1_C2_C3(): Maximum likelihood estimation for exponential series systems.

  • β€’

    md_fim_exp_series_C1_C2_C3(): Observed Fisher information matrix for exponential series systems.

  • β€’

    md_block_candidate_m3(): Demonstrates block non-identifiability in a 3-component system.

B.2 Optimization Details

MLE is computed using the L-BFGS-B algorithm [2] with analytically computed gradients. The optimization is initialized using a method-of-moments estimator based on the total system hazard:

Ξ›^init=nuncensoredβˆ‘i=1nsi,Ξ»^j,init=Ξ›^initm. (72)

For challenging optimization landscapes, simulated annealing can be used to find a good starting point before local optimization.

Appendix C Additional Simulation Results

Additional simulation results supplement the findings in SectionΒ 5. The complete results are available in the R package’s simulation directory.

  1. 1.

    Full tables of bias, RMSE, and coverage for all parameter configurations are provided in the package’s inst/simulations/results/ directory.

  2. 2.

    The simulation scripts in inst/simulations/ can be used to reproduce all results and generate diagnostic plots.

  3. 3.

    Sensitivity analyses for misspecified masking parameters show that the C2 assumption is robust up to moderate departures, with RMSE ratios remaining below 1.10 across tested configurations.

  4. 4.

    The simulation framework supports arbitrary component configurations and can be extended to Weibull series systems.