CONFIDENCE BOUNDS FOR THE RELIABILITY OF A SYSTEM
FROM SUBSYSTEM DATA
O. Hryniewicz
Systems Research Institute, Warsaw, Poland e-mail: Olgierd.Hryniewicz@ibspan.waw.pl
ABSTRACT
The paper is concerned with the construction of lower bounds for the reliability of a system when statistical data come from independent tests of its elements. The overview of results known from literature and obtained under the assumption that elements in a system are independent is given. It has been demonstrated using a Monte Carlo experiment that in case when these elements are dependent and when their dependence is described by Clayton and Gumbel copulas these confidence bounds are not satisfactory. New simple bounds have been proposed which in some practical cases perform better than the classical ones.
1 INTRODUCTION
Reliability indices of complex systems can be estimated from the results of lifetime tests. When a system is treated as a one entity we can distinguish two different types of reliability tests. In the first one, we observe consecutive failures of a system, and after each of them a failed system is completely renewed. In such a case, random times between consecutive failures are described by random variables having independent and identical probability distributions. If this assumption is true, we can estimate a required reliability characteristic using a sample of observed lifetimes. In the second case, we have to observe several identical systems working in the same conditions. Times to first failures of these systems constitute a sample which may be used for the estimation of the considered reliability characteristic. In both cases, however, we need to have either sufficiently long time of test or sufficiently large number of observed systems. Both these requirements are seldom met in practice. Thus, this method of the reliability estimation is rarely used in practice despite the fact that from a statistical point of view the required estimators are obtained in the simplest possible way. Moreover, in such a case we do not profit from the information about the structure of the considered system, and from the knowledge of times to failure of its elements.
In practice we are frequently faced with a different problem: how to evaluate reliability characteristics of a system on its design stage. There exist many methods for the prediction of reliability using available statistical data. In this paper we consider the simplest one, when we can utilize the results of reliability tests of system's elements performed in presumably the same conditions as the conditions of work of the designed complex system.
Research studies on statistical methods aimed at the estimation of system's reliability using the results of reliability tests of its elements were initiated independently in the 1950s in the United States and the Soviet Union, where they were performed by prominent mathematicians and statisticians. Some strong mathematical results were obtained, and these results can be used for both point and interval estimation of system's reliability using the data obtained for its elements or subsystems. In this paper we will focus our attention on the interval estimation. The reason for the
importance of the results of this type stems from practice. Usually we can use scarce reliability data, and thus the obtained point estimators are not very precise. Therefore, we need to know some lower bounds for the predicted reliability characteristics.
Preliminary analysis of the theoretical results shows undoubtedly that even in the cases of simple systems exact analytical methods require utilization of complex mathematical tools such as nonlinear mathematical programming. On the other hand, interesting approximate results, obtained mainly by American researchers, can be used in practice when a sufficiently large number of failures have been observed. For this reasons already in the 1980s the reliability theoreticians lost their interest in further research on those problems. However, the problem is still interesting for practitioners who need approximate, or even heuristic, methods which may be used for the prediction of reliability using existing statistical data.
The paper is partly based on the lecture for young scientists working in the area of reliability. Therefore the purpose of this paper is two-fold. In first six sections we give a short overview of different methods for the construction of confidence intervals for the reliability of systems. In all these methods it has been assumed that the elements of a system are independent. In the last section of the paper we present new heuristically designed bounds which seem to be robust to deviations from this assumption in certain practical cases.
2 GENERAL METHODOLOGY FOR THE EVALUATION OF SYSTEM'S RELIABILITY
Evaluation of reliability of complex systems became the subject of intensive theoretical investigation in the beginning of 1960s. Fundamental results were summarized in the famous book (Barlow & Proschan 1965). In the developed mathematical models we assume that both the system as a whole, and system's elements at any time instant t>0 are either in the state of functioning (or failure-free state), when the random variable X(t) describing the reliability state adopts the value 1, or in the state of failure, when this random variable adopts the value 0. When the considered system consists of m elements, then its reliability state is described by the random vector X=(Xi,X2,.. .,Xm), and the probability of the observation of any reliability state is given by
m
p(x)=nA' (1 - pi )1-X' (1)
'=1
where
p, = P(X' = 1) = E(Xr),' = 1,...,n .
(2)
In the above formulae we have omitted time t assuming that in case of specific calculations it adopts the same value for all components of the random vector.
Reliability state of the whole system depends on the states of all individual system's elements. Denote by Q the set of all 2m possible states of system's elements. We can divide this set into two exclusive subsets: the subset of all functioning states of the system G, and the system of all failure states of this system G (G u G = Q). The function
f(x)={' X 6 G
w [0 X g G (3)
is called the structure function, and it describes the relation between reliability state of the whole system and reliability states of its elements. The effective construction of this function is the subject of numerous research works. Particular results may be found in all classical textbooks on reliability such as (Barlow & Proschan 1965, 1975).
Probability that the considered system is in the failure-free state depends on the vector p=(pi,p2,.. .,pm) that describes the probabilities of failure-free functioning of system's elements, and system's reliability structure function. It is given by the function called the reliability function which is given by the following formula
R(p) = P(X e G ) = £№)] =
m
Z^(x)npX (i - p)i-X .
XeG i=1
(4)
Below, we present the respective formulae for the reliability structures which are most frequently met in practice.
a) In case of a system with series reliability structures which consists of m groups of identical ni, i=1,...,m elements we have:
m
R(p ) = n Pn (5)
i=1
b) For the system with a parallel reliability structure which consists of m elements the respective formula is given by
mm
r(p )=l-n(i - pi )=i-n q.
i=1 i=1
(6)
c) In case of a series-parallel reliability system which consists of m connected in series groups, where each of these groups consists of ni connected in parallel identical elements, the reliability function is given by the formula:
R(p ) = n &-(i - Pi)"' ].
(7)
d) For a parallel-series system consisting of m connected in parallel groups, where each of these groups consists of ni identical elements connected in series, the reliability function is given by the formula:
m
R(p) = i-n
i=1
1 -n Pj
_ 1=1
(8)
In formulae (5) - (8) piy denotes the probability that the j-th element in the i-th subsystem is in a failure-free state.
The systems with structures described above belong to a more general class of systems called coherent systems, or systems with monotonic structure. The system has monotonic structure if
^(x)>^(Y ) (9)
holds when Xi>Yi, i=1,...,m, and when
^(0)= 0 ,....(l) = 1, (10) with 0=(0,...,0) and 1=(1,...,1). For systems with a monotonic structure the reliability function can be always computed. However, for large and complex systems this can be a hard computational task.
In order to compute the probability that the system is in the failure-free state we need to know the estimates of the elements of the vector p. These estimates can be obtained from the results of reliability tests. We assume that for each of system's elements we have the results of independent reliability tests. From these tests we obtain the vector of estimates p =(p1 , pi ,..., pm ). The estimators pi are unbiased estimators of unknown probabilities pi only in certain particular cases.
However, in the majority of practical cases, when we apply the maximum likelihood method of estimation, these estimators are asymptotically unbiased, but in practice the conditions of asymptotics usually do not hold due to the limited number of the pertinent statistical data. The knowledge of estimates p =(p1 , p1 ,..., pm) allows for simple estimation of the reliability R(p). In such a case we apply the method of substitution. We substitute in (4) unknown probabilities p with their estimates p . The estimator of the reliability of the whole system R(p ) is unbiased only in a particular case of systems with a series reliability structure and unbiased estimators ofpi. In all other cases R(p ) is biased or at best asymptotically unbiased. Therefore, in practical situations the estimates of the system's reliability are very uncertain, and we need to have methods for the computation of lower bounds for its possible value. Such bounds may be obtained by the calculation of confidence intervals for R(p).
Let us now consider a system consisting of element s of m different types. Suppose that the reliability of the element of the i-th type, i=1,...,m, is a certain function of a parameter 6i whose value is unknown. Thus, we may assume that the reliability of the whole system is described by a function R(9) which depends on the vector 9=(6>i, &2,..., dm) of parameters describing the reliability of system's elements. Moreover, we assume that the information from reliability tests of system's elements is denoted by xi, i=1,...,m. Thus, the results of the tests are described by a vector x=(x1, x2,..., xm). We have to note that the values of 6i and xi only in special cases are represented by single numbers. In a general case they are represented by vectors of numbers. The interval (R,R), where R = R(x) and R = R (x) is the two-sided confidence interval, calculated on the confidence level y for the unknown value of R(9) if the following condition is fulfilled
In an analogical way we can define one-sided lower and upper confidence intervals for the reliability function R(9). In the sections which follow we present methods for the calculation of such confidence intervals. In this presentation we use notation given in the book (Gnedenko et al. 1999).
3 CONFIDENCE INTERVALS FOR SYSTEM'S RELIABILITY IN THE CASE OF DISCRETE RELIABILITY DATA
Let us consider the problem of reliability estimation when the results of reliability tests of system's elements are available in a discrete form. Let us assume that the elements of all types are independently tested in exactly the same conditions as the work conditions of the considered system. In the simplest case we test samples of size Ni, i=1,...,m, for all m types of elements, and the duration of all tests is the same, and is equal to t. In this simplest case we assume that we know the reliability state of each tested element at the end of the test. Thus, we assume that we know the numbers of elements di, i=1,.,m, which failed during the test. The test result is described, therefore, by pairs of integer numbers (di,Ni), i=1,...,M In such a case we say that the reliability tests, also known as pass-fail tests, are performed according to a binomial scheme. In this simple case there exists an unbiased estimator of the reliability of a tested element given by a simple formula
Calculation of the confidence interval for the reliability pi is not simple. For a given confidence level y one can calculate the confidence interval using a so called fiducial approach. The respective formulae are known as the Clopper-Pearson formulae, and in the considered case of reliability
(11)
p = 1---, i = 1,...,m
The random number of the observed failures is thus described by the binomial distribution
(12)
(13)
estimation they have the form given in (Gnedenko et al. 1999). The lower bound p of the one-sided confidence interval for the reliability p is given as the solution of the following equation
£ if](1 - p) (pY-k = 1 - r , (14)
k=0 V k J
and the upper bound p of the one-sided confidence interval for the reliability p is given as the solution of the equation
iffy (1 - p)N-k = 1 - ' . (15)
k=0 V k J
In case of d=N we have p = 1, and when d=0 we have p = 0 . It is worth noticing that if we replace
1-y in (14) - (15) with 0,5<a<1 and 0,5<^<1, respectively, we can use these formulae for the calculation of a two-sided confidence interval for the reliabilityp on the confidence level equal to 1-
a-p.
When the probability of a failure is low, or when reliability is high, i.e. when the strong inequality qi=1-pi<<1, i=1,...,m holds, and when the number of tested elements Ni, i=1,...,m is large, the probability distribution of the number of failed elements di, i=1,...,m can be approximated by the Poisson distribution with the parameter A1=qiNi, and the probability mass function given by the formula
d *
p(d. = d*)= ^d-e-Ai,i = 1,. ,m (16)
V 1 l} d*!
This approximation is valid when in case of q ^ 0 and N ^ ro the condition Nq=const holds.
One-sided confidence intervals for the parameter A of the Poisson distribution can be found by solving the following equations
d-1 Aj
eA£ ^ = 1 - P (17)
j=0 j!
- d Aj
e A £ A = « (18)
j=0 j!
When d = 0 we have A = 0. For further calculation we can use the connection between the Poisson distribution and the special case of the gamma distribution, namely the chi-square distribution. The confidence intervals can be thus calculated from the formulae:
A = 2 x] (2d), (19)
A = 2 %L(2d + 2), (20)
where ^ (n) is the quantile of order y of the chi-square distribution with n degrees of freedom.
Similarly, as in the case of the binomial distribution, for 0,5<a<1 and 0,5<^<1 we can use (19) -(20) for the calculation of the two-sided confidence interval for the parameter A on the confidence level 1 - a - (3.
The Poisson distribution can be also used when the times to failure are described by the exponential distribution. When all elements failed during the test are replaced by new ones, and the duration of the test is equal to T, the observed number of failures is described by the Poisson distribution with the parameter A=XNT, where X is the failure (hazard) rate in the exponential distribution, and N is the number of simultaneously tested elements. Confidence intervals for the parameter A (and for the failure rate A) are in this case calculated from the formulae (19) - (20).
4 CONFIDENCE INTERVALS IN THE ABSENCE OF OBSERVED FAILURES
Contemporary technical systems are built of very reliable elements. For such elements we usually do not observe failures during reliability tests. In such a case, the point estimation of system's reliability is trivial, and is equal to 1. However, we are interested in the lower bound for this characteristic which may be interpreted as kind of guaranteed reliability. Suppose, that for each of the m types of elements the system is built of we test Ni, i=1,...,m, elements, and in every case the number of observed failures is di = 0, i=1,...,m. For such results tests the upper bound for the confidence interval is always equal to R = 1. On the other hand, it is possible to calculate the lower bound R of the confidence interval for the reliability of the considered system. In the book (Gnedenko et al. 1999), where results of many works were summarized, it has been shown that the computation of this bound is equivalent to solving the following optimization problem
R = minRip), (21)
peH o
where the set H0 contains all values of the vector p=(p1,p2, ...,pm) such that
m
n pN > 1 - y (22)
i=1
and
0 <pt< 1,i = 1,...,m . (23)
In many interesting cases there exist closed solutions to this optimization problem. In case of a series system such solution was given in (Mirnyi & Solovev 1964). They showed that the lower bound of the confidence interval for system's reliability is given by a simple formula
R = minp (24)
i -i
where p is the lower bund of the one-sided confidence interval, calculated according to the
Clopper-Pearson method (14). It is easy to show that this bound can be calculated from an equivalent formula
R = i1 - rt' (25)
where
N * = minNi. (26)
i
For systems with a more complicated structure very strong theoretical results were obtained in (Pavlov 1982) who considered systems with a convex cumulative risk function defined as follows
Rit) = e"H it). (27)
He has shown that for such systems
R = minR^,... ,1,p,1,... ,1) (28)
where
p^ =i1 -y)l1Ni,i = 1,.,m . (29)
The solutions of this problem for parallel, series-parallel, parallel-series, and &-out-of-n systems have been presented in the book (Gnedenko et al. 1999). For example, in the case of a system with a parallel reliability structure, consisting of n different elements the lower bound of the one-sided confidence interval for system's reliability is given by:
« t
R = 1 -n—^ (30)
- j=f t + Nj y ;
where t is the solution of the following equation
n f t ^
2 Njin 1+— =-m(1 - r) (31)
j=i I N J
In a particular case, when N1=.=Nn=N Tyoskin and Kurskiy obtained a simple analytic solution (see (Gnedenko et al. 1999))
r = 1 -1 -(1 - ^J. (32)
For systems with a more general coherent structure such simple solutions do not exist. However, in the book (Gnedenko et al. 1999) two boundaries for the lower bound of the confidence interval have been proposed. Consider the set of all minimal cuts of the system, and assume that the minimal cut with the smallest number of elements consists of b elements. Then, consider the set of all possible minimal paths. For this set consider its all possible subsets consisting of independent, i.e. having no common elements, paths. Let a be the number of such paths in the subset with the largest number of independent paths. Assume additionally, that for each type of system elements exactly N elements have been tested. The boundaries for the lower bound for the system's reliability are the given by
1 -1 -(1 - ^ j < r < 1 -1 -(1 - yyn j (33)
In a particular case of a = b we have
r = 1 -1 -(1 - yyn J (34)
The authors of (Gnedenko et al. 1999) notice, that this case is typical for many reliability structures such as lattice or radial structures which are typical for large network systems.
Another very interesting method for the calculation of the lower bound of the confidence interval for system's reliability was presented in (Gnedenko et al. 1999). Let us assume that the same vector of reliabilities p=(p1,p2,.,pm) is used for the calculation of reliability of two systems: the reliability R(p) of the considered complex system, and the reliability R (p) of a simple (e.g. series) auxillary system. For this auxillary system we must know the lower bound of the respective confidence interval R (p). In order to find the lower bound of the confidence interval for the reliability of the considered system we have to solve the following optimization problem:
R = minR(p) (35)
p
where the element of the vector p must fulfill the following constraint
m
n pi > R:, 0 < pi < 1,i = 1,...,m . (36)
i =1
The lower bound calculated in this way fulfills all the requirements for a lower bound of a confidence interval, but the length of such interval is usually not the shortest possible.
5 CONFIDENCE INTERVALS IN THE PRESENCE OF OBSERVED FAILURES
When failures are observed during reliability tests of system's elements the problem of building confidence intervals for the reliability of the whole system becomes much more complicated. Comprehensive information about available methods can be found in the fundamental book (Gnedenko et al. 1999). Below, we present only some basic results considered in this book and related literature.
Let us assume that the considered system consists of elements of m different types. For each of these types we test a sample of Ni elements, and for each sample we observe di >0, i=1,...,m failures. Let
S = RP!, p 2,-, pm J (37)
A
be the point estimator of system's reliability, where pi, i = 1,...,m are the estimators of the
reliability of systems elements calculated according to (12). Now, denote by d*= (d^, d2*,..., dm*) the vector of numbers of observed failures. Moreover, denote by S*=S(d*) the observed value of the estimator of system's reliability presented as the function of the vector d . The lower bound of the confidence interval for the system's reliability is now calculated from the formula
m f N ^
max E nl N pN di (1 - pt)di = 1 - r, (38)
pGA* S(d j^d') i=1 Id J
where maximum is calculated over the set AR of vectors (p1 ,p2,...,pm), such that
R(p,p2,...,pm) = R, 0 < pt < 1,1 = 1,.,m . (39)
The sum in (38) is calculated over all possible values of the vector d = (di , d2,..., dm) that fulfill the condition given for this sum in (38). In certain cases other formulation of this optimization problem is more suitable for computations. According to this formulation we denote by n(d)=n(d1,d2,.. ,,dm) a non-decreasing, with respect to all components, series of vectors. The first element of this series is the vector (0,0,...,0), and then we have the vectors of the type (0,..,0,1,0,...,0), etc. The lower bound of the confidence interval for system's reliability can be calculated from
R = minR(p1 p ,.. ,pm ), (40)
where minimum is taken over the set of all values of the vector (p1,p2, .,pm) such that
E( n fN W - p )di>1 - * (41)
n(d )<«(d*) i=1 V ai J V+1;
0 < pt < 1,i = 1,...,m
The optimization problem given by (40) - (41) was formulated first time in (Buehler 1957) where a system consisted of two elements was considered. This was the first result of the calculation of the confidence interval for system's reliability.
Let us now consider the series system consisted of m different elements. The optimization problem is now the following:
m
R = min n pi, (42)
i=1
where minimum is taken over all vectors (p1,p2, .,pm) such that
E n [NJ ViNi-di(1 - pi )di * 1 - y,0 < p. < 1,i = 1,-,m (43)
R(d )>R(d") i=1 V di J
The calculation of the lower bound of the confidence interval for system's reliability R can be simplified when the probabilities of failures are small, i.e. when the inequality qi=1-pi <<1, i'=1,...,m holds. In such a case we can assume that the number of failures is described by the Poisson distribution with the parameter Ai=qiNi, i=1,...,m. It has been shown in (Gnedenko et al. 1999) that in this case we have
R = e"f (44)
where
m ^ih.
1n j
f = max E A , (45)
and the maximum in (45) is taken over all vectors A = (Ai,A2,...,Am) such that
m f \dt ^
2 ne — > 1 - Y,^ > 0, i = 1,...,m (46)
R(d)>R(d*) i=1 V ! J
This practical result was obtained first time in (Bol'shev & Loginov 1966) for the case of equal values of N, and, independently, in (Pavlov 1973) and (Sudakov 1974) for any values of these numbers.
6 APPROXIMATE CONFIDENCE INTERVALS FOR SYSTEM'S RELIABILITY
Computation of exact bounds of confidence intervals for system's reliability requires, with only few exceptions, solving difficult optimization problems. Therefore, its practical applicability is somewhat limited unless specialized software is available. For this reason several authors, mainly American, have tried to obtain approximate, but relatively easy for computation, solutions. Different approximate solutions have been proposed in the papers like (Madansky 1965, Myhre & Saunders 1968, Easterling 1972, Mann 1974 a, b, Mann & Grubbs 1972, 1974). Comprehensive review of such results can be found in a well known book (Mann, Shaefer & Singpurwalla 1974). However, probably the most interesting from a practical point of view result was presented in one of the first textbooks on reliability (Lloyd & Lipow 1962). These authors presented a heuristic method, attributed to Lindstrom and Madden, for the calculation of the approximate confidence interval for the system with a series reliability structure. This method utilizes the concept of so called equivalent tests. To present this method we consider, following the book (Gnedenko et al. 1999), a system with a series-parallel structure which has the same elements in its parallel subsystems. Let R be the estimated value of the reliability function for the considered system, and Ni, i=1,...,m be the number of tested items for the element of the i-th type. The equivalent number of failures Di for the element of this type is then calculated from the equation
Rf 1,—,1,1 - D/,1,.. ,1j = R* (47)
At the next stage of the computation procedure, for each equivalent test (Ni,Di*) we calculate the lower bound of the confidence interval P.(n;,D*) by solving the equation
Bp (n. - d;,D+1)=1 - 7, (48)
where
p
J xa-1(1 - x )b-1 dx
Bp (a,b ) = ^--(49)
J xa-1(1 - x )b-1 dx
0
is the incomplete beta function whose values can be computed using an available numerical procedure. The lower bound of the confidence interval is now calculated from a simple formula
R = minR(1, — ,1,P. (n.,D*),1, — ,1). (50)
The Lindstrom-Madden method was proposed as an approximate heuristic method. However, it has been proved (see (Gnedenko et al. 1999) for additional information) that for many simple reliability structures it produces exact confidence intervals.
Another method which uses the concept of equivalent tests, and which can be used for the analysis of complex systems consisted of many simple subsystems, was proposed in (Martz & Duran 1985). In this method it is assumed that for each simple subsystem we are able to calculate the value
of its reliability estimator R, and the lower bound for the respective confidence interval Ri. Next, from a set of equations
1 - — = R (51)
M
and
Ri = P {Mt,rt) (52)
we calculate the parameters (Mi,ri) of the equivalent binomial reliability tests. In further analysis the considered subsystem is treated as a single element described by the equivalent test. Note, that for the application of this method it is not important how we have found the values of Ri and Ri.
7 SOME REMARKS ABOUT OTHER METHODS FOR THE CALCULATION OF CONFIDENCE INTERVALS FOR SYSTEM'S RELIABILITY
In the previous sections we have presented methods for the calculation of confidence intervals for system's reliability for the case of discrete reliability data from tests, i.e. when the numbers of tested elements and the numbers of observed failures are known. It is a well known fact that the knowledge of lifetime distributions combined with the knowledge of observed times to failures may increase the accuracy of reliability estimation. Moreover, this knowledge may be sufficient for the prediction of reliability at time instants other than the times of the performed reliability tests. Unfortunately, even in the simplest case of the exponential distribution of lifetimes the exact and practically applicable solutions are known only in few cases when lifetime tests are performed according to the type-II censoring scheme (a fixed number of observed failures). For example, (Lentner & Buehler 1963) considered the case of a series system with only two elements. Their result was generalized in an unpublished PhD thesis (El Mawaziny 1965) who proposed an iterative method for the calculation of the lower bound of the confidence interval for reliability of a series system consisted of m elements. Because of its complicated nature this algorithm has not been described in reliability textbooks. However, there exists a good approximation proposed in (Mann & Grubbs 1972), and in a simplified version in (Mann 1974b).
Consider the case when the lifetimes are exponentially distributed, and reliability tests provide type-II censored data. For each type of system elements we test a sample of ni items, and observe times tij of the first r>0, i=1,.. ,,m failures. The respective value of the total time on test z, is given by
ri
zt = E t>j+ n - r K, i=^ • • •, m (53)
j=1
Denote by zw the minimal value of zi, i=1,...,m. (Mann 1974b) has shown that the estimator of the hazard rate of the series system has approximately the expected value given by
M = t~ + — > (54)
¿=1 zi z(1)
and the variance given by
v = + (55)
i=1 zi z(1)
To approximate the optimum lower bound on series system reliability Rs (t) at confidence level (, using the Wilson-Hilferty transformation, one calculates
RS (t )=
exp
-1^
1/2 V
yyv'
1--^ + :
9^2 3^
(56)
where yr is the quanitile from the standardized normal distribution. For systems with more complex structures an interesting approach has been proposed in (Gnedenko et al. 1999). According to this approach first we have to calculate upper bounds for the hazard rates of system's elements using the following simple formula
X = ^^ ,i = 1, — ,m, (57)
. 2S. V '
where x2y (2r) is the quantile of the y order from the chi-square distribution with 2r degrees of
freedom. When we insert these lower bounds into a formula for the calculation of the system's reliability function instead of respective hazard rates, i.e. if we calculate
R = R(ll.I2, . . . ), (58)
the obtained value usually fulfills the requirements for a confidence interval. (Pavlov 1982) has shown that in case of />0,778 this approach allows to calculate confidence intervals for a broad class of reliability structures for lifetime distributions having non-decreasing (in time) hazard rates (i.e. for elements with the ageing property).
The general methodology for the calculation of confidence intervals for system's reliability was proposed in (Belyaev 1966, 1968). Other, but completely equivalent general method, was proposed in (Bol'shev & Loginov 1966). Below, we present the main results of Belyaev.
Suppose that we know the statistic S which can be used as a point estimator of system's reliability, i.e. s = R. Moreover, we assume that this statistic is a function of a vector of parameters 9 describing probability distributions of lifetimes of system's elements. Additionally, we assume that the probability distribution of this statistic is known, i.e. we know
F(t,0) = p (S < t). (59)
For a given value of the vector 9 we can now introduce two functions t1(9) and t2(9), such that
F (tj ,0 ) = a (60)
and
F (t2,0) = 1 - p. (61)
Now, let's denote by
ar ={0 : R(0) = R} (62)
the set of all values of the vector 9 for which the reliability function adopts a given value R. Next, introduce two functions
K1(R) = mintAe) (63)
0GAR
and
K2 (R ) = mint2 (e). (64)
0GAR
The lower and upper bounds of the confidence interval on the confidence level 1-a-P for system's reliability can be found by solving equations
K1(R) = S * (65)
and
t K2 (r ) = S *, (66)
where S is the observed value of the statistic S.
The described general methodology is based on the original methodology for the construction of confidence sets proposed in (Neyman 1935), and is valid for any type of reliability data, and any type of reliability structure. However, its practical applicability is limited only to rather simple cases.
8 APPROXIMATE LOWER BOUNDS FOR SYSTEM'S RELIABILITY BASED ON MINIMUM VALUES OF THE RELIABILITY OF SYSTEM'S ELEMENTS
Computation of optimal (i.e. the shortest) and exact confidence intervals is, with a few exceptions, a very difficult task. Moreover, in all published results it is assumed that the elements in a system are mutually independent. Additional problems arise from a fact that confidence intervals used for the description of test results may be conservative, as in the case of intervals based on the Clopper-Pearson formula. In this section we present approximate bounds for system's reliability which, under certain conditions, may replace lower bounds of confidence intervals.
In order to investigate the robustness of the confidence intervals for system's reliability against the departure from the assumption of independence of system's elements let us introduce the notion of a copula. According to a famous theorem of Sklar (see e.g. (Nelsen 2006)) any two-dimensional probability distribution function H(x,y) with marginals F(x) and G(y) is represented using a function C called a copula in the following way:
H (x, y ) = C (F (x), G(y)) (67)
for all x,y e R. Conversely, for any distribution functions F and G and any copula C, the function H defined by (67) is a two-dimensional distribution function with marginals F and G. Moreover, if F and G are continuous, then the copula C is unique. In our investigation we will consider three types of copulas:
a) Clayton copula defined as
H(x, y)=[f(x) + G-e -1]- 1e, 0 > 0 (68)
b) Gumbel copula defined as
H (x,y ) = exp [(- In F(x)) + (- In G(y)) Y9 j 0 > 0, (69)
c) Fairlie-Gumbel-Morgenstern (FGM) copula defined as
H(x, y) = F(x)g(y)(1 + 0(1 - F(x))(1 - G(y))),-1 < 0 < 1 (70)
The Clayton and Gumbel copulas can be used for modeling a positive stochastic dependence. The FGM copula can be used for modeling both negative (6k1) and positive (6>1) dependence. The Clayton copula is especially interesting in reliability applications as it describes stronger dependence for smaller lifetimes than for larger ones. If this type of dependence exists the reliability of a series system with dependent elements is greater than in the case of independence. On the other hand, for a parallel system the reliability of a system with dependent elements is smaller.
In the majority practical cases the reliability of tested elements is high, and even for moderate sample sizes the number of observed failures is small. This suggests utilization of the result obtained for the case of zero-failure tests for the calculation of the lower bounds for reliability of a series system given by the expression (24). To analyze the properties of this approximation let us consider a two-element series system whose elements are equally reliable. We also assume that the sample sizes for both elements are the same. On Figure 1 we present the comparison of the values of our simple approximate bound with the bounds calculated for this system using a substitution method. For obtaining the presented results we performed a Monte Carlo simulation experiments, and in each of them we generated 500 000 test cases, Our approximate bound, plotted against the expected number of observed failures in a sample (for a probability of failure equal to 0,01), is represented by a continuous upper curve. The middle curve represents the bound calculated by the insertion into (5) the respective lower bound of the confidence intervals for the reliability of elements, calculated for
the same confidence level (y=0,9).The lower curve is a similar to the previous one, but calculated for the confidence level equal to -Jy , as it is suggested in statistical literature.
Then, we calculated the coverage probability of the considered confidence intervals. The results of the comparison are presented on Figure 2 for our approximate bound, and the bound represented by a middle curve on Figure 1.
As we can see, our simple bound fulfills requirements for a confidence interval not only for zero-failure reliability tests, but for all tests with the expected number of failures not greater than 1,95. The classical and much wider confidence intervals have the probability of coverage close to 1, i.e. much greater than the designed value of 0,9.
Figure 1. Lower bounds for a series system
Figure 2. Coverage probabilities for a series system in case of independence
Now, let us consider the case when the elements of the system are dependent. On Figure 3 we show the coverage probability when this dependence is described by the Clayton copula with
dependence parameter 0=2, and the Gumbel copula, with dependence parameter 0=2. For this value of the parameter the Kendall measure of dependence rfor both copulas is equal to 0,5. It means that the dependence is positive and fairly strong.
The coverage probability in the case of the Clayton copula (solid line) is greater than the designed value for tests with the expected value of observed failures greater than 5. However, in the case of the dependence described by the Gumbel copula (dashed line) this feature is guaranteed only for this value not greater than 2. It shows, how the type of dependence influences the results despite the fact that the popular measure of dependence, such as Kendall r in both cases gives exactly the same value.
Figure 3. Coverage probabilities for a series system in case of dependence
Now, let us consider the case of the system with elements connected in parallel. For such systems a simple for computation bound which is similar to that for a series system does not exist. Instead we propose the following approximation
R = 1 - min q , (71)
i
where qi is the upper bound of the confidence interval for the probability of failure. The lower bound calculated according to (71) is always smaller than the bound obtained by substitution of the probabilities of failures qi with their respective upper bounds qi. Thus, the coverage probability in
case of independent elements of the system, calculated according to (71), is always greater than the respective confidence level. It can be seen at Figure 4, where this probability is always equal to 1. (Note that the coverage probability in case of the bound obtained by substitution is also much greater than the confidence level which is equal to 0,9).
The situation changes dramatically when the elements of the system are dependent, and when their dependence is described either by the Clayton copula or by the Gumbel copula. On Figure 5 we present the coverage probabilities in such cases when the confidence intervals are calculated using the substitution method.
Figure 4. Coverage probabilities for a parallel system in case of independence
Figure 5. Coverage probabilities for a parallel system in case of dependence
The coverage probabilities (the left-most curve for the Clayton copula, and the curve next to it for the Gumbel copula) show dramatically that the confidence intervals obtained by substitution under the assumption of independence are too narrow. On the other hand, the interval calculated according to (71) has the coverage probability (depicted by a dashed curve for the Clayton copula, and equal to one for the Gumbel copula) greater than the confidence level.
9 CONCLUSIONS
Many prominent authors, mainly from USA and the Soviet Union, contributed to the problem of computing the lower confidence bounds for system's reliability using the data from tests of separate elements or subsystems. The proposed exact bounds are usually difficult to compute. Good approximations exist, but they are usually obtained under the assumption that failures of all elements or subsystems are observed during the tests. In the paper we have shown using Monte Carlo simulation that in case when elements working together in a system are dependent these bounds are inaccurate or even useless, as it is the case of parallel (redundant) systems. In the paper, we have proposed very simple bounds characterized by satisfactory performance, at least for highly reliable system elements, which are robust against the presence of positive dependence of the elements of a system.
REFERENCES
Barlow, R. E. & Proschan, F. (1965). Mathematical Theory of Reliability. John Wiley, New York.
Barlow, R. E. & Proschan, F. (1975). Statistical Theory of Reliability and Life Testing. Probability Models. Holt Rinehart and Winston, Inc., New York.
Belyaev, Y. K. (1966). Confidence intervals for the functions of many unknown parameters. Proc. of the Sov. Acad. Sci, 169: 755-758 (in Russian)
Belyaev, Y. K. (1968). On simple methods of Confidence Limits Construction. Eng. Cybernetics, No.5: 96-101 (in Russian, exists English translation of this journal)
Bol'shev, L.N. Loginov E.A. (1966). Interval estimates under noisy parameters. Theory of Prob. and its Appl. v.11, No.1: 94-107 (in Russian, exists English translation of this journal)
Buehler R.J. (1957), Confidence Intervals for the Product of two Binomial Parameters. Journal of the American Statistical Association, 52: 482-493.
Easterling R.G. (1972), Approximate Confidence Limits for the System Reliability. Journal of the American Statistical Association, 67: 220-222.
El Mawaziny, A.H. (1965), Chi-square distribution theory with applications to reliability problems. PhD Thesis, Iowa State Univ., Ames, Iowa.
Gnedenko B., Pavlov I., Ushakov I. & Chakravarty S. (1999), Statistical Reliability Engineering. John Wiley, New York.
Lentner, M.M. & Buehler, R.J. (1963), Some inferences about gamma parameters with an application to a reliability problem. Journal of the Amer. Stat. Assoc. 58: 670-677.
Lloyd D.K. & Lipow M. (1962), Reliability Management. Methods and Mathematics. Prentice-Hall, Englewood Cliffs, NJ.
Madansky A. (1965), Approximate Confidence Limits for the Reliability of Series and Parallel Systems.
Technometrics, 7: 493-503.
Mann N.R. (1974a), Approximate Optimum Confidence Bounds on Series and Series-Parallel System Reliability for Systems with Binomial Subsystem Data. IEEE Transactions on Reliability, R-23: 295-304.
Mann N.R. (1974b), simplified expressions for obtaining approximately optimum system-reliability confidence bounds from exponential subsystem data. Journal of the Amer. Stat. Assoc., 69: 492-495
Mann N.R. & Grubbs F.E. (1972), Approximately Optimum Confidence Bounds on Series-System Reliability for Exponential Time to Fail Data. Biometrika, 50: 191-204.
Mann N.R. & Grubbs F.E. (1974), Approximately Optimum Confidence Bounds for System Reliability Based on Component Test Plan. Technometrics, 16: 335-347
Mann N.R., Schaefer R.E. & Singpurwalla N.D. (1974), Methods of Statistical Analysis of Reliability and Life Data. John Wiley, New York.
Martz H.F. & Duran I.S. (1985), A Comparison of Three Methods for Calculating Lower Confidence Limits on System Reliability Using Binomial Component Data. IEEE Transactions on Reliability, R-34: 113-120.
Mirnyi, R.A. & Solovyev, A.D. (1964), Estimation of system reliability on the basis on its units tests. In: Cybernetics in Service for Communism, vol.2, Energiya, Moscow: 213-218.
Myhre J. & Saunders S.C. (1968), Comparison of Two Methods of Obtaining Approximate Confidence Intervals for System Reliability. Technometrics, 10: 37-49.
Nelsen, R.B. (2006) An Introduction to Copulas (2nd edition), Springer, New York.
Neyman J. (1935), On the Problem of Confidence Intervals. Annals of Mathematical Statistics, 6: 111-116.
Pavlov, I.V. (1973), Confidence limits for system reliability on the basis on its components testing. Eng. Cybernetics, No.1: 52-61 (in Russian, exists English translation of this journal).
Pavlov, I.V. (1982), Statistical methods of reliability estimation by tests results. Radio i Svyaz, Moscow (in Russian).
Sudakov, R.S. (1974), About interval estimation of reliability of series system. Eng. Cybernetics, No.3: 86-94 (in Russian, exists English translation of this journal).