Научная статья на тему 'Reliability-directed computer-aided design system'

Reliability-directed computer-aided design system Текст научной статьи по специальности «Компьютерные и информационные науки»

CC BY
84
15
i Надоели баннеры? Вы всегда можете отключить рекламу.

Аннотация научной статьи по компьютерным и информационным наукам, автор научной работы — Oleg Abramov

A theoretical apporoach and applied techniques for designing analogus engineering devices and systems with due account of random variations in system parametrs and reliability specifications are considered. For solving this problem a socalled operational/parametric approach is used. This approach is based on the computer-aided simulation of system capability and availability, parameter deviations and techniques of optimal parametric synthesis in terms of reliability criteria. Special attention is paid to algorithms that reduce the labour content of parameters optimisation problems. For seeking a numerical solution of the parametric design problem a computer-aided reliability-oriented design system is proposed.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «Reliability-directed computer-aided design system»

RELIABILITY-DIRECTED COMPUTER-AIDED DESIGN SYSTEM

Oleg Abramov

(Vladivostok, Russia)

Abstract.

A theoretical apporoach and applied techniques for designing analogus engineering devices and systems with due account of random variations in system parametrs and reliability specifications are considered. For solving this problem a so-called operational/parametric approach is used. This approach is based on the computer-aided simulation of system capability and availability, parameter deviations and techniques of optimal parametric synthesis in terms of reliability criteria. Special attention is paid to algorithms that reduce the labour content of parameters optimisation problems. For seeking a numerical solution of the parametric design problem a computer-aided reliability-oriented design system is proposed.

1. Introduction

The design process consists of group of particular tasks such as analysis or synthesis one. Any technical object ^^ can be stated as a pair W=<X,S>, whereX=(x1,...,xn) are parameters and S is the object structure. We can consider a function ®=C(W,Q) as a criteria of object's operation quality (here Q is the operation conditions).

The synthesis problem can be formulated as an optimization task

®(W, Q) ^ extr ^ WQ, W efi Q

vr t^w

where Q is the possible set of object variations and wQ is the optimal object under the conditions Q. The

analysis task can be divided to the parametric and structural ones

®(X, S, Q) ^ extr extr ^ XQ, SQ ,

^ W eQ XW eQ s Q Q

where QX is the set of possible parameters and QS is the set of possible structures.

Therefore, the synthesis of engineering systems consists of two basic parts: developing of structure (structural

synthesis) and internal parameter values choosing (parametric synthesis).

The studies have shown that the engineering system parameters are subject to random variations and the variations may be considered as non-stationary stochastic processes. The conventional methods for choosing parameters (for parametric synthesis) generally do not take account of parameter production and field deviations of parameters from their design values. As a result, the engineering systems designed in such a manner are not optimal in the sense of their gradual failure reliability.

This paper proposes the approach and some algorithms for seeking a numerical solution of the parametric optimization problem with a stochastic (reliability) criterion, which has made possible to implement and extend the concepts discussed in [1], [2], [3].

On the basis of the proposed methods and algorithms a computer-aided reliability-oriented design system called CARD has been developed. The CARD system builds mathematical models and calculates ratings of component parameters so that achieve the highest precision, acceptability (manufacturing yield) or reliability of engineering systems under design.

2. Parametric synthesis problem

Let S(x) be a system which depends on a set of n parameters x=(x1,..., xn). We say that system S(x) is acceptable if Y(x) satisfy the conditions (1):

a < Y(x) < b (1)

where Y, a and b are m--vectors of system responses(output parameters) and their specifications, e.g. Y1 (x) - average power, Y2(x) -delay, Y3(x) - gain. The inequalities (1) define a region D in the space of design parameters

D ={x| a < Y(x) < b} (2)

m

D is called the tolerance margin domain (region of acceptability) for S.

Fig.1 illustrates such a region. The values given for a and b are the specifications for the system.

Pg (x-Xr) -probability density function

Fig. 1. Region of acceptability D defined by system response functions.

As it was mentioned above, the engineering system parameters are subject to random variations (aging, wear, temperature variations) and the variations may be considered as stochastic processes:

^ X(t) = {x 1(0,..., Xn (t)}.

In general the optimal parametric design (synthesis) problem can be stated as follows. Given the characteristics of random processes X(t) of system parameters variations, a region of admissible deviation-D and a service time T, find such a deterministic vector of parameter ratings (nominals) X r =(x1r ,...,xnr) that the probability

P (xr, T) = Pr {[[(t) - x„,..., Xn (t) - Xnr ] e D, Vt e [0, T]} be maximized.

Any optimization technique requires, first, a method of objective function calculation and, second, an extremum searching method which allows to find a solution with a minimum cost.

3. Reliability estimation

The proposed method of reliability estimation, called "the method of critical sections", is based on the following fundamental ideas.

As known, any random process can be presented as a set of random values, or as an ensemble of functions yra(t) in some functional space, where functions yra(t) depend on the parameter ra which characterizes the realization or the selective function of a random process. The entire set of this functions will define the output parameter degradation process.

The probability Prt(DY) of matching to given constraints Dy = {Y / a < Y < b} at the time t for any t-section of output parameters space

Pr, ( Dy ) = ] ft (y)dy,

where f(y) is a one-dimensional distribution density of a random process.

Let us select from the entire set of random processes a subset S^ of such realizations that their values will belong to the numeric ensemble D at the times t1, t2, ... t 1^. The value of the probability index corresponding to this subset can be defined as

b b

Pr (S,) = {...{ ft1,...,tM (y<1,..., ym)dyx...dy^, (3)

a ^a ^

where ft1 V is the joint distribution density of random values for t^,...,^- sections.

If the number of t-sections within [0;T] is large enough, the Pr(Sy) may be considered as a probability of system non-fault operationality.

Usage of equation (3) is very difficult, but it can be simplified by applying some constraints to the selective functions. For example, for smooth processes we can suppose that in order to determine acceptability of any realization it will suffice to determine acceptability of this realization only for few t-sections (called critical sections). In usual practice the number of such critical sections is not more than three for real engineering devices.

Let the random parameter changing process Y(t) be monotonous. In this case in order to any it realization y(t) lie in the region of acceptable changes [a,b] during the time interval T, it is necessary and enough that they lie in this region in the boundary (critical) sections t=0

m

and t=T. The conditions of lying of Y(t) in the region of acceptable values during time T will be as follows:

a < Y(0) <b, a < Y(T) <b,

where Y(0) and Y(T) are random values appeared on the

corresponded t-sections of process Y(t).

The probability of lying of Y(t) inside [a, b] during given time can be stated as

Pr(T) = Pr{[ a <Y(0)<b]n [a<Y(T)<b]}

If the tendency of parameter changing is known, then using some allowances we can write

f f dy

Pr(T) = j fo (y)dy - j It (y)dy, for - >0,

Pr(T) = j It (y)dy - j fo (y)dy, for dy <0,

a b

where f0(y) and fT(y) are the probability density functions at the sections t=0 and t=Trespectively.

An important problem of system design is to choose the nominal design xr, so that the "yield" is maximized. The yield is the probability that the system will work, i.e., pass specifications (1) at the time t=0. Now suppose that a joint probability density function (PDF) for internal parameters p(x) is given which characterizes the random variations of the parameters at t=0. The designer usually has design control on the nominal xr of this distribution. To emphasize this dependence of the PDF on the nominal design, we write the probability density function as p(x -xr). In terms ofp(x -xr) the yield is

Pr(xr)= j p(x -xr)dx

(D - is the region of acceptability). Thus the design problem is to choose xr so that Pr(xr) is maximum.

The practical algorithm of the stochastic criterion calculation is based on the conventional Monte Carlo method.

At the beginning, the random vector of parameters is generated (this vector means random manufacturing device realization), and then the internal parameters degradation is simulated using degradation model. For example, parameters variations can be approximated as follows

X(t)= £ xruk (t)

where xk is a random variable; {uk (t)}mk=0 are continues deterministic functions of time.

The Monte Carlo method approximates Pr(xr,T) by the ratio of number of acceptable realizations (falling in region D)-Na to the total number of trials - N.

P =

N„

/N

Unfortunately, often the region D is unknown. It is given only implicitly through system's equations and the systems response functions. If we do not know the region D, then a Monte Carlo evaluation of probability Pr(xr,T) at particular nominal value xr requires N system analyses for each trial set of parameter xr. Typically, hundreds of trials are required to obtain a reasonable estimate for Pr(xr,T).

Optimization requires the evaluation of our probability Pr(xr,T) for many different values of the nominal values of the parameters xr. Therefore to make practical the use of Monte Carlo techniques in statistical system design, it is necessary to reduce the number of system analysis required during optimization.

As a solution, the following two-steps technique and the corresponding algorithms can be used for practical reliability optimization.

4. Optimization techniques

The first step consists in replacing the original stochastic criterion with a certain deterministic one, allowing nearby optimum solutions to be obtained. The two such objective functions are possible. One of them is a so-called a "minimal serviceability reserve" that can be presented in the general form:

F (x) = min

i= 1 ,m

at - Yt (x)

-1

where Yi(x) - the i-th output value, ai - the i-th constraint (Y(x) < a) and wi - the i-th weight coefficient.

From this, we have a following optimization problem: x r = argmax F (x).

xeD

It means that such a nominal point should be found that will have the largest distance from the acceptability region margins.

An other method which can be used for the reliability optimization is so-called "equal densities method". This

b

a

w

k=0

method is of combined type, that uses statistical data and a deterministic optimization technique.

At the first step we should estimate distribution density function (DDF) for output parameter. As can be shown

analytically, probability maximum will be achieved, if

DDF will be shifted such, that both lower and upper constraints will cut equal densities on DDF [1]. Advantages of this approach are fairly clear, but there is one serious restriction - this method may be applied only in case of one output parameter.

Now the first design step is completed. At the same time the next design step must be made if the reliability index that was achieved by using deterministic methods is not high enough. This step is a direct probability optimization, i.e., methods of stochastic optimization should be used here.

First, it should be pointed out that most of optimization methods have the highest convergence speed when they start at a "good" initial point. Therefore, it would be most natural to get a previous solution as an initial point for the next design step.

As it was mentioned above, the stochastic methods are very time-consuming. And now some techniques that allow us to decrease computational cost will be considered. The first method, we dwell on, is the "correlated samples method".

tIt is clear that the optimization process needs only the value of difference AP between to points probabilities, not probability values. And the variance of AP is estimated to be

a2 {aP } = a2 {} + a2 {} - 2 R{ ; P2 }a{ },

where R is the correlation coefficient between two estimates. When the samples are independent, the R is equal to zero and the variance

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

a2 {aP} = a2 {aP} = a2 { } + a2 { }.

If a positive correlation can be achieved, R will be more than zero and, therefore,

a2

{aP }<a2 {AP },

i. e., for the same precision, simulation with correlated samples will require the less number of model runs, than independent samples. Intuition suggests that the positive correlation can be achieved by use the equal random influences for both device realizations to be compared.

The next approach is based on the supposition that the device operation state do not change inside the little

region near the previously checked point. Therefore, if a simulated point exists such that the distance from it to the point to be checked is less than a definite value, it is not need to perform the new system simulation.

In practical algorithm the internal device parameters quantization is accomplished. And every generated random point can be assigned to the one of the existing quants. The system simulation will be performed if no early checked points are exist within this quant. Each newly simulated point is presented as a vector of quant coordinates with the acceptability flag (0 or 1), so a database of checked quants can be build, and it is not necessary to make a system simulation more than one time per quant.

The experimental results show that the method in general is highly effective when compared to a conventional Monte Carlo technique, especially in conjunction with the "correlated samples".

The next way to decrease total design time on the phase of statistical optimization is to use modern supercomputing technologies and parallel processing. The easiest implementation of this idea would be multithreading programming on computers with symmetric multiprocessors (SMP) architecture and operating systems that can assign different threads to different processors (like Windows NT or modern clones of Unix). But computational power of conventional SMP systems is very limited - usually they consist of not more than two or four processors. Systems with massive parallel processing (MPP) architecture can contain much more processors (up to few thousands) but are very expensive.

It seems more applicable for most installations usage of distributed processing technologies. In this case computational tasks can be distributed over the set of networked workstations. And, by the way, here each computer can use multithreading and SMP architecture. Now the CAD system to be developed is modified for usage of parallel Monte Carlo method. The implementation of this technology is based on PVM (Parallel Virtual Machine) software package. The work is still in progress.

5. Interactive system for design optimization

All techniques described above were included in the computer-aided reliability-directed system called SPORA. The system SPORA was developed for parametric synthesis of analogous engineering devices with respect to reliability requirements. SPORA is designed to permit a user to observe, interrupt, diagnose, modify and restart a computation as it progress, resulting in very substantial savings not only in computing time, but olso in the overall time needed to

carry out a design. For example, when a design or optimization procedure is not succeeding in an attempt to meet specifications, the failure can be identified in an interactive CAD system by observing the output. The designer may therefore stop the computation and, making use of the heuristic information displayed on the screen, the designer could,

• modify the structure of the design,

• relax the constraint specifications,

• reset parameters of the procedure such as scaling, weights etc.,

• change problem formulation or algorithms.

This involvement in the optimization process put the designer in an ideal position to perform tradeoffs.

The system uses a widely distributed SPICE-like circuit simulation program, that allows to simulate a large class of analogous devices in direct current, frequency and time domains. SPORA also consists of features for nominal design, design centering, tolerances assignment, etc. The system has special graphic postprocessor for the best presentation of design results. SPORA is now in service and used for the reliability-oriented design of electronic equipment.

The SPORA system has been tested on a number of complex designs involving integrated circuits, filters and control systems.

6. Design example

•a

S

m

(16) jj016 R19

VT3

\® 1 i i

j VT2 1 1

jh+c1° HI

Çr9 DrIPD R12 □

R13 R14

®

! C17

VT4 <11_2

R16

©

= C15

Fig. 2 Band-width amplifier

Requirements:

Frequency band: 800...1800 MHz

Amplifier ratio: 25±2 DB

Probability (manufacturing yield) >0.9

Table 1. Initial nominal values (Pr = 0.44)

L1 C4 C9 C10 C11 C13 C14 C17

0.01m 0.05p 20.6p 15p 25p 10p 1.5p 1.5p

Table 2. Nominal values after first optimization step (Pr = 0.7)

L1 C4 C9 C10 C11 C13 C14 C17

0.03m 5.6p 8.2p 12p 8.2p 2.2p 12p 1.5p

Table 3. Optimal values (Pr = 0.92)

L1 C4 C9 C10 C11 C13 C14 C17

0.012 6.5p 8.5p 10p 8.9p 2.4p 11p 1.4p

7. Conclusion

A theoretical approach and applied techniques for designing analogous engineering devices and systems with due account of random variations in systems parameters and reliability specifications were considered. For solving this problem we use a so-called operational/parametric approach based on the computer-aided simulation of system capability and availability, parameters deviations and techniques of optimal parametric synthesis with respect to reliability criteria.

Under this approach the optimal parametric synthesis task is formulated as an optimization problem with a stochastic criterion.

The replacement of original stochastic criterion with a certain deterministic one is used in order to reduce computational cost of statistical algorithms. Moreover, a few techniques for decreasing the number of model runs in Monte Carlo method were implemented.

On the basis of the proposed methods and algorithms a computer-aided reliability-oriented design system called SPORA has been developed. The SPORA system builds mathematical models and calculates ratings of component parameters so that achieve the highest precision, acceptability (manufacturing yield) or reliability of engineering systems under design.

References

[1] O.V.Abramov, "Reliability-Directed Parametric Synthesis of Stochastic Systems", Nauka Publishers, Moscow (in Russian), 1992.

[2] O.V. Abramov and K.S.Katuyev, "Effective Methods for Parametric Synthesis of Stochastic Systems", Proc. First Asian Control Conference, Tokyo, vol. 3, pp. 587-589, 1994.

[3] R.Brayton, G.Hachtel and A. Sangiovanni-Vincentelli "A survey of Optimization Techniques for Integrated-Circuit Design", Proc. IEEE, vol.69, no. 10, pp 1334-1362, 1981

i Надоели баннеры? Вы всегда можете отключить рекламу.