Научная статья на тему 'A guaranteed control problem for a linear stochastic differential equation'

A guaranteed control problem for a linear stochastic differential equation Текст научной статьи по специальности «Математика»

CC BY
92
10
i Надоели баннеры? Вы всегда можете отключить рекламу.
Журнал
Ural Mathematical Journal
Scopus
ВАК
Область наук
Ключевые слова
GUIDANCE PROBLEM / GUARANTEED CLOSED-LOOP CONTROL / LINEAR STOCHASTIC DIFFERENTIAL EQUATION

Аннотация научной статьи по математике, автор научной работы — Rozenberg Valeriy L.

A problem of guaranteed closed-loop control under incomplete information is considered for a linear stochastic differential equation (SDE) from the viewpoint of the method of open-loop control packages worked out earlier for the guidance of a linear control system of ordinary differential equations (ODEs) to a convex target set. The problem consists in designing a deterministic control providing (irrespective of a realized initial state from a given finite set) prescribed properties of the solution (being a random process) at a terminal point in time. It is assumed that a linear signal on some number of realizations is observed. By the equations of the method of moments, the problem for the SDE is reduced to an equivalent problem for systems of ODEs describing the mathematical expectation and covariance matrix of the original process. Solvability conditions for the problems in question are written.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «A guaranteed control problem for a linear stochastic differential equation»

URAL MATHEMATICAL JOURNAL, Vol. 1, No. 1, 2015

A GUARANTEED CONTROL PROBLEM FOR A LINEAR STOCHASTIC DIFFERENTIAL EQUATION1

Valeriy L. Rozenberg

N.N. Krasovskii Institute of Mathematics and Mechanics of the Ural Branch of the Russian Academy of Sciences, Ekaterinburg, Russia, [email protected]

Abstract: A problem of guaranteed closed-loop control under incomplete information is considered for a linear stochastic differential equation (SDE) from the viewpoint of the method of open-loop control packages worked out earlier for the guidance of a linear control system of ordinary differential equations (ODEs) to a convex target set. The problem consists in designing a deterministic control providing (irrespective of a realized initial state from a given finite set) prescribed properties of the solution (being a random process) at a terminal point in time. It is assumed that a linear signal on some number of realizations is observed. By the equations of the method of moments, the problem for the SDE is reduced to an equivalent problem for systems of ODEs describing the mathematical expectation and covariance matrix of the original process. Solvability conditions for the problems in question are written.

Key words: Guidance problem, Guaranteed closed-loop control, Linear stochastic differential equation.

Introduction

The problem of constructing optimal strategies of guaranteed feedback control under conditions of uncertainty is one of the most important in mathematical control theory and its applications. In the present paper following the theory of closed-loop control developed by N.N. Krasovskii's school [1-3], the approach based on the so-called method of open-loop control packages originating from the technique of nonanticipating strategies from the theory of differential games [4] is applied to solving the guidance problem for a linear SDE. The method tested on the guidance problems under incomplete information for linear controlled systems of ODEs consists in reducing the problems of guaranteed control formulated in the class of closed-loop strategies to equivalent problems in the class of open-loop control packages. The latter class contains the families of open-loop controls parameterized by admissible initial states and possessing the property of nonanticipation with respect to the dynamics of observations [5-7].

This paper is devoted to the study of the problem of guiding (with a probability close to one) a trajectory of a linear SDE to some target set. The statement means that we should form a deterministic control providing (irrespective of the realized initial state from a specified finite set) prescribed properties of the solution (being a random process) at a terminal point in time. Here, we observe a linear signal on some number of realizations. Similar problems arise in practical situations, when it is possible to observe the behavior of a large number of identical objects described by a stochastic dynamics. By the equations of the method of moments [8,9], the problem for the SDE is reduced to an equivalent problem for systems of ODEs describing the mathematical expectation and covariance matrix of the original process. The technique of the method of open-loop control packages developed in [5-7] is applied to the systems obtained.

The paper has the following structure. In Section 1, the main problem of closed-loop guaranteed control for a linear SDE under incomplete information is formulated. In Section 2, the procedure reducing the original problem to two auxiliary guidance problems for systems of ODEs is described and the equivalence of the problems for SDE and ODEs is established. In Section 3, a brief summary of results obtained earlier within the framework of the method of open-loop control packages for

1This work was supported by the Russian Science Foundation (project no. 14-11-00539).

ODEs is given. In Section 4, the necessary statistical estimates are analyzed. In Section 5, the main results of the paper are presented: a solvability criterion for the original problem and an assertion connecting the guidance accuracy and the number of trajectories of the original process that are available for measuring. In Section 6, we consider an example illustrating the application of the algorithm to solving the guidance problem for a first-order SDE. Note that the procedure reducing a problem for a linear SDE to the corresponding problem for the ODE, which is conceptually close to the one proposed in the paper, was used, in particular, in [10] for solving the problem of dynamic reconstruction of an unknown disturbance characterizing the level of random noise in a linear SDE on the base of measuring some realizations of the SDE's phase vector.

1. Problem statement

Consider a system of linear SDEs of the following form:

dx(t,w) = (A(t)x(t,w) + Bi(t)ui(t) + f (t)) dt + B2(t)U2(t) d£(t,w), x(to,w) = xo. (1.1)

Here, t e T = [t0, #], x = (x1,x2,..., xn) e Rn, £ = (£1,£2,..., £k) e Rk (all the vectors are treated as columns); w e H, (H,F, F) is a probability space; £(t, w) is a standard Wiener process (i.e., a process starting from zero with zero mathematical expectation and covariance matrix equal to It (I is the unit matrix from Rkxk)); f (t) is a continuous vector function with values in Rn; A(t) = [aij(t)}, Bi(t) = [buj(t)}, and B2(t) = [b2ij(t)} are continuous matrix functions of dimensions n x n, n x r, and n x k, respectively.

Two controls act in the system: a vector u1(t) = (u11(t),u12(t),..., u1r(t)) e Rr and a diagonal matrix U2 (t) = [u21(t),u22(t),...,u2k (t)} e Rkxk, which are Lebesgue measurable on T and take values from specified instantaneous control resources Su1 and Su2 being convex compact sets in the corresponding spaces. The control u1 enters the deterministic component and influences the mathematical expectation of the desired process. Since U2d£ = (u21d£1,u22d£2,... ,u2kd£k), we can assume that the vector u2 = (u21,u22,... ,u2k) characterizes the diffusion of the process (the amplitude of random noises).

The initial state x0 belongs to a finite set of admissible initial states X0, which consists of normally distributed random variables with numerical parameters (m0,D0), where m0 = Mx0 is the mathematical expectation, m0 e M0 = [m0,m0,... }, D0 = M(x0 — m0)(x0 — m0)* is the covariance matrix (the asterisk means transposition), D0 eD0 = ..., Dn2}. Thus, the set

X0 contains n1n2 elements. Note that, if 0 e D0, then there are deterministic vectors in X0. We assume that the system's initial state belongs to X0 but is unknown.

Equation (1.1) is a symbolic notation for the integral identity

t t x(t, w) = xo + J(A(s)x(s, w) + B1(s)u1(s) + f (s)) ds + J B2(s)U2(s) d£(s, w). (1.2)

to to

The latter integral on the right-hand side of equality (1.2) is stochastic and is understood in the sense of Ito. For any w e H, the formulated Cauchy problem has a unique solution and specifies the corresponding realization of the stochastic process x(t,w), t e T .A solution of equation (1.1) is defined as a stochastic process satisfying integral identity (1.2) for any t with probability 1. Under the above assumptions, there exists a unique solution, which is a normal Markov process with continuous realizations [11].

Note that equations similar to (1.1), (1.2) describe simplest linearized models, for example, of changing the size of a multi-species biological population in a stochastic medium, of price dynamics on goods markets under the influence of random factors, or of particle motion in some field.

The problem in question consists in the following. Let convex closed target sets Me Rn and D e Rnxn and a continuous matrix observation function Q(t) of dimension q x n be given.

At any time, it is possible to receive the information on some number N of realizations of the stochastic process x(t) (we omit the symbol w if we mean the process rather than its realization). The following signal is available:

y(t)= Q(t)x(t). (1.3)

Assume that, for a finite set of some specified times Ti £ T, i £ [1 : l], we can construct, using N realizations of the process x(t), a statistical estimate mN of the mathematical expectation m(Ti) and a statistical estimate Df of the covariance matrix D(Ti) such that

p( max{\\mN - m(Ti)||Rn , \\Dn - D(Ti)\\RnXn} < h(N)) = 1 - g(N), (1.4)

where h(N) and g(N) — 0 as N — oc. We show below that standard procedures of obtaining the estimates mN and DN admit modifications providing the validity of relation (1.4) and the specified convergences (a similar procedure was proposed in [10]).

Problem 1 of guaranteed closed-loop e-guidance consists in forming a control (u1(^), u2( )) guaranteeing, whatever the initial state xo from the set Xo, prescribed properties of the process x at the terminal time Here, we mean that, for an arbitrary small (in advance specified) e > 0, the mathematical expectation m($) and the covariance matrix D($) reach the e-neighborhoods of the target sets M and D, respectively. In the motion process, the sought control is formed using the information on N realizations of the signal y(t). By virtue of estimate (1.4), it is reasonable to require that the probability of the desired event should be close to 1 for sufficiently large N and algorithm's parameters concordant with N in a special way.

2. Reduction of the original problem

Let us reduce the guidance problem for the SDE formulated above to two problems for systems of ODEs. By virtue of the linearity of the original system, the mathematical expectation m(t) depends only on u1(t); its dynamics is described by the equation

m(t) = A(t)m(t) + 5i(t)ui(t) + f (t), t £ T = [to,tf], m(to) = mo £ Mo. (2.1)

We assume that N (N > 1) trajectories xr(t), r £ [1 : N], of the original SDE are measured; then, according to the statement of the guidance problem, we know values of signal (1.3) yr (t) = Q(t)xr(t).

The signal on the trajectory of equation (2.1) is denoted by ym(t) = Qm(t)m(t); its estimate formed by the information on yr, r £ [1 : N], by yf (t). The latter is constructed as follows:

1 N 1 N

yf (t) = N Eyr(t) = Q(t)mN(t), mN(t) = NY,xr(t). (2.2)

r=1 r=1

Obviously, Qm(t) = Q(t) and, for the finite set of times Ti £ T, i £ [1 : l], in view of relation (1.4), it holds that

P{Vi £ [1 : l] \\yN(Ti) - ym(Ti)\\R, < C1h(N)) = 1 - g(N), (2.3)

where the constant C1 can be written explicitly. Recall that M is the target set for the trajectory of equation (2.1).

The covariance matrix D(t) depends only on U2(t); its dynamics is described by the so-called equation of the method of moments [8,9] in the following form:

D(t) = A(t)D(t) + D(t)A*(t) + B2(t)U2(t)U^(t)B^(t), t £ T = [to,tf], D(to) = Do £ Do. (2.4)

For our purposes, matrix equation (2.4) is conveniently rewritten in the form of a vector equation, which is more traditional for the problem under consideration. By virtue of the symmetry of the matrix D(t), its dimension is defined as nd = (n2 + n)/2. Let us introduce the vector d(t) = {ds(t)}, s £ [1 : nd], whose coordinates are found by the elements of the matrix D(t) = {dij(t)}, i,j £ [1: n]:

ds(t) = dij(t), i < j, s = (n - i/2)(i - 1)+ j.

(2.5)

Actually, the vector d(t) consists of successively written and enumerated elements of the matrix D(t), taken line by line starting with the element located at the main diagonal. Let us write the symmetric right-hand side of (2.4) in detail:

A(t)D(t) + D(t)A*(t) =

( 2(a1,d1)

{a,2, di) + (ai,d2)

(ai,d2) + (a2, di)

2(a2, ¿2)

(ai,dn) + (an,di) \ (a2,dn) + (an,d2)

V (an,di) + (ai,dn) (an,d2) + (a2,dn)

2 (an, dn)

/

B2(t)U2(t)U2i(t)B*2(t) =

/ k k

Y1 b2ir u2r Y1 b2ir b22r u2r

r=i r=i

kk

Y b22rb2irY b22ru2r

r=i r=i

kk

Y1 b2nr b2ir u2r Y1 b2nr b22r u2r

r=i r=i

k2 Y b2irb2nru2r

r=i k

b22rb2nru22r

r=i

Y b2nr u2r

r=i

/

Here, ai, di, i £ [1 : n], are the rows of the matrices A and D; the symbol (■, ■) stands for the scalar product of vectors in the corresponding space. Note that the element (ij) of the matrix

nn

A(t)D(t) + D(t)A*(t), namely (ai, dj) + (aj,di) = ^ airdjr + ^ ajrdir, determines the coefficients

r=i r=i

for the coordinates of the vector d in the sth equation of the required system, where s is found from relation (2.5). Based on the representations above, by the matrices A(t) and B2(t), we form the matrices A(t) : T — RndXnd and B(t) : T — RndXk acting as follows.

1. To form the sth row (s £ [1 : nd]) of the matrix A, we should uniquely define by formula (2.5) the indices i and j (i,j £ [1 : n], i < j) corresponding to s, i.e., such that (n-i/2)(i — 1)+j = s. Further, we find the elements asr (r £ [1 : nd]). If s is such that i = j, then there can be at most n nonzero elements in the sth row: asr = 2aihl, hi £ [1 : n], r corresponds to (in the sense of (2.5), and the first index is not greater than the second) the pair of indices h]_i (or ihi). In the case when s is such that i = j, to fill in the row of the matrix A, one should analyze the three variants: (i) r does not correspond to any pair of indices from the set {1i, 2i,... ,ni,i1,i2, ...,in, 1j, 2j,... ,nj,j 1,j2,... ,jn}, then asr = 0; (ii) r corresponds to some (unique) pair hii (ihi), hi £ [1 : n], then asr = ajhl, or to a pair h2.j (jh2), h2 £ [1 : n], then asr — aih2

; (iii) r corresponds to two pairs of indices, in this case, hi = j and h2 = i, then asr = aii + ajj. Since such a coincidence is unique for i = j, obviously, we have at most 2n — 1 nonzero elements in the sth row of the matrix A.

2. To form the sth row (s £ [1 : nd]) of the matrix B, we should uniquely define by formula (2.5) the indices i and j (i,j £ [1 : n], i < j) corresponding to s, i.e., such that (n — i/2)(i — 1) + j = s. Further, we find the element bsr (r £ [1 : k]) of the matrix B by the rule bsr = b2irb2jr.

Obviously, the elements of the matrices obtained are continuous. Performing standard matrix operations, it is easy to verify that system (2.4) is rewritten in the form

d(t) = A(t)d(t)+ B(t)v(t), t £ T =[to,tf], d(to) = do £V0.

(2.6)

The initial state d0 is obtained by Do; the notation for the set is the same. The multiplication of the diagonal matrices U2(t)U|(t) results in the appearance of the control vector v(t) = (u^(t), u22(t),... ,u2k(t)), whose elements take values from some convex compact set Sv £ Rk for all t £ T.

The signal on the trajectory of equation (2.6) is denoted by yd(t) = Qd(t)d(t); its estimate formed by the information on yr, r £ [1 : N], by yN (t). The latter is constructed as follows:

1

N

N1

E(yr (t) - yN (t))(yr (t) - yN (t)T

r=l

1

N

Q(t)(t) - mN(t))(xr(t) - mN(t)TQ*(t) = Q(t)DN(t)Q*(t), (2.7)

r=l

where DN(t) = {dN(t)}, i,j £ [1 : n] is the standard estimate of the covariance matrix D(t) for an unknown (estimated by mN(t)) mathematical expectation m(t). Symmetric q x q-matrix (2.7) can be represented in the form

/

Q(t)DN (t)Q* (t) =

E E qirqipdN E E q2rqipdN

N

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

r=1p=1 r=1p=1 n n n n

E E q1rq2pdN E E q2rq2pdjyr

r=1p=1 r=1p=1

E E q1rqqpdN E J2 q2rqqpdpr \ r=1p=1

r=1p=1

nn

Z Z qqrq1pdpr r=1p=1 nn

Z Z qqr q2pdNr r=1p=1

E E qqrqqpdpNr

r=1p=1

Obviously, it is sufficient to consider nq = (q2 + q)/2 elements of this matrix. The element

nn

(i1j1), i1 < j1, i1,j1 £ [1 : q], namely Z E qiirq^pd^, defines the S1th row

r=1p=1

(s1 = (q — i1 /2)(i1 — 1) + j1 by analogy with (2.5)) in the formed relation

yN (t) = Qd(t)dN (t),

where Qd(t) is the continuous nq x nd-matrix, dN(t) is the nd-vector extracted from DN(t) by rule (2.5). Then, to write the element Qd[s1,s2], it is necessary to find, using s1 and s2, the indices i1, j1 (i1, j1 £ [1 : q], i1 < j1, (q - i1/2)(i1 - 1) + j1 = S1) and i2, j2 (i2,j2 £ [1 : n], i2 < j2, (n - i2/2)(i2 - 1) + j2 = s2), respectively. It is easy to verify that

Qd[S1,S2]

= ) qiii2 qjlj2 , i2 = j2 qhi2 qjlj2 + qilj2 qjli2 , i2 = j2

Thus, by matrix (2.7), we construct the estimate yN(t) = Qd(t)dN(t) of the signal yd(t) = Qd(t)d(t). Obviously, for the finite set of times Ti £ T, i £ [1 : l], we have the relation of type (1.4):

P{Vi £ [1 : l] \\yN(Ti) - yd(n)\\Rnq < C2h(N)) = 1 - g(N),

(2.8)

where the constant C2 can be written explicitly. We denote the target set for the trajectory of equation (2.6) by the prior symbol D.

Thus, original problem 1 of guaranteed closed-loop e-guidance for the SDE can be reformulated as follows.

Problem 2. For an arbitrary small (in advance specified) e > 0, it is required to choose controls u1 (•) in equation (2.1) and v(-) in equation (2.6) such that the trajectories of (2.1) and (2.6) reach the e-neighborhoods of the target sets M and D at the terminal time whatever the initial states m0 from the set M0 and d0 from the set D0. It is important that the probability of the desired event should be close to 1.

The required controls are formed through the estimates of the signals y and yd satisfying relations (2.3) and (2.8); actually, these controls define the control in SDE (1.1). The dependence

of the number N of measurable trajectories (necessary for the estimation) on the value e is given below. The next theorem follows from the procedure of constructing problem 2.

Theorem 1. Problem 1 and problem 2 are equivalent.

Thus, to solve problem 1, one should establish some conditions of consistent solvability of the problems of e-guidance for ODEs (2.1) and (2.6), i.e., solvability conditions for problem 2, and should find the form of concordance of parameters N and e as well.

3. The method of open-loop control packages: a brief review of results for ODE

Let us present briefly the approach by A.V. Kryazhimskii and Yu.S. Osipov to solving the problem of closed-loop guidance under incomplete information for a linear ODE [5-7]. Consider a dynamical control system

where x(t) £ Rn, u(t) £ P C Rm (P is a convex compact set); A(-), B(■), f (■) are continuous matrix function of dimension n x n, n x m, and n x 1, respectively; X0 is a finite set of possible initial states. The real initial state of the system is assumed to be unknown. A convex closed target set M C Rn and a continuous matrix observation function Q(t) of dimension q x n are given.

The problem of guaranteed closed-loop e-guidance consists in forming by the signal y(t) = Q(t)x(t) a control guaranteeing that the system's state x(ft) reaches an arbitrarily small e-neighborhood of the target set M (e > 0 is in advance specified) at the terminal time ft. The solution of the problem is sought in the class of closed-loop control strategies with memory. The correction of the values of a control u(-) is possible at in advance specified times. In [6], the equivalence of the formulated problem of closed-loop control to the so-called problem of package guidance is established; the terminology of the latter problem and conditions of its solvability are discussed in [7].

Consider the homogeneous system

its fundamental matrix is denoted by F(■, ■). For any admissible initial state x0 £ X0, the homogeneous signal corresponding to x0 is the function gx0(t) = Q(t)F(t,t0)x0, t £ [t0,ft]. The set of all admissible initial states x0 corresponding to a homogeneous signal g(-) till a time t is denoted by X0(t|g(-)) = {x0 £ X0: gx0(Ol^r] = gt)l[t0,r]}, where g(^)\[t0,T] is the restriction of the homogeneous signal g(-) onto the interval [t0,T].

A family (ux0 (■))X0ex0 of open-loop controls is called an open-loop control package if it satisfies the condition of nonanticipation: for any homogeneous signal g(-), time t £ (t0, ft], and admissible initial states x'0, x'0 £ X0(t\g( )), the equality uX.'o(t) = uXo(t) holds for all t £ [t0,T]. An open-loop control package (uX0( ))XoeXo is guiding if, for any x0 £ X0, the motion from x0 corresponding to uX0(■) takes a value exactly in the target set M at the time ft. If there exists a guiding open-loop package, we say that the idealized problem of package guidance [5,6] corresponding to the original problem of guaranteed closed-loop is solvable. Note that it is assumed in these problems that the signal is known exactly.

Let G be the set of all homogeneous signals (their number does not exceed the number of elements of the set X0). We introduce the set G0(g() of all homogeneous signals coinciding with g(-) in a right-sided neighborhood of the initial time t0. The first splitting moment of the homogeneous signal g(-) is the time

x(t) = A(t)x(t) + B(t)u(t) + f (t), t £ T = [t0, ft], x(t0) = x0 £ X0,

(3.1)

x(t) = A(t)x(t), t £ T =[t0,ft], x(t0) = x0 £ X0;

If ti(#(•)) < then, by analogy with G0(g(-)), we introduce the set G1(g(-)) of all homogeneous signals from G0(g(-)) coinciding with g(-) in a right-sided neighborhood of the splitting moment Ti(g(-)). Actually, the difference G0(g(-))\G1(g( )) informs on the number of the signals from G0(g(-)) noncoincident with g(-) after T1(g(-)). By analogy with n1(g(-)), we define the second splitting moment of the homogeneous signal g(-) and so on. Finally, we introduce the set of all the splitting moments of the homogeneous signal g(-): T (g()) = {Tj (g()): j = 1,... ,kg}, kg > 1, Tkg (g( )) = Then, we consider the set (ascending ordering) of all the splitting moments of all the homogeneous signals (in [7], it is shown that these moments are possible switching moments for the "ideal" guiding open-loop control): T = Ug(.)eG T(g(0), T = {ti,.. .,tk}, K <Y.g(.)ec kg(-) is the number of elements of the set T. Obviously, the sets T(g( )) and T are finite due to the finiteness of the sets X0 and G. For any k = 1,...,K, the set X0(Tk) = {X0(Tk|g(-)): g( ) E G} is called the cluster position at the time Tk, whereas each of its elements X0k is called the cluster of initial states at this moment.

The constructions described are used for the characterization of open-loop control packages. Further, in [7], by the introduction of an auxiliary extended open-loop guidance problem for the system consisting of copies of system (3.1) parameterized by the admissible initial states, a criterion for the solvability of the original problem based on solving a finite-dimensional optimization problem is obtained. Let us formulate the main result of [7], as far as possible without using the terminology of the extended problem.

Theorem 2 [7, Theorem 2]. The problem of package guidance for system (3.1) is solvable if and only if

sup Y((1xo)xoeXo) < 0,

(lx0 )x0eX0 £<S

K Tk

Y ((Ixo )xoeXo )= £ (Ixo ,F (0,t0)x0) + £ J £ p- ( £ B *(s)F*(V,s)lx0\p) ds

xo€Xo k=1 Tk-1 Xok eXo(Tfc) xoeXok

i9

^ E lxo, F (0,s)f (s)) ds - Y, p+(lxo IM). (3.2)

to xo&Xo xo&Xo

Here, (lxo)xoeXo is a family of vectors from Rra parameterized by the admissible initial states (the number of vectors coincides with the number of elements in X0); S is the set of families (lxo)xoeXo such that ZxoeXo \\lxo = 1; (',') is the scalar product in the corresponding finite-dimensional Euclidean space; p-(-|P) is the lower support function of the set P, and p+(-|M) is the upper support function of the set M.

4. Properties of the statistical estimates

Lemma 1. For a finite set of some specified times Ti E T, i E [1 : l], the standard estimates mN of the mathematical expectation m(Ti) and DN of the covariance matrix D(Ti) constructed through N (N > 1) realizations x1(Ti), x2(Ti),... ,xN (n) of the random variables x(Ti) by the following rules [12]:

N

m

N

= ^E x^ (Ti)

N

(4.1)

r=1

DiN

N

1

N

N1

¿2(xr(Ti) - mN)(xr(Ti) - mN)*

(4.2)

r=1

provide the validness of relation (1.4) (consequently, (2.3) and (2.8)).

Proof. We present the proof of the lemma for case n = 1 (the argument for n > 1 is similar; however, the formulas are too cumbersome). In this case, the scalar variable x(Ti) is normally distributed with the mathematical expectation m(Ti) and dispersion D(Ti). Consider estimate (4.1). Let us show that it has the following property:

P (Vi £ [1: l] \m? - m(Ti)|< hm(N)) = 1 - gm(N); hm(N), gm(N) - 0 as N - o. (4.3)

First, we prove that

Vi £ [1: l] P (\mf - m(Ti)\< hm(N)) = 1 - fm(N); fm(N) - 0 as N - o. (4.4)

As is known [12], the random variable £ = (mN - m(Ti))^JN/D? has the t-distribution (Student's distribution) with N-1 degrees of freedom, which is close to the standard normal law N(0, 1) for N > 30. Our aim is to modify the procedure of constructing the estimate of the mathematical expectation in case of unknown dispersion in such a way that relation (4.4) is fulfilled. Consider the following expression for 0 < y < 1/2:

P(-NY < (mf - m(Ti))\JN/D.N < N^ = P(-ND.N/N < mf - m(n) < N^D?/N)

= P(|mN - m(Ti)| < NyDN/n) = 1 - fm(N).

We can assume that hm(N) = C1NY—1/2 in (4.4) (here and below, by Ci we denote auxiliary constants, which can be written explicitly). On the other hand,

P(-NY < (mN - m(Ti))y/N/DN < N^ = 2(Ft,n-i(NY) - Ft, n-1 (0)) = 2Ft, n-i(Ny) - 1

= 2(Ft,N-1(NY) - $(N V1 - 2/N)) + 2$(N V1 - 2/N) - 1 > 2$(NV1 - 2/N) - 1 - 2 \Ft?N-1(NY) - $(NV1 - 2/N )\ = 1 - 2(1 - $(NV1 - 2/N)) - 2 \Ft?N-1(NY) - $(NV 1 - 2/N )\.

Here, Ft,?-1(x) is the probability function of the t-distribution with N - 1 degrees of freedom,

/X 1 2

_e-y /2 dy.

-<x \ 2n

Using the inequality describing the closedness of the functions Ft,?-1(x) and $(x) for large N in the form [13]

\Ft,N—1 (x) - $(xV 1 - 2/N )\ < C2/N, (4.5)

as well as the asymptotics of the normal distribution as x 00 in the form [12]

1 / p — X2/2 N

1 - §(x) = -^ e—X /2 + o(p-), (4.6)

xV2n V x J' v '

we obtain

P(-NY < (mf - m(Ti))\JN/D? < NY) > 1 - 2C\/N - C3p—n2''/2/NY + o(P-?2''/2/NY).

Since the asymptotics of the normal distribution (see (4.6)) is suppressed by the error of the normal approximation (see (4.5)), we can assume (see (4.4)) that fm(N) < C4/N. Relation (4.4) holds. Denote

Ami = {\mf - m(Ti)\ < hm(N)} , P(Ami) = 1 - fm(N) Vi £ [1 : l]. Then, ( \ \ )

P (Vi £ [1 : l] \mf - m(Ti) \ < hm(N)) = P (Am1Am2 ... Ami)

z-1

= P(Am1)P(Am21Am1)P(Am3|Am1Am2)... P^Am \ JJ Amk) > (1 - fm(N))z > (1 - lfm(N)) .

k=1

The last but one inequality is based on the assumption that all the conditional probabilities are not less than the corresponding unconditional ones due to the continuity of trajectories of the process under consideration; the last inequality follows from the smallness of fm(N). Obviously, we can assume that gm(N) = lfm(N) = lC4/N in (4.3). Thus, it holds that

P (Vi E [1 : l] \mN - m(Ti)\< C\NY-1/2) =1 - lC4/N. (4.7)

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Now, pass to the proof of the following property of estimate (4.2):

P (Vi E [1: l] \DN - D(Ti)\< hd(N)) =1 - gd(N); hd(N),gd(N) ^ 0 as N ^ to. (4.8) First, we show that the relation

Vi E [1 : l] P (\DN - D(n)\ < hd(N)) = 1 - fd(N), fd(N) ^ 0 as N ^to (4.9)

is valid.

As is known [12], the random variable £ = (N - 1)DN/D(ni) has the distribution x2 with N -1 degrees of freedom and the distribution of the variable - yj2N - 3 is close to the standard normal law N(0, 1) for N > 30. Our aim is to modify the procedure of constructing the estimate of the dispersion in case of unknown mathematical expectation in such a way that relation (4.9) is fulfilled. Let us choose the numerical parameters:

x2 = (V2N - 3 - N2 /2 = N2y/2 - NY V2N - 3 + N - 3/2,

x2 = (V2N - 3 + N2 /2 = N2y/2 + NY V2N - 3 + N - 3/2, 0 < y < 1/2. For such a choice, we obtain x1/N ^ 1 and x2/N ^ 1 as N ^ to. We have

P (x2 < (N - 1)DiN/D(Ti) < x2)

= P ((x1/(N - 1) - 1) D(Ti) < DN - D(Ti) < (x2/(N - 1) - 1) D(Ti)) = 1 - fd(N). Then, by virtue of the definition of x12 and x22, we get

P (\dN - D(Ti)\ < hd(N)) = 1 - fd(N), where hd(N) = max {\x1/(N - 1) - 1\ D(ni), \x2/(N - 1) - 1\ D(ni)} = C5NY-1/2. On the other

ie[1:Z]

hand, ( )

P (x2 < (N - 1)DiN/D(Ti) < x2)

= (V2N - 3 - N7) 2 /2 < (N - 1)DN/D(Ti) < (V2N - 3 + N^ 2 /2^ = P (-NY 2(N - 1)DN/D(Ti) - V2N - 3 < NY^J = P ^^J2(N - 1)DN/D(Ti) - V2N - 3 < N^ - P ^yj2(N - 1)DN/D(m) - V2N - 3 <

= FhN-1 (NY) - FiN-1(-NY),

where Fi>N —1(x) is the distribution function of the variable y 2(N - 1)D?/D(Ti) - V2N - 3. Using the improved Berry-Esseen inequality [12] describing the closedness of the function FiN? —1(x) and the function of the normal distribution in the form

\Fi?—1 (x) - $(x)| < C6/(2VN-1(1 + |x|3)), (4.10)

as well as the specified above asymptotics of the normal distribution (see (4.6)), we derive

FiNN—1(NY) - FiNN —1(-NY) = FiNN—1(NY) - $(NY) + $(-NY) - F%,?—1-NY) + $(NY) - $(-NY)

> 2$(NY)—1 — |Fi,?—1(NY)-$(NY^-Fn—1(-NY)-$(-NY)| > 2$(NY)-1-c6/(Vn-1(1+n3y))

= 1 - 2p—n2/2/(NY- C6/(VN=1(1 + N3Y)) + o[p—N2Y/2/Ny) .

Since the asymptotics of the normal distribution (see (4.6)) is suppressed by the error of the normal approximation (see (4.10)), we can assume (see (4.9)) that fd(N) < C7N—1/2—3. Relation (4.9) holds. Introducing the notation

Adi = {\D? - D(Ti)\ < hd(N)} , P(Adi) = 1 - fd(N), Vi £ [1 : l], by analogy with the events Ami, we conclude that

P (Vi £ [1 : l] \D? - D(Ti) \ < hd(N)) = P (Ad1Ad2 ... Adi) > (1 - lfd(N)). Then, obviously, we can assume that gd(N) = lfd(N) = lC7N—1/2—3~( in (4.8). Thus,

P (Vi £ [1: l] \D? - D(Ti)\< CN1—1/2) =1 - lCjN—1/2—3Y. (4.11)

Comparing (4.3) and (4.8), we conclude that, in relations (1.4), (2.3), and (2.8), we can consider the same parameters, namely,

h(N) = ChNY—1/2, g(N) = Cg Nmax{—1'—1/2—3y}, (4.12)

where Ch and Cg are constants, which can be written explicitly. For example, choosing the value Y = 1/6, we obtain h(N) = ChN—1/3, g(N) = Cg/N. For y — +0, h(N) and g(N) have the power exponents of the value 1/N, which are asymptotically equal to 1/2. □

5. Criterion for the solvability of problems 1 and 2.

Concordance conditions for parameters

In addition, we define several notions for ODEs (2.1) and (2.6) based on the information from Section 3. Let G1 = {g1(-)} and G2 = {g2(•)} be the sets of all homogeneous signals for (2.1) and (2.6), respectively. The sets of all splitting moments of all homogeneous signals for (2.1) and (2.6) are denoted by T1 = {t¡,... ,tK 1} and T2 = {t2, ... ,tK2 }; the cluster positions and clusters of initial states at the times t¡1 and t|, by Mo(tI) and M0k, Do(t|) and D0k, respectively. Recall that the last splitting moment always coincides, by definition, with ft: t^1 = tK2 = ft. To simplify the presentation, we assume t^ = Tq = t0 (t0 is not a splitting moment). Let us introduce the sets of pairs of homogeneous signals splitted (in the sense of definitions from Section 3) at the moments tI k £ [0 : K1 - 1] and t2, k £ [0 : K2 - 1]: = {(gK^gjT1))}, Gf = {(g2(T2),g2(T2))}, i = j. A moment from the interval T ,tI + Ce] T + Ce < tI+1, the sense of the constant C is clarified below), at which all the pairs from Gl* are distinguishable, is denoted by tI * and is called a distinguishing moment for all the signals splitted at the time t^. Similarly, we define a distinguishing moment t2*. Obviously, we can believe that, for all t1 £ T1, k £ [0 : K1 - 1] and Tk £ T2, k £ [0 : K2 - 1], the corresponding moments Tk * and Tk * are defined uniquely; at these

moments, the signal's values in all the pairs from G^* and Gk* are different. Indeed, if we assume that, in some time after splitting, the signals coincide once again (that is stipulated by the form of the observation matrix Q), then the finiteness of the number of the signals, their continuity, and the definition of splitting moments imply the existence of a minimum moment from (t^,t1 + Ce] ((Tk,Tk + Ce]) such that the signals in any pair from Gk* (Gk*) do not coincide till this moment. The set of all such distinguishing moments for (2.1) and (2.6)

T* = T1* U T2*, T1* = {t1*,...,tk*1-1}, T2* = {t02*,...,tK*!-1}, (5.1)

determines both the aforesaid set of l (l < K1 + K2) times, at which the N trajectories of the original process are measured, and the set of times, which are possible for switching the closed-loop control. Note that the case T* = 0 takes place only if all the signals coincide on the whole interval [io,0].

Let us formulate according to (3.2) the solvability condition for the guidance problem for (2.1):

sup Y1((lmo )moEMo ) < 0,

(Zm0 )m0 EMo

eSi

T i

Ki Tk

Y1((lmo )moeMo )=E ^mo, F1(&,to)mo) + E / E E B1(s)F1*(§, s)lmo ds

moEMo k= 1 i Mok EMo(t1) mo EMok

Tk-1 k

&

+ /( E lmo ,F1(0,s)f (S)) ds - E p+(lmo M), (5.2)

to moEMo moEMo

and the solvability condition for the guidance problem for (2.6):

sup Y2((ldo )doEDo) < 0, Y2((ldo )doEDo )= E (ldo ,F2(^,t0)d0)

(zdo) doe-Do ES2 doeVo T 2

K2 k

+ E / E P-( E B*(s)F№s)ldo K) ds - E p+(ldo D). (5.3)

k=1 T 2 Dok EDo(t2) doEDok doEDo

' k — l k

Here, (lmo)moEMo, (ldo)doEDo are families of vectors parameterized by the corresponding initial states; S1, S2 are the set of families (lmo)moEMo, J2moEMo ¥mo III« = 1 and (ldo)doEDo J2doEDo IIldoII«d = 1; F1(', •) and F2(-, ■) are the fundamental matrices of systems (2.1) and (2.6). Let us prove the following statement actually being the main result of the paper.

Theorem 3. Let conditions (5.2) and (5.3) be fulfilled, let the information on N trajectories of SDE (1.1) be received at the times belonging to the set T* (5.1), let the constants Ch, Cg, and y be taken from the Lemma 1 (see (4.12)) and

N > (2Ch/p(e)f(1-2Y) , (5.4)

P(e) = min{ 1 1 mil 1 |gi1(Tk1*) -gjCkOb«, 2 2 mi9n 2N 2 Hg»2(t1*) -gj^Hh«*}.

T)l*ET1(gl,gj)EGk* j TI*ET2*, (g2,g2)EG2C J J

Then, problem 1 of e-guidance is solvable with the probability 1 - CgNmax^ 1 1/2 37^ and there exists an e-guiding control in equation (1.1) based on open-loop control packages for systems (2.1) and (2.6).

Proof. Conditions (5.2) and (5.3) provide the solvability of the idealized package guidance problems corresponding to the problems of guaranteed closed-loop control composing problem 2. Thus, there exist open-loop control packages (u\m0())m0eM0 and (v*o(•))d0ev0 parameterized by the initial states and solving the idealized package guidance problems for systems (2.1) and (2.6). The solution of the problems of guaranteed e-guidance implies the possibility of synthesis (by u\mo and v*o) of the corresponding closed-loop controls bringing the trajectories to e-neighborhoods of the target sets M and D, respectively. The input information is estimates (2.2) and (2.7) constructed by realizations of signal (1.3). Thereby, according to Theorem 1, we can design an e-guiding control solving problem 1 for system (1.1).

Let make the aforesaid more precise using system (2.1) as an example. This system starts at the time t0 from some unknown initial state mo £ M0; the guiding open-loop control for this state is denoted by u*1fho(•). Let T1 * contain at least one distinguishing moment. We construct the closed-loop control Ul(-) as follows. Before the motion starts, we decide to apply to system (2.1) the test control U\(t) = U10 £ Su1, t £ [t^r^ *). At the moment r0* of initial distinguishing of the signals, we measure N trajectories of (1.1) and, using them, construct the estimate yN(tq *) of the realized signal y-mT*) = QmT*)m(T01 *). Using the Cauchy formula, we obtain the homogeneous signal

QmT*)Fi(t1 *,to)mo = ym(Tl*) - Qm(r1 *) J Fi(t1 *,t)(Bi(t)U*w + f (t)) dT.

to

Now, basing on the estimate yN T1 *), we find the cluster of initial states M0k containing m0. Taking into account (2.3) and (4.12), we have \\y?T *) — ymT*)||Rq < h(N). Further, if

h(N) < p1(e)/2 = 1 min 1 Wg1(r1 *) — g1(r1 *)\\r/2, (5.5)

then the cluster M01 = {m0 £ M0 : gmo(•)|[t0;T)] = gm0(•)l[to,r)]} is defined uniquely (since there is no any splitting of the homogeneous signal till the moment r1). Consequently, at every time t £ [tq1 *,r1 *), we apply the control u*1Moi(t) from the package (u1mo( ))m0eM0 corresponding to the cluster M01, i.e., the control u1m0(t). A similar procedure is applied for all the splitting moments rl* of the signals. Obviously, the control u\ (t) constructed in such a way differs from the guiding open-loop control u1m,0 (t) only on at most K1 intervals, each of which has a length not more than Ce. Denote the solutions of (2.1) for the initial state m0 corresponding to the controls u1fho(t) and u1(t) by m*(•) and m*(•), respectively. Note that m* (ft) £ M. In view of the boundedness of all the functions and the control in (2.1), we estimate the value

i9

Wm* (ft) — m* (ft)WRn = f F1(ft,r)B1(r)(u1(r) — v*1m o (r)) dr < KmCe, (5.6)

J Kn

t0

where the constant Km can be written explicitly. Hence, choosing C < 1/Km, we guarantee \\m*(ft) — m*(ft)W < e, i.e., the value m*(ft) reaches the e-neighborhood of the target set M.

For system (2.6), all the argument is similar. The attainability of the e-neighborhood of the target set D by the corresponding trajectory is guaranteed by the choice of C < 1/Kd in the relation similar to (5.6). The form of the function p(e) (all the minimums in it exist in view of the finiteness of all the sets involved) is explained by the necessity to require estimates similar to (5.5) for any r* £ T*. Thus, we obtain h(N) < p(e)/2. This and (4.12) imply relation (5.4) connecting N and e. Finally, by (4.12), all the actions described above, including the e-guidance of solutions (2.1) and (2.6), consequently, the solution of (1.1), hold with the probability 1 — CgNmax{-1'-1/2-il^ converging to 1 as N oo.

The theorem is proved. □

1

0

6. Illustrative example

Consider the linear SDE of the first order:

dx(t) = —x(t)dt + ui(t)dt + U2(t)d£(t), t e T = [0,2], ui,u2 e [0,1], (6.1)

with the unknown initial state xo e X0, X0 consists of four normally distributed random variables with numerical parameters (m0,d0), where the mathematical expectation m0 e M0 = m,J = (3 — e)e, m^ = e2, and the dispersion d0 e D0 = {d^dl}, d^ = e2/2, d0 = e4. Incomplete observations have the following form:

y(t) = Q(t)x(t), Q(t) = j 0— 1, t e (1,2] • (6.2)

Let us write ODEs for the mathematical expectation and dispersion, as well as the observed signals, using the formulas from Section 2:

m(t) = —m(t) + Ui(t), m(0) = m0 e {m0, m2}, ym(t) = Q(t)m(t); (6.3)

d(t) = —2d(t) + u2(t), d(0) = d0 e {d0,d2}, yd(t) = Q2(t)d(t). (6.4)

Let the target sets for m and d be as follows: M = [2/e, 1], D = [1/2, 1].

The problem consists in forming, for an arbitrary small (in advance specified) e > 0, an open-loop control (ui,u2) guaranteeing, whatever the initial states m0 e M0 and d0 e D0, by the information on N trajectories of equation (6.1), the attainability (with a probability close to 1) of e-neighborhoods of the target sets M and D by the the mathematical expectation m(2) and the dispersion d(2), respectively. Let us present the solutions of equations (6.3) and (6.4):

t t m(t) = e-tm0 + e-t J eTu,(t)dr, d(t) = e-2td0 + e-2t J e2ru2(t)dr. 00

Obviously, the splitting moments of the homogeneous signals for equations (6.3) and (6.4) coincide. The form of the function Q(t) and the structure of the sets M0 and D0 imply that the number K of the splitting moments is equal to 2, r, = 1 is the first splitting moment, after which it is possible to distinguish the homogeneous signals corresponding to the different initial states, the terminal time r2 = 2 is, by definition, the second splitting moment. Note that, since the guiding open-loop control package is presented below, the solvability criterion for the original problem (written by the formulas form Section 5), obviously, holds.

In the case when the controls u, and u2 are in the form of piecewise constant functions

uio={e C;v uKt)={^t e

we obtain

m(2) = e-2m0 + (e — 1)e-1(e-1u[0;i] + u^]), d(2) = e-4d0 + (e2 — 1)e-2(e-2V[0,i] + V(i>2])/2.

It follows from the form of m(2) that the solution of equation (6.3), starting from the greater initial state m0 = e2, reaches (at t = 2) the set M (namely, its upper boundary m = 1) only under the action of zero control ui on the whole interval [0,2], i.e., u[0 i] = u(i 2] = 0. At the same time, if the real initial state coincides with the smaller possible value m^ = (3 — e)e, then, after the necessary action of zero control till the splitting moment t = 1 of the homogeneous signal, only the choice of u(i 2] = 1 forces the trajectory to reach the lower boundary m = 2/e of the target set M.

Obviously, for the set of parameters chosen in the example, the open-loop control of equation (6.3) guaranteeing the attainment of the trajectory to M at the time t = 2 is unique.

A similar argument is applicable to equation (6.4). Its solution, starting from the greater initial state dl = e4, reaches (at t = 2) the set D (namely, its upper boundary d = 2) only under the action of zero control u2, on the whole interval [0,2], i.e., v[0,1] = v^^] = 0. At the same time, if the real initial state coincides with the smaller possible value d1 = e2/2, then, after the necessary action of zero control till the splitting moment t = 1 of the homogeneous signal, only the choice of v(1>2] = 1 forces the trajectory to reach the lower boundary d = 1/2 of the target set D. The uniqueness of the open-loop control of equation (6.4) guaranteeing the attainment of the trajectory to D at the time t = 2 is also obvious.

We pass to constructing the guaranteeing closed-loop control using the guiding open-loop control package. Note that, in the given example, since the splitting moments of the homogeneous signals for equations (6.3) and (6.4) coincide (this may not be fulfilled in the general case), to find the real initial state of system (6.1), it is sufficient to perform measurements of N (N > 1) trajectories x1(r*),x2(r*),..., xN(r*) of the original SDE at the unique distinguishing moment r* = 1 + e; the zero controls are fed onto equations (6.3) and (6.4) till this time. Using these measurements, we construct the estimates yN(r*) and yN(r*) of the signals ym(r*) and yd(r*) satisfying the relations like (2.3), (2.8), and (4.12):

1 N

yN(r*) = Q(r*)mN(r*), mN(r*) = (r*),

r=1

1 N 2

2

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

yN(r*) = Q2(r*)dN(r*), dN(r*) = —(xr(r*) — mN(r*))

r=1

P (max {\yN(r*) — ym(r*)\ , \y%(r*) — yd(r*)|} < h(N)) = 1 — g(N).

Let us derive the condition providing the determination at the time r* (using the estimates yN (r*) and y?(r*)) of the real initial states of equations (6.3) and (6.4) (respectively, m1 or m2 and d1 or d2) and, consequently, of the initial state of equation (6.1). Actually, taking into account that u1(t) = 0, u2(t) = 0, t £ [0,1+ e], and the form of the observation function Q, we should distinguish the values yN(r*) = ee-(1+s"lm1 and yN(r*) = ee-(1+£^m2, as well as yld(r*) = e2e-2(1+s)d1 and yd(r*) = e2e-2(1+s)d2. Therefore, N must be such that

h(N) < min ^(ee-(1+£)lm1 — m20\)/2, (e2e-2(1+£)d — dg|)/2} . (6.5)

Then, only one of the inequalities \yN(r*) — yN(r*)\ < h(N) or \yN(r*) — yN(r*)\ < h(N) holds with the probability 1 — g(N); the same is valid for the inequalities \yN(r*) — y\(r*)\ < h(N) or \yN(r*) — yd(r*)\ < h(N). In case equation (6.3) starts from the initial state m1, we decide to apply the control u1(t) = 1 on the interval (1 + e, 2); otherwise (from the state m%), the control u1(t) = 0. In the first variant, in view of the time delay in switching the control to optimal, m(2) takes the value not 2/e but 2/e — (e£ — 1)/e; i.e., for small e, m(2) reaches the e-neighborhood of the set M. In the second variant, as a result, we have exactly m(2) = 1. By analogy, we proceed with equation (6.4): if the real initial state is d1, then we apply the control u2(t) = 1 on the interval (1 + e, 2); if d2, then u2(t) = 0. In the first case, d(2) takes the value not 1/2 but 1/2 — (e2e — 1)/2e2; i.e., for small e, d(2) reaches the e-neighborhood of the set D. In the second case, we have exactly d(2) = 1.

Thus, the closed-loop control method with incomplete information described above solves the original e-guidance problem: it guarantees the attaintment of the solution of equation (6.3) to the e-neighborhood of the target set M and the attaintment of the solution of equation (6.4) to the e-neighborhood of the target set D at the time t = 2 with a probability close to 1. The computations by formulas (4.12) and (6.5) showed that N = 103 guarantees the guidance accuracy e = 0.1 with a probability P > 0.96, whereas N = 105 guarantees e = 0.01 with P > 0.996.

REFERENCES

1. Krasovskii N.N., Subbotin A.I. Positional differential games. Moscow: Nauka, 1974. 456 p. [in Russian].

2. Krasovskii N.N. Control of a dynamical system: problem on the minimum of guaranteed result. Moscow: Nauka, 1985. 516 p. [in Russian].

3. Krasovskii N.N., Subbotin A.I. Game-theoretical control problems. New York: Springer, 1988. 517 p.

4. Subbotin A.I., Chentsov A.G. Guarantee optimization in control problems. Moscow: Nauka, 1981. 288 p. [in Russian].

5. Osipov Yu.S. Control packages: an approach to solution of positional control problems with incomplete information // Russ. Math. Surv. 2006. Vol. 61, no. 4. P. 611-661.

6. Kryazhimskii A.V., Osipov Yu.S. On the solvability of problems of guaranteeing control for partially observable linear dynamical systems // Proc. Steklov Inst. Math. 2012. Vol. 277. P. 144-159.

7. Kryazhimskii A.V., Strelkovskii N.V. An open-loop criterion for the solvability of a closed-loop guidance problem with incomplete information. Linear control systems // Proc. Steklov Inst. Math. 2015. Vol. 291, Suppl. 1, P. S103-S120.

8. Chernous'ko F.L., Kolmanovskii V.B. Optimal control under random perturbation. Moscow: Nauka, 1978. 272 p. [in Russian].

9. Pugachev V.S., Sinitsyn I. N. Stochastic differential systems. Moscow: Nauka, 1990. 642 p. [in Russian].

10. Rozenberg V.L. Dynamic restoration of the unknown function in the linear stochastic differential equation // Autom. Remote Control. 2007. Vol. 68, no. 11. P. 1959-1969.

11. 0ksendal B. Stochastic differential equations: an introduction with applications. Berlin: SpringerVerlag, 1985. 408 p.

12. Korolyuk V.S., Portenko N.I., Skorokhod A.V., Turbin A.F. Handbook on probability theory and mathematical statistics. Moscow: Nauka, 1985. 640 p. [in Russian].

13. Johnson N.L., Kotz S., Balakrishnan N. Continuous Univariate Distributions. New York: John Wiley, 1995. Vol. 2. 717 p.

i Надоели баннеры? Вы всегда можете отключить рекламу.