Научная статья на тему 'ROBUST IDENTIFICATION OF AN EXPONENTIAL AUTOREGRESSIVE MODEL'

ROBUST IDENTIFICATION OF AN EXPONENTIAL AUTOREGRESSIVE MODEL Текст научной статьи по специальности «Математика»

CC BY
52
18
i Надоели баннеры? Вы всегда можете отключить рекламу.
Ключевые слова
EXPONENTIAL AUTOREGRESSION / ROBUST ESTIMATE / CONSISTENCY / ASYMPTOTIC NORMALITY / ASYMPTOTIC RELATIVE EFFICIENCY

Аннотация научной статьи по математике, автор научной работы — Goryainov A.V., Goryainov V.B., Khing W.M.

One of the most common nonlinear time series (random processes with discrete time) models is the exponential autoregressive model. In particular, it describes such nonlinear effects as limit cycles, resonant jumps, and dependence of the oscillation frequency on amplitude. When identifying this model, the problem arises of estimating its parameters --- the coefficients of the corresponding autoregressive equation. The most common methods for estimating the parameters of an exponential model are the least squares method and the least absolute deviation method. Both of these methods have a number of disadvantages, to eliminate which the paper proposes an estimation method based on the robust Huber approach. The obtained estimates occupy an intermediate position between the least squares and least absolute deviation estimates. It is assumed that the stochastic sequence is described by the autoregressive equation of the first order, is stationary and ergodic, and the probability distribution of the innovations process of the model is unknown. Unbiased, consistency and asymptotic normality of the proposed estimate are established by computer simulation. Its asymptotic variance was found, which allows to obtain an explicit expression for the relative efficiency of the proposed estimate with respect to the least squares estimate and the least absolute deviation estimate and to calculate this efficiency for the most widespread probability distributions of the innovations sequence of the equation of the autoregressive model

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «ROBUST IDENTIFICATION OF AN EXPONENTIAL AUTOREGRESSIVE MODEL»

UDC 519.234 DOI: 10.18698/1812-3368-2020-4-42-57

ROBUST IDENTIFICATION OF AN EXPONENTIAL AUTOREGRESSIVE MODEL

A.V. Goryainov1 V.B. Goryainov2 W.M. Khing1

agoryainov@gmail.com

vb-goryainov@bmstu.ru

waimyokhing@gmail.com

1 Moscow Aviation Institute (National Research University), Moscow, Russian Federation

2 Bauman Moscow State Technical University, Moscow, Russian Federation

Abstract

One of the most common nonlinear time series (random processes with discrete time) models is the exponential autoregressive model. In particular, it describes such nonlinear effects as limit cycles, resonant jumps, and dependence of the oscillation frequency on amplitude. When identifying this model, the problem arises of estimating its parameters — the coefficients of the corresponding autoregressive equation. The most common methods for estimating the parameters of an exponential model are the least squares method and the least absolute deviation method. Both of these methods have a number of disadvantages, to eliminate which the paper proposes an estimation method based on the robust Huber approach. The obtained estimates occupy an intermediate position between the least squares and least absolute deviation estimates. It is assumed that the stochastic sequence is described by the autoregressive equation of the first order, is stationary and ergodic, and the probability distribution of the innovations process of the model is unknown. Unbiased, consistency and asymptotic normality of the proposed estimate are established by computer simulation. Its asymptotic variance was found, which allows to obtain an explicit expression for the relative efficiency of the proposed estimate with respect to the least squares estimate and the least absolute deviation estimate and to calculate this efficiency for the most widespread probability distributions of the innovations sequence of the equation of the autoregressive model

Keywords

Exponential autoregression, robust estimate, consistency, asymptotic normality, asymptotic relative efficiency

Received 09.10.2019 Accepted 13.12.2019 © Author(s), 2020

Introduction. The exponential autoregression model is one of the most popular of time series models (random processes with discrete time) [1]. The advantage of this model lies in the possibility of obtaining, with its use, a description of the nonlinear effects of a number, in particular, limit cycles, jumps resonance, and the dependence of the oscillation frequency on the amplitude, which is impossible in the framework of the linear autoregressive model. The exponential autoregressive model has proven itself in technology [1], economics [2], climatology [3], oceanology [4], biology [1].

The main task in identifying an exponential autoregressive model is to evaluate its parameters — the coefficients of the autoregressive equation that describes this model. The most common methods for estimating coefficients are the least squares method and the least absolute deviation method. These methods have more than two hundred years of history and are well studied for linear models. In particular, if time series observations are Gaussian (normal) random variables, then the least squares method gives the best results. If observations of the time series were made with large measurement errors, then the least absolute deviation method is more effective. In the second half of the 20th century, the M-estimates method was developed, which includes the advantages of both the least squares method and the least modulus method. For linear models, the M-estimates almost not inferior to the least squares estimate in the Gaussian case and surpasses it even with a small deviation of the probability distribution of observations from the Gaussian. The M-estimate is almost always better than the least absolute deviation estimates, second only to probability distributions with the so-called heavy tails, which is typical for observations obtained with gross measurement errors.

For nonlinear models, a comparison of the above three methods is poorly understood. Separate results were obtained for threshold autoregression and autoregression with random coefficients [5, 6]. In the present work, a comparative study of these methods is carried out by computer simulation by evaluating the parameters of exponential first-order autoregression.

Problem statement. The time series Xt, t = 0, ±1, ±2,..., described by the first-order model of exponential autoregression satisfies the recurrence equation

The coefficients ao, bo, cq of equation (1) are real numbers and are model parameters. The updating process Sf, i = 1,2,..., is a sequence of independent identically distributed random variables with zero expectation function

(1)

Esf = 0 and finite variance Dst-Esf-a2 <<x>. Suppose that model (1) is stationary and ergodic. A sufficient condition for this is, for example, the simultaneous fulfillment of conditions | a |< 1, c > 0 and a2 < oo [7].

Model (1) is an example of smoothing another popular nonlinear model — a threshold autoregressive model [8] of the form

' aoXf-i+et, if | Xf-i |> C; (oo + fcoPQ-i + Si, if |Xf_i|<SC,

Xt=-

where C > 0 is some threshold constant. Indeed, if | Xt~i | it is large, then the

value e-coxt-i is dose to zero; therefore, the right-hand side of (1) is practically indistinguishable from aoXt-i+et. As the | Xt-\ | decreases, the role of the coefficient bo increases.

Consider the problem of estimating the parameters (ao,bo,co) of equation (1) from observations Xi,X2,--.,Xn of a process Xt, satisfying this equation. We define the M-estimates of the coefficients (ao,bo,co) as the

A

minimum point (a, b, c) of the function

g(a, b,c)= XP

t = 2

Xt-

a + be cXt-1 I Xf—i

(2)

where p is the so-called p-function. It is usually assumed that p is an even and downward-convex function. The least squares and the least absolute deviation estimates are a special case of the M-estimates for p(x) = x2 and p(x)=|x|, respectively.

The most common p-functions are [9] p-Huber function

ph(*h ":":"'/' (3)

xif \x\<k; 2k\x\-k2, if \x\>k,

and the p-Tukey function

PH(*) =

1-

/ 2 Л

1-

V V. fc > у

1,

, if \x\<t, if \x\>k.

(4)

Here k e (0, oo) is the timing parameter, the change of which allows to achieve the maximum efficiency of the M-estimates, depending on the specific type of probability distribution density function / of the members st of the updating process.

According to (2), (3) and the curves shown in Fig. 1, the M-estimates with the p-Huber function is a compromise between the least squares estimate and the least absolute deviation estimate. Since Ph(x) it coincides with x2 in the neighborhood (~k,k) of the origin and behaves linearly outside this neighborhood, similar to | x |, the large residuals Fig. 1. p-Huber function:

Xt-(a + be~cXt-i}Xt-i, generated by I-pM^-x2; 3-2k\x\-k>

large errors in the observations (large perturbations et), they affect the target function g(a,b,c) in a linear rather than quadratic manner, thereby reducing the effect of these large errors on the minimum g(a, b, c), and on the accuracy of parameter estimation. The p-Tukey function, which ignores large residuals (larger in magnitude than k), ignores it even more, reduces the contribution of sharply distinguished observations, replacing them with unity.

The purpose of the work is to study the accuracy of the M-estimates with the p-Huber function depending on the probability distribution of the members of the update sequence zt.

Simulation studies of the properties of the M-estimates. At the first stage, the non-bias and consistency of the M-estimates were studied by computer simulation. For definiteness, it was assumed that ao = -0.3, bo = -0.8, cq = 1, the sample size n varied from 100 to 800 in increments of 100. Random values et, t = l,...,n, were modeled N = 1,000 once usingMATLAB random number generator simulating the normal, logistic, and double exponential probability distributions. Based on the generated values 8,1,812,..., e,n, using the recurrence relation (1) with a zero initial condition X0 =0, the realizations xa,xi2,--.,xin, z = 1,2, ...,2V, of the time series xa, ,..., x,«, i = 1,2,..., N, observations were calculated. For each realization of the observations xu, xi2,..., , i = 1,2,..., N,

the realization was found (flj,^»^), an M-estimates of the parameters

(ao, bo, Co), was determined, which was determined as the minimum point of the objective function g(a,b,c) of the form (2). As p-functions in (2), p-functions (3) and (4) were used. The function g(a, b, c) was minimized using the Levenberg — Marquardt optimization algorithm [10], the essence of which is a combination of the Newton method with the gradient descent method.

To study the probabilistic properties of the M-estimates, its expectation

function (Eg, Eb, Ec) was approximated by averaging the realizations (a,-, c.{)

N N л N

by i = l,2,...,N, over the vector (a,b,c)=

¡'=1 i=1 1=1 У

, since

it follows from the law of large numbers that (a,b,c) —» (E a, E b, Ec) at N —> oo,

A

by definition, the non-bias of the estimate means that (Ea,Eb, Ec) = (ao,bo,co).

Thus, under the condition of non-bias, the difference (a,b,c)-(ao,bo,co) in the simulation should take values close to zero.

A

The errors d-b - bo of estimation of the second coordinate bo for n = 100,200,..., 800 the double exponential probability distribution of random variables et and p-function (3) are given in the table. Small values of the 8

A

allow us to conclude that the estimate b is not biased. For estimates of the other two coordinates obtained including with the p-function (4) and for other probability distributions of the updating sequence et, the simulation results are similar. This makes it possible to draw a similar conclusion about the non-bias of M-estimates a and c of parameters, ao and Co, respectively.

The bias 5 and approximation A of the second moment of parameter b estimation for various values of n of the sample size

и 100 200 300 400 500 600 700 800

8 -0.00262 -0.0203 0.00247 -0.00431 0.00828 -0.001832 -0.000646 -0.00227

Ä 0.0802 0.0399 0.0272 0.0198 0.0157 0.0132 0.0112 0.0105

A

The validity of the M-estimates (a, b, c) was verified by approximating the

a A

second moments (E(a-ao),E(b-bo),E(c-co)2) and variances (Da,Db,Dc)

A

of its coordinates. Consistency by definition means that (a, b, c) —»(ao>^o>co )> in probability n -»oo. From the Chebyshev inequality it follows that for

A A

consistency, for example b, it is enough that if a A = E(b-bo)2 —»0 n—>oo,

A

taking into account the established nonbias of b, it is enough if n -» oo the

A

condition Dfe—»0 is fulfilled. According to the law of large numbers, N A _

A = n_1X(&, -bo)2 tends to A when N—» oo, therefore, by the behavior of A, ¡■=i

A

one can judge the validity of the b estimation. The table also shows the values A for n-100,200,...,800. With increasing n values A decrease, approaching

A

zero, which indicates the validity of the assessment b estimation.

Usually, in mathematical statistics, the rate of convergence of a consistent

estimate is proportional n~112. Such estimates are called Vn-consistent. To test

the hypothesis that the rate of convergence of the M-estimates to the estimated parameter is proportional to ri~1/2 the expression n~1/2 as a function of n was approximated according to the table by a polynomial of the first degree. Least squares obtained

>fnA «1.0699 - 0.0032683« + 4.9245 • 10"6 n2 - 2.5734 • 10"9 n3,

which confirms the hypothesis of Db ~ const n~m. Similarly established 4n -consistency of estimates a and c.

At the second stage, the asymptotic normality of the M-estimates was studied by computer simulation. For definiteness, it was assumed that ao = -0.3, bo - -0.8, Co = 1, are random variables of St have a double exponential distribution of probability, the p-function is described (3). Figure 2 shows the histogram constructed from the results of 10,000

A

simulated implementations b. The symmetric and bell-shaped histograms suggest that the probability distribution of a

A

random variable b is close to normal.

Testing the hypothesis of normality

A

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

of the b estimate for 10,000 implemen-tations using the %-test allowed us to accept it at a significance level of 0.001. A similar result was obtained for estimates a and c.

Asymptotic properties of the M-estimates. The exact probability distribution of estimates is very difficult to find and is possible only in the simplest cases. It is usually possible to establish the asymptotic distribution of the estimate, i.e., the distribution to which the probability distribution of the estimate converges (weakly, by distribution) with an unlimited increase in the number of observations. As a rule, such a limiting distribution turns out to be normal, and the estimates in this case are called asymptotically normal. The

results obtained above suggest that the M-estimates (a, b, cj is asymptotically

normal. Below we will justify the asymptotic normality of the M-estimates

(a, b, c) with p-functions (3), (4).

Having expanded the objective function g(a, b, c) of the form (2) according to the Taylor formula at a point (ao, bo,co) up to the second order inclusive, we obtain

2000,— 1800 -1600 -1400 -1200 -1000 -800 -600 -400 -200 -ol—

-2.5 -2.0 -1.5 -1.0 -0.5 0 0.5

A

Fig. 2. Histogram of M-estimate of b

_г-п~Г

.Ihn_

g(a, b, с) = g(a0 ,bo,co) + ATQ + - QTBQ + 8(a, b, c),

2

where

A -

T

1 f6g(ao,bo,CQ) 8g(ao,bo,co) dg(ao,bQ,cQ)^

4n

da

db

8c

-S p'(zt)Xt-i, - X 9\zt)Xt_ie-c^, Zp'i^XlM"0*1

V t=2 t=2 t=2

Q = [y[n(a-ao),y[n(b-bo),y[n(c-co)y, d2g(a0,b0,c0) d2g(ao,bo,co) d2g(a0,b0,c0)^

вЛ

да2 дадЬ дадс

82g(ao,bo,co) 82g(ao,bo,co) d2g(ao,bo,co)

дадЬ 8b2 dbdc

82g(a0,b0,c0) d2g(ao,b0,c0) d2g(ao,bo,co)

dadc dbdc 8c2

5 (a,b,c) is an infinitesimal function of a higher order when (a,b,c)->

-> (flo > bo, Co) compared with (a - oq )2 + (b - bo )2 + (c - Co )2. The matrix B elements have the form

= i P"(s,)XlP ^№£0) = j. p.^,)^^, _

da oaob

t=2

82g(a0,b0,C0)

t=2

= -Zp'bt)X?_1b0e-CoXt-1,

dadc t=2

8bz

t=2

8bdc t=2 t=2

t=2

f=2

We show that there is lim B. For example, find lim — ^

tt-»oo n—>00 M 3c

By the assumption, random sequences sf and Xf are both stationary and ergodic. Therefore, the sequences p"(s^Xf^e'200**-*- and p'(s^X^e-60*'-!

will also be stationary and ergodic [11]. Under the law of large numbers for stationary and ergodic sequences (see [11]), we obtain that with n—»00 probability

- i P"(et)XUble~2c^ E[p"(si)]E(x06b02e-2c«xo ); »t=2 V '

\ t p'i^xlM"0*11 E №)] E (x0V~c°X° ).

From the form of the p-functions defined by formulas (3), (4), it follows that Ep'(ei) = 0. So from independence et from Xt-i it follows:

E [p'fe)] E [x%b0e-c°x° = 0. Thus, with n 00 probability

I d2g(ao,bo,co) ^ £[p„(si)]E/xeh2e-2cQx% \ n 8c1 \ >

Calculating similarly the remaining elements of the matrix lim B, we

n—> 00

obtain that in probability limß - E[p"(si)]K, where

n—>00

E(X2)

Eix2e"c°XoN

-E ( ХцЬое~с°Хо

EiX2e"c°Xo 1 е(хУ2с°хо ) -e(X^b0e~2c°Xo

2\

K =.

J \ J

j -E (x^o<r2c°X° j E [x$by2c°X0 ^

Consequently

g(a, b, c) = g(a0 ,b0,c0) + AT0 + ^ E [p"(si)] QTKQ + 6 (a, b,c) + 6 (a, b, c),

where with n -» 00 5(a, b,c)—> 0 in probability.

Reasoning as in [12], we find the following: the asymptotic distribution of

the M-estimates which is the minimum point of the objective function

g(a,b,c), coincides with the asymptotic distribution of the minimum point

of the function

g(a, b, c) = g{ciQ, bQ, c0)+ ATQ +1 E[p"(si )]QTKQ,

£

representing a quadratic form. It is easy to show that

(a, b, c) - -(E [p"(£l) A

Let us prove the asymptotic normality of the random vector A, from which the asymptotic normality of the random vector (a,b,c) will follow, and

therefore the asymptotic normality of the M-estimates Denote by the

At-1 cr-algebra of events generated by the set of random variables X0,..., Xt-\. As was noted Ep'(£f) = 0, then it follows from (1) that st does not depend on Xt_i. By the definition of a-algebra At-1, a random variable Xt-i is measurable with respect to At-1. Therefore, from the properties of conditional mathematical expectations [13] we have

Elp'feJXi-i | At-1] = Xí_iE[p'(Eí) I A-1] = Xt_iE[p'(et)] = 0.

Given the form of (3), (4) p-functions, we obtain: 0 < E[(p'(ef)Xf_i)2] < oo. Therefore, by the central limit theorem for martingales (for example, [14]) a random sequence

1 %(*o,bo,co)=l£pXet)Xti

yfñ да 4ñ t=2

is asymptotically normal with a expectation function

and variance

limE

n—>00

limD

n-> OO

( 1 n \

-¡-Yp'MXt-i

ynt=2 J

f 1 n ^

-^XP '(et)Xt-i y»t=2 .

Since the random variables st and Xt-i are both independent and Ep'(£f) = 0, so

f t P'(£f)Xf-il = ^ t E[p'(et)Xf_i] = Z Eip'feHEiXf-i] = 0. Vn t=2 y vn t=2 Vn t=2

Since for s < t the random variables bb and dd are independent, and a - b, then for all s < t.

The random variables p'fe) and p'(ss)Xs_iXi_i are independent and Ep'(ei) = 0. Then for all s < t

E[p'(es)Xs_1p'(ei№-i] = E[p'(ei)]E[p'(es)Xs_1Xi_1] = 0. Further, independence of et and Xt_i implies

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

E[(p'(ei)XH)2] = E[p'(ei)2]E[^1] = E[p'(ei)2]E[X§]. So at n —» oo

D

f 1 П \ f л « \2

тЕрЬ№-1 =e vn i=2

VV"i=2

= -Z ¿E[p'(6,)W(Gi)XH]-»- t E[(p'(Bt)XH)2] = E[p'(Bi)2]E[Ag].

n t=2 s=2 n t=2

Thus, the random sequence ]— djKgo>fro>co) asymptotically normal with

Vn da

zero mean and variance E [p'(si )2 ] E [Xq ].

It is similarly proved that the random sequence ]— dff(aO'fy)>co) is

yfn

db

asymptotically normal with zero mean and variance E[p'(ei)2]E^Xoe 2c°xo J,

and the random sequence %(go>fro>co) ^ asymptotically normal with zero

Vn 8c

mean and variance

E[p'(8i)2]E X$b2e~2c°Xo To find the asymptotic covariance matrix of the vector A we get:

r 1 8g(ao,b0,c0) 1 dg(ao,bo,cQ

lim E

/1—>oo

л/п

да

4n

db

lim E

и-»оо

lim E

n-> 00

1 ög(ao>fybCo) 1

л/й

да

4п

дс

= E[P'(81)2]E(X0VCOXO); = -Е [p'(Ei )2 ] Е | X$b0e~c°x° J;

1 dg(ao,bo,co) 1 dg(a0,bo,cQ)

4п

db

4п

de

= -E[P'(81)2]E(X0V-2COX°2).

Thus, the random vector A is asymptotically normal with zero mean and the covariance matrix E[p'(si)2]K. Since = -(E[p"(si)]iC)_1A, then the

covariance matrix of the vector and, therefore, the covariance matrix

ofM-estimates (a,b,c ) is equals (see for example [15])

E[p'(si)2] (E[p"(ei)])2

К

-l

Substituting into this expression p(x) = x2, we obtain the following: since (x2)'-2x and (x2)" - 2, the least squares estimate [a ,b*is asymptotically normal with zero mean and the covariance matrix

E[2(ei)2] , , , , , v ' K'1 = EelK~l = g K .

(E[2]) 1

Let us compare the quality of the M-estimates (a, b, c) and the least-

squares estimate (a,b*,c* Among the two scalar estimates, we will consider

the best one that deviates less from the estimated parameter. Estimates are random; therefore, their deviations from the estimated parameter are also random. It is logical to measure the accuracy of the estimate by expectation of the square of the difference between the estimate and the parameter being estimated, which for an asymptotically unbiased estimate coincides with its asymptotic variance. Thus, it is advisable to compare the accuracy of two scalar asymptotically unbiased estimates by comparing their asymptotic variances. If the estimated parameter is a vector, and its estimates are asymptotically unbiased and have asymptotic covariance matrices proportional to each other, then it is natural to compare the accuracy of these estimates with the ratio of the proportionality coefficients of their asymptotic variances.

Estimates (a,b,c^j and are asymptotically normal with

proportional covariance matrices; therefore, we will compare the accuracy of the estimates (a,b,c) and (a*,£>*,c*) with each other by comparing the quantities

E[p'(si)2]

and a2. The ratio e(f,k) =

(E[p"(si)])2

a2 ct2(E[P"(ei)])2

E[p'(ei)2] E[p'(si)2]

(E[p"(si)])2

will be called the asymptotic relative efficiency of the M-estimate with respect to the least squares estimate. Asymptotic relative efficiency shows how many times more observations are needed to the least squares estimate compared with the number of observations required by the M-estimate to achieve the same accuracy. For example, e = 2 means that to achieve the same accuracy, the M-estimate requires 2 times less observations n, than the least squares estimate.

Asymptotic relative efficiency depends on the type of probability distribution density f(x) of the updating process et. Let us calculate the value e for various distributions of the updating process zt.

If 8f are Gaussian random variables with Est = 0 and Dst = 1, so f(x) = fg(x), where

Ш =

1

e 2

So

E[(p'(si))2] = \ZJ,P\x))2 fg{x)dx = 4 f

' bß -k- ( к к2-Ц^е 2 + (l-fc2)erf

V7I Vv2.

= 4

1c fy -—

к2-!Ц£е 2 +2(l-fc2)Oo(/c)

V7T

E[p"(ei)]= J =

erf2

—oo i—4

f—'

= 4O0(fc); 4Ф 2p(k)

+(i-fc2)erfi-^) k2-^e~k2 +2(1 -к2)Ф0(к) V7I Vv2 У V7t

fc2

* 2 2 * 1 -— Here erf(x) = J-j=e_i di is error function; O0(x) = J ,— e 2 dt is Laplace's

0 v 7T o yl2n

function.

Figure 3 shows the relationship between e(fg,k) and k with a Gaussian distribution Ef. The asymptotic relative efficiency of the M-estimate with increases of k monotonically increases, that is, the effectiveness of the M-estimate increases, at k —> oo approaching the least-squares efficiency. We note that lim e(fg,k) = 1, i.e., k —> oo, when the M-estimate goes over to the

k-> oo

least-squares estimate, the quality of both estimates becomes the same. Also, when k = 0 the M-estimate becomes the least absolute deviation estimate. Since when x —» 0

O0(x) = -erf 2

' x л

Jl) л/2л

+ 0(x3), e 2 -1-х2 +0(x3),

(5)

then, lim e(f„,k) = 2/n. Therefore, with a Gaussian distribution of random fc—>o *

variables £t, the least squares estimate is approximately 1.5 times more effective

than the least absolute deviation estimate. In other words, the accuracy of the least squares estimate constructed from n = 100 the observations will be achieved using the least absolute deviation estimates only based on approximately « = 150 observations.

The assumption that et are Gaussian random variables is usually justified by the central limit theorem of probability theory. However, this theorem is limiting in nature and the assumption that the probability distribution of 8( deviates slightly from the Gaussian one looks more realistic. A typical model for violating the assumption of st Gaussianity is the assumption that it has a contaminated (clogged) Gaussian distribution, or Tukey distribution, with a density [16]

Fig. 3. Relationship between e(fg,k) and к on normal st distribution

1 -— 1 -fT{x) = {l-y)-j=e 2 2t2, o<y<l, t>1.

V 2% v 2ttt

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

In this case, the least squares method usually sharply loses its effectiveness [17].

Estimate e(fr,k). In this case the values CT2=ESf, E[p'(ei)2] and E[p"(gi)] are the following

ct2 = 1 + x2y - y;

E[p'(Bi)2] = 4Ä:2^l + (Y-l)erf^j-Yerf (^L^ + 4<u2yer/ k ^

vW2y

+

f

+ 4(l-y)

erf

2 Л

42) л/л

ьБ -Ч

e 2

Axyk42

yfn

e 2T =

= 4 \k2 + 2( 1 - у)(1 - к2 )Ф0(к) + 27(t2 - к2) Ф0

+

+

ksfeiy-1) — _kvfs[2 ~

а Л

yfn

yfn

21T

E[p"(ei)] = 2(l-7)erf

r h ^

vV2y

+27 erf

r , = 4 (1-у)Ф0(/с) + уФ0 -

k2 V

.2

VV V7T

Using (5), at t —>go get e(fr,k)-Cx2 + O(x), where C>0 is some constant depending on k and y.

Therefore, with increasing t value e(fr,k) the value increases indefinitely for any k>0 and ye(0,1). Thus, the asymptotic relative efficiency of the M-estimates with respect to the least squares estimate for x—*•<» can be arbitrarily large.

When A:—»0 we obtain the asymptotic relative efficiency of the least absolute deviation estimate with respect to the estimate of least squares

which, as is easily seen, coincides with 2/n at x = 1 and y = 0.

Conclusion. Using computer simulation, it was found that the M-estimates of the coefficients of the equation of the exponential autoregression is unbiased, consistent and asymptotically normal. The relationship betweem the asymptotic variance of the M-estimates and the form of the probability distribution density of the updating process of the autoregressive equation is found. The values of the asymptotic variance of the M-estimates are calculated for the main types of density. It is shown that in conditions close to practical, the M-estimates is more efficient than the least squares estimate and the least absolute deviation estimate.

REFERENCES

[1] Ozaki T. Time series modeling of neuroscience data. CRC Press, 2012.

[2] Olugbode M., El-Masry A., Pointon J. Exchange rate and interest rate exposure of UK industries using first-order autoregressive exponential GARCH-in-mean (EGARCH-M) approach. Manch. Sch., 2014, vol. 82, iss. 4, pp. 409-464.

DOI: https://doi.org/10.! 111/manc. 12029

e(fT,0)= lim e(fT,k) =

fc-> 0

2(1 + т2у-у)(т2 -2t2y + 2ут + у2т2 -2у2т + у2)

Т27Г

2( 1 - у + т2у)(ту - т - у)2

Translated by К. Zykova

[3] Gurung B. An exponential autoregressive (EXPAR) model for the forecasting of all India annual rainfall. Mausam, 2015, vol. 66, no. 4, pp. 847-849.

[4] Ghosh H„ Gurung B„ Gupta P. Fitting EXPAR models through the extended Kalman filter. Sankhya B, 2015, vol. 77, no. 1, pp. 27-44.

DOI: https://doi.org/10.1007/sl3571-014-0085-8

[5] Goryainov A.V., Goryainov V.B. M-Estimates of autoregression with random coefficients. Autom. Remote Control, 2018, vol. 79, pp. 1409-1421.

DOI: https://doi.Org/10.1134/S0005117918080040

[6] Goryainov V.B., Goryainova E.R. Comparative analysis of robust and classical methods for estimating the parameters of a threshold autoregression equation. Autom. Remote Control, 2019, vol. 80, no. 4, pp. 666-675.

DOI: https://doi.Org/10.l 134/S0005117919040052

[7] Chan K.S., Tong H. On the use of deterministic Lyapunov function for the ergodicity of stochastic difference equations. Adv. Appl. Probab., 1985, vol. 17, iss. 3, pp. 666-678. DOI: https://doi.org/10.2307/1427125

[8] Tong H. Threshold models in time series analysis 30 years on. Stat. Interface, 2011, vol. 4, no. 2, pp. 107-118. DOI: https://dx.doi.org/10.4310/SII.2011.v4.n2.al

[9] Huber P., Ronchetti E.M. Robust statistics. Wiley, 2009.

[10] Rhinehart RR. Nonlinear regression modeling for engineering applications: modeling, model validation, and enabling design of experiments. Wiley, 2016.

[11] White H. Asymptotic theory for econometricians. Academic Press, 2000.

[12] Goryainov V.B. M-estimates of the spatial autoregression coefficients. Autom. Remote Control, 2012, vol. 73, no. 8, pp. 1371-1379.

DOI: https://doi.Org/10.1134/S0005117912080103

[13] Shiryaev A.N. Veroyatnost' [Probability]. Moscow, Nauka Publ., 2011.

[14] Billingsley P. Convergence of probability measures. Wiley, 1999.

[15] Magnus J.R., Neudecker H. Matrix differential calculus with applications in statistics and econometrics. Wiley, 1999.

[16] Hampel F.R., Ronchetti E.M., Rousseeuw P.J., et al. Robust statistics. Wiley, 2005.

[17] Goryainov V.B., Goryainova E.R. The influence of anomalous observations on the least squares estimate of the parameter of the autoregressive equation with random coefficient. Herald of the Bauman Moscow State Technical University. Series Natural Sciences, 2016, no. 2 (65), pp. 16-24 (in Russ.).

DOI: http://doi.org/10.18698/1812-3368-2016-2-16-24

Goryainov A.V. — Cand. Sc. (Phys.-Math.), Assoc. Professor, Department of Probability Theory and Computer Modeling, Moscow Aviation Institute (National Research University) (Volokolamskoe shosse 4, Moscow, 125993 Russian Federation).

Goryainov V.B. — Dr. Sc. (Phys.-Math.), Professor, Department of Mathematical Simulation, Bauman Moscow State Technical University (2-ya Baumanskaya ul. 5, str. 1, Moscow, 105005 Russian Federation).

Khing W.M. — Post-Graduate Student, Department of Mathematical Simulation, Bauman Moscow State Technical University (2-ya Baumanskaya ul. 5, str. 1, Moscow, 105005 Russian Federation).

Please cite this article as:

Goryainov A.V., Goryainov V.B., Khing W.M. Robust identification of an exponential autoregressive model. Herald of the Bauman Moscow State Technical University, Series Natural Sciences, 2020, no. 4 (91), pp. 42-57. DOI: https://doi.org/10.18698/1812-3368-2020-4-42-57

И.В. Фомин, C.B. Червон, A.H. Морозов

ГРАВИТАЦИОННЫЕ ВОЛНЫ РАННЕЙ ВСЕЛЕННОЙ

В Издательстве МГТУ им. Н.Э. Баумана вышла в свет монография авторов И.В. Фомина, C.B. Червона, А.Н. Морозова

«Гравитационные волны ранней Вселенной»

Рассмотрены применение скалярных полей в космологии и методы построения моделей ранней Вселенной на основе их динамики. Выполнен анализ динамики Вселенной на различных стадиях ее эволюции. Проведен расчет параметров космологических возмущений. Представлены методы верификации инфляционных моделей и новые методы детектирования гравитационных волн.

По вопросам приобретения обращайтесь:

105005, Москва, 2-я Бауманская ул., д. 5, стр. 1

+7 (499) 263-60-45

press@bmstu.ru

https://bmstu.press

i Надоели баннеры? Вы всегда можете отключить рекламу.