Научная статья на тему 'PARAMETER ESTIMATION OF SCALE MUTH DISTRIBUTION(SMD) UNDER TYPE-1 CENSORING USING CLASSICAL AND BAYESIAN APPROACHES'

PARAMETER ESTIMATION OF SCALE MUTH DISTRIBUTION(SMD) UNDER TYPE-1 CENSORING USING CLASSICAL AND BAYESIAN APPROACHES Текст научной статьи по специальности «Математика»

CC BY
52
37
i Надоели баннеры? Вы всегда можете отключить рекламу.
Область наук
Ключевые слова
Scale Muth distribution / type-I censoring / classical approach / Bayesian approach / real data analysis

Аннотация научной статьи по математике, автор научной работы — Agni Saroj, Prashant K. Sonker, Shalini Kumari, Rakesh Ranjan, Mukesh Kumar

The lifetime distributions are used to understand and explain the real life circumstances in various fields (medical, engineering, etc.). Many times it is very tough task to complete an experiment with complete data due to lack of time, money or some other factors and get the data in incomplete form. to draw the information from such type of data (incomplete data), we use some censoring techniques. In the field of statistics, there are several censoring techniques available where type-I censoring is most commonly used due to its simplicity. In this article, the scale Muth distribution (SMD) is considered as a lifetime distribution under type-I censoring scheme. The parameter estimation has been done by classical as well as Bayesian approach. Under the classical paradigm, two most popular methods were used maximum likelihood estimation (MLE) and the maximum product of spacing estimation (MPSE). And under the Bayesian paradigm, we used the informative priors for each parameter and obtained the estimates by considering the squared error loss function using an approximation method, Metropolis Hasting (MH) algorithm. The performance of each estimator is evaluated by their mean squared error or simulated risk. Also, a real data set is used to illustrate the real phenomena and to estimate the parameter using above-mentioned techniques under type-I censoring scheme.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «PARAMETER ESTIMATION OF SCALE MUTH DISTRIBUTION(SMD) UNDER TYPE-1 CENSORING USING CLASSICAL AND BAYESIAN APPROACHES»

RT&A, No 4 (76) Volume 18, December 2023

PARAMETER ESTIMATION OF SCALE MUTH DISTRIBUTION(SMD) UNDER TYPE-1 CENSORING USING CLASSICAL AND BAYESIAN APPROACHES

Agni Saroj1, Prashant K. Sonker2, Shalini Kumari3,Rakesh Ranjan4

1,2,3Department of Statistics, Banaras Hindu University, Varanasi, 221005, India.

4 DST Centre for Interdispinary Mathematical Sciences, Banaras Hindu University,

Varanasi, 221005, India.

5Department of Statistics, MMV, Banaras Hindu University, Varanasi 221005, India. e-mail: [email protected], [email protected], [email protected], [email protected], *[email protected] *Corresponding Author

Abstract

The lifetime distributions are used to understand and explain the real life circumstances in various fields (medical, engineering, etc.). Many times it is very tough task to complete an experiment with complete data due to lack of time, money or some other factors and get the data in incomplete form. to draw the information from such type of data (incomplete data), we use some censoring techniques. In the field of statistics, there are several censoring techniques available where type-I censoring is most commonly used due to its simplicity. In this article, the scale Muth distribution (SMD) is considered as a lifetime distribution under type-I censoring scheme. The parameter estimation has been done by classical as well as Bayesian approach. Under the classical paradigm, two most popular methods were used maximum likelihood estimation (MLE) and the maximum product of spacing estimation (MPSE). And under the Bayesian paradigm, we used the informative priors for each parameter and obtained the estimates by considering the squared error loss function using an approximation method, Metropolis Hasting (MH) algorithm. The performance of each estimator is evaluated by their mean squared error or simulated risk. Also, a real data set is used to illustrate the real phenomena and to estimate the parameter using above-mentioned techniques under type-I censoring scheme.

Keywords: Scale Muth distribution, type-I censoring, classical approach, Bayesian approach, real data analysis.

A new popularized lifetime distribution, Muth distribution was introduced by J. E. Muth [1] in the field of reliability theory. Muth distribution has less probability mass on right tail in

& Mukesh Kumar*5

1. Introduction

comparison of some other well known distribution like Weibull, Gamma and log-normal. The cumulative distribution function (cdf) and probability density function (pdf) of this distribution are defined respectively:

F(y; a) = 1 - e{ay-a(eay-1)} y > 0, a e (0,1] (1)

((eay - a)eixy-a(a-1)} y > 0, a e (0,1]

f (y; a)= < (2)

[0 otherwise

A very few properties of this distribution was given in [1] and mainly focuses on its strict positive memory and mean residual lifetime. After this Rinne [2] used this distribution as lifetime model with German data set of car prices for fitting. In 2008, Leemis and McQueston [3] given a well defined figure and showed the relationship with exponential distribution. When the shape parameter a ^ 0 it converges to standered exponential distribution having parameter 1, and given its name as Muth distribution. In 2015, Jodra et.al. [4] again highlighted this distribution and derived some other statistical properties such as, moment generating function (mgf), Quantile function, rth moments, moments of order statistics , mode, median and given a scale transformation form of this distribution named as SMD with scale parameter ft >0 and shape parameter a e (0,1].

Estimation of parameters of SMD with classical methods has been done in [4]. After that many of authors used this distribution with different transformation and derived their properties and applications such as Jodra et. al. [5] define power Muth distribution, Chesneau and Agiwal [6] given inverse power Muth distribution and observed that there is no restriction to select the values of a in eq.[1] as a < 0 and a e (0,1] but a = 0. In 2022, Saroj et.al. [7] given a new lifetime distribution with up-side down bathtub (UBT) shape hazard, named as inverse Muth distribution(IMD) by taking inverse transformation of Muth distribution, and derived some statistical properties of IMD where it was found that the moment of any order for IMD does not exist for a e (0,1]. Also, the scale transformation was considered from the IMD named as scale inverse Muth distribution with real data applicability [7]. Recently in 2022 Sonker et. al.[8] estimated the parameter under simple stress-strength reliability and multi-component stress strength. In 2018, Almarashi and Elgarhy [9] given a Muth generated (M-G) class of distribution with the help of Muth distribution as a generator and T-X generated family [10]. Then Al-Babtain et. al. introduced a Transmuted M-G class of distribution [11] and Almarashi et. al. introduced a new truncated M-G family of distribution[12]. Some censoring schemes are also used by some authors for different form of Muth distribution such as power Muth distribution under progressive censoring scheme [13] and SMD under type-II censoring scheme [14]. Generally lifetime distributions are used under the life testing experiments, where we get the data in ordered form by failure (or death) of testing units. Many of the times, it is very tough task to record failure time of each event due to lack of time cost effectiveness and many more. In all these cases we need a technique that helps in drawing inferences about population after observing a part of testing units. For this, there are several censoring techniques are available

for different cases and situations. There are many censoring methods available in statistics one of them is type-I censoring scheme which is a simplest censoring technique and the form of right censoring. Under the type-I censoring scheme, we select a time point To to terminate the experiment and record the failure time of the event which fails before To, and after To it assumed that all the remaining items are censored. A very impressive experimental example for real life situation under type-I censoring was given in [15]. For the more details about type-I censoring see [16] & [17].

In this article, classical as well as Bayesian method of estimation are used to estimate the parameter of SMD under type-I censoring scheme. In classical method two most popular method MLE and MPSE are used and Bayesian estimation is performed under squared error loss function (SELF) and an approximation method is used to find the estimates because our posterior is not in simplest form.

2. Classical methods of estimation

2.1. Maximum Likelihood Estimation

In the classical inference, one of the most popular estimation method was proposed by R. A. Fisher in 1920 based on likelihood function called MLE. The main principle of this method is to search for the value of parameter which maximizes the likelihood function. In this article we have to estimate the parameter of SMD under type-I censoring scheme. In this case the likelihood function was defined by Cohen(1965)[18] as:

n

L(6\x) = Cnf (xt)]*[1 - F(To)](1-Si) (3)

i=1

where 5( is an indicator function which indicates that an experimental unit is observed or censored. If To is the pre-assigned time point to terminate the experiment then S( is defined for the Ith; i = 1,2,3,..., n unit as:

= i 1 if Xi < To

i [0 if xi > To

Let us suppose that the failure time of m units are observed from n before time point To, then the

constant C will be defined as , n! m, and the Likelihood function is:

(n-m)!'

m

L(0\x)= nif(Xi)][1 - F(To)](n-m) (4)

i=1

Let x ~ SMD(a, fi) having pdf and cdf as:

ax-if e x-1

F(x; a, ß) = 1 - Aß SV )) x > 0, a e (0,1] (5)

ß (eßx - a)Aß a ( ^ x > a e (0,1 (6)

0 otherwise

then the Likelihood function is expressed as:

L(a, ß\x)

n!

(n - m)!

1, (a x{) ,\f Xi - a (e( a Xi)-1)

ß(rß " - a)eu

1 - <M - e

{f Xi - a (e( ß Xi)-1)}

L(a,ß/x) = cßm n

i=1

(e(ßXi) - a)e

X -1 (e( a X)-1)

Ix - a (e( aXi)-1)

(7)

where C

n!

(n-m)!

. after taking the Log of both side of eq.[7] the log-likelihood defined as

(8)

m / i* \ \ m (a 1 (* \ \ log(L(a, ft/x)) =log(C) - mlog(ft) + £ log (e(Xi) - a) + £ -xt - -(e(Xi) - 1)

i=1 i=1 \ft ' + (n - m)( ft To - a (e( To) - 1))

To obtain the MLE for a and ft we have to search the value of a and ft which maximizes the eq.[8]. now, differentiate the eq.[8] w.r.t. a and ft and equating to 0 we get,

Ja log(L(a, ß/x)) = 0

k Xjeß x - ß + k J x + efXi - 1) XiilM +( _ ) k To + efTo - 1) x{ efTo ¿1 ß(e f xi - a) Mß+ a2 aß j + (n m) k { ß + a2 aß

0

(9)

and,

J log(L(a,ß/x)) = 0

m iaxießX'

m

-ß- ¿1 (elxi - a) ' ß2i=1

+ lx> (eß- - a) + ^^ k (eT - a) = 0

2

i=1

(10)

now solve these two non-linear eq.[9] and eq.[10] to get the value of aand ft with satisfying the condition

'd log(L(a,ft|x))

< 0

(11)

dadft , ......

As we can see that eq.[9] and eq.[10] are the complex function for a and ft i.e. we can not solve both the equations analytically. To obtain the value of a and we used Newton Raphson iteration method and get the the value of a and ft as ami and ftmi which satisfying the condition eq.[11] and maximize the eq.[8].

2.2. Maximum Product of Spacing Estimation

This (MPSE) method is an another method of estimation alternative to MLE which is originally proposed by Cheng and Amin in 1983 [19] and Ranneby in 1984 [20]. It provides the consistent and asymptotically unbiased estimate. This method cover the drawback of MLE where MLE fails to estimate three or more parameter distribution like gamma, weibull and log-normal [21] and also the major limitation of MLE is that it does not work satisfactorily for Heavy tailed continuous distribution with unknown location and scale parameter [22]. The estimators obtained by MPSE

nm

nm

e

method possess most of the large sample properties like sufficiency, consistency and asymptotic efficiency, possessed by MLE under more general condition [[23] & [21]] and MPSE holds the invariance property same as the MLE, shown by Coolen and Newby [24]. The consistency of MPS estimator have been discussed in detailed in [25]. MPS method is mostly similar as MLE where in case of MLE we use the likelihood function and search the value of parameter which maximises the likelihood function but in the case of MPSE method we tried to find the estimated value of parameter which maximises the product of spacing function based on cdf of the target density function.

Let Xi, X2,..., xn, be an ordered random sample of size n from the univariate distribution. For the MPSE, the general condition is that the density function is considered to be strictly positive in an interval (a,b) and zero elsewhere (a = — to and b = to). In this article we have the cdf F(x) and pdf f(x) defined in eq.[5] and [6] respectively with supporting value x > 0;0 < a < 1 and fi > 0. By this F(x) = 0 and f (x) = 0 for x < 0, and F(x) = 1 and f (x) = 0 for x > to. The Ith spacing function is defined as Di = F(xi) — F(xi—1) and the product of spacing function is S = nn+1 Di.

The average of this is measured by geometric mean of spacing denoted by 'G' and defined as

1

G = Sn+1

If the sample is very likely or most probable, it is assumed that they gives the magnitude of spacing more or less equal. Hence, MPSE will be that value of parameter which maximizes the S or G. This study considered under type-I censoring so, the sample values belong in (0,to) can be divided in no. of parts of interval like [0,x1], (x1,x2], (x2, x3]...(xk—1, xk], (x^, To], (To, to). MPSE method described in detail for the type-I censoring case in [14]. According to [14] we may write the product of spacing function for chosen distribution as :

c. . if (To) • n=1 Di • Dm+j}; if lTo — xm | < e

S (a, filx) <x < { , 1 } (12)

\ Ds •{nm=1 Di }•{ n"=i Dm+jj ; if otherwise

where, Di = F(x{) — F(x—), Dm+j = (1 — F(To))/(n — m), = F(To) — F(xm) and F(x), f (x) given in eq.(5), (6) respectively. Thus the MPSE can be obtained by maximizing S (a, filx) = log(S (a,fi|x)). Now differentiating log of eq.(12) w.r.t. a and fi we get,

{f'(To) + ^m F'(xi)—F'(xi—1) (n—m)F'(To) :f \T _ x I < ^

f(To)+ ^i=1 F(xi)—F(xi—1) 1—F(To) if lTo xm 1 < 6

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

(13)

F'(To )—F' (xm ) + srm F' (xi)—F' (xi—1) (n — m)F' (To) other?nise

F(To)— F(xm) + Li=1 F(xi)— F(x—1) 1—F(To) otherwise

where,

—/'(x) = -2fi (a — efi ) {ax(efix — a) — fiefix + fi\ + a2(xefi da a fi '

lF' (x) = ^ [we'x—»>- ^ "f x+fi}

a x

and A1 = e( fi x—1 (e?—11))

And the partial derivative of log of eq.[12] w.r.t. fi will be similar as eq.[13]

{f'(To) + srm F'(xi)—F'(xi—1) (n—m)F'(To) if |T _ x I < 6

f(To)+ ^i=1 F(xi)—F(xi—1) 1—F(To) if lTo xm 1 < 6

F'(To)—F'(xm) + rm F'(xi)—F'(x—1) (n—m)F'(To) o w

F(To)—F(xm) + U=1 F(xi)—F(x—1) 1—F(To) ^

where,

d , A -i r a ^ a ^ a v n

dßf (x) = -ß [- a) + axeßX + x(eßX - a)2

èF'(x) = -ß (eaX - a)

Now equating eq.[13], [14] to zero and solve it for a and /3, we get the estimated value of a and / which maximizes the eq.[12]. But the analytical solution is not possible because those equations are not in closed form. Hence we use Newton Raphson method to solve it numerically and get estimated value as amp and pmp.

3. Asymptotic Confidence Interval

In this section, we define the interval estimation for aand / . Since the exact distribution of MLE and MPSE is not easy to find because of the maximum likelihood and product of spacing function is not in the explicit form. Several authors show that the MPSE method is asymptotically equivalent to the MLE see [[19], [26], [27], [28]]. So, by using the large sample theory we may propose the Asymptotic Confidence Interval based on MLE and MPSE function and we may write the asymptotic distribution of both kind of estimator is:

(6 — 6) = N(0,I—1 (6)); (15)

where,

6 is the estimate of parameter

6 is the true value of parameter

I-1 (6) is the inverse of Fisher information matrix

The 100(1 — a*)%confidence interval is defined as:

± Zah]/Var(d); (16)

where Var(d) = -E ^^gXp j „ is the digonal element of the inverse of Information matrix I 1(9) In our case the information matrix is defined as:

I (a;ß )

I1,1 I1,2 I2,1 I2,2

Here I1,1 = dOi, 12,2 = and h,2 = I2,i = dad/ are the second order derivative of log-likelihood function eq.[8] and the log of the product spacing function eq.[12] . So, for a the 100(1 — a*)% confidence interval is defined as:

a ± Za* Var(a); where a is ami and amp (17)

and for

/3 ± Za / 2\J Var( 3); where 3 is 3 ml and 3 mp (18)

4. Bayes estimates of a and fi under squared error loss function

In this section we considered Bayesian approach to estimate the parameter a and fi under type-I censoring scheme. In the Bayesian method it is assumed that our parameter are random in nature instead of a fixed quantity. As the random nature of the parameter, we can associate it with distribution function such as called prior distribution. Let us consider that gamma prior for scale parameter and beta prior for shape parameter a, as informative prior, therefore

n(a) a aa—1 • (1 — a)b—1 (19)

and

n(fi) a fic—1 • e—dfi (20)

where, a,b and c,d are hyper parameters for beta prior and gamma prior respectively. Thus the joint prior of a and is

n(a, fi) a aa—1 • (1 — a)b—1 • fic—1 • e—dfi (21)

Now, according to the Bayes rule the formula for joint posterior of a and is

n(a P,\x) =_n(a,fi)L(a,fi|x)__(22)

n^filx) j ^(a, fi)L(a, filx)da dfi ( )

where, L(a,fi|x) is likelihood given in eq.[7]. So, combining the eq.[7] and [21] the posterior for a and can be written as:

a t a x.

n(a, filx) a C—1 • fi(c—m—1) • a(a—1) • (1 — a)(b—D • e(n—m)(aT—h(efi —1))—dfi • ^fixi—h(efi '—1)) (23)

where, C1 = f fi(c—m—1) • a(a—1) • (1 — a)M • e(n—m)(fiT—1 (e? o —1))—dfi • eE(axi—b(efi i—1)) da dfi In the case of Bayesian techniques, loss function play very important role. In this article, to evaluate the Bayes estimator of a and we use a symmetric loss function, squared error loss function (SELF) which gives the equal weight for under and over estimation problem. The general form of SELF can be expressed as:

L(§, 9)= X(S — 9)2; A > 0 (24)

where 9 is the estimate of parameter 9. the Bayes estimator of parameter under SELF is posterior mean i.e. posterior expectation E9 [9 |x]. In our case it is found that the posterior mean for both parameter is the ratio of two mathematically non-tractable integrals; so we use the Markov Chain Monte Carlo (MCMC) with Metropolis-Hasting (MH) algorithm to obtain Bayes estimate. MH algorithm is generalization form of basic algorithm originated by Metropolis and Ulman(1949) [29], then Metropolis et. al. (1953) used first time in statistics study the equation of a state of a two-dimensional rigid sphere system. In 1970, W. K. Hasting gives the original method by using arbitrary proposal distribution and it popularized by name MH algorithm. For more detail see[[30], [31] & [32]]. By using this technique we can generate the sample from posterior eq.[23] and evaluate the Bayes estimate. The following iterative steps are involved in MH algorithm: 1. Set an initial guess value 90.

2. By using previous point Generate the next value of 6 as 6> from the proposal density p* (6j |6j—1). Here we take normal distribution as the proposal density for parameter.

3. Then we generate u U(0,1)and calculate the ratio:

p(6j|x) ■ p*(6j—1|6j)

r

4. Set

p(Qj-1 \x) • p*(Qj\Qj-1)

p(e '\x)-p*(Q '-1\Q')

6j = { 6 if U - P(6'—1\X)-p* (9!\9!—1)

[6j—1 otherwise

5. Repeat step (2-4) for j=1,2,3,...,M and obtain 61,62,...,6m

6. To find Bayes estimate under SELF, we take

ep m = (e

M - M0 Mo+1

where Mo is burn-in period of Markove chain.

5. Highest Posterior density interval

Nowadays, in the field of Bayesian inference is considered that the parameter of interest behaves like a random variable, then what is probability that the parameter (say 6) lies within a specified interval. Edwards et al. (1963) [33] suggested a method to calculate this interval and named it credible interval. By using this many of the statistician summarizes the marginal posterior density by 100(1 — a*)% posterior credible intervals for the parameters. These intervals can be easily calculated by analytically or MCMC simulation technique. The shortest interval among all of the Bayesian credible intervals is called highest posterior density (HPD) interval. The HPD intervals for parameter can be obtained by using method explained by Chen and Shao (1999)[34] on the basis of MCMC sample from posterior density in Eq.[23]. There are also mentioned that when a marginal distribution is not symmetric, a HPD interval is more desirable. Box and Tiao (1973)[35] was already discused that the HPD interval has two main properties as: (i) the density for the points which are lie within the interval is more than that for the points outside the interval, and (ii) for a given probability say, (1 -a*), the interval is of the shortest length. If we have the posterior as p(6|x) then the credible interval for 6 is defined as:

(6[(a/ 2)M], 6[1—(a/ 2)M] )

and the HPD interval is defined as:

(6j*, 6j* + [1—(a /2)])

where, j* is chosen by which length lj* = min(6j+[1—(a/2)] — 6 j) for 1 — j — M — [(1 — a)M]

6. Simulation Study

Under this section, we study about the behavior of the estimated value of parameters using the simulation study and compare the performances of MLE, MPSE and Bayes estimator on the basis

of their mean squared error (MSE) values. To compute the value of MSE, we use the formula as; MSE(9)=N (9 — 9)2, where N is the number of simulated random sample (we choose N=10000), 9 is the true value of parameter and 9 is the estimated value of the parameter for ith simulated random sample. all the computation was done by R-software like sample generation from the considered distribution and finding the estimated value. To generate the random sample, we considered the following steps:

Step 1. Provide some selected values of the parameters, a and fi, and sample size n.

Step 2. Then generate a random sample from a standard uniform distribution U(0,1), which also

be size n denoted as Ui; i — 1,2,3..., n

Step 3. Now by using the quantile function of SMD given in [4], obtain uth quantile (random

sample) from the SMD xi.

Step 4. Repeat steps 2 and 3, for each i = 1, 2, ... , n we get a random sample of size n from the SMD.

In step 3 the Lambert-W function W—1 (.) can be computed using R-command lambertWm1() from 'lamW ' package, see Adler [2]. In the simulation study, we choose the different combination of parameter values as (a = 0.3, fi = 0.5), (a = 0.5, fi = 0.5), (a = 0.7, fi = 0.5), (a = 0.3, fi = 2.5), (a = 0.5, fi = 2.5), (a = 0.7, fi = 2.5), and observed the behavior of MSE by varying sample size. For this we choose the different sample sizes as n = (30, 50, 70). In case of Bayesian estimate we need the value of hyper parameters for which we take prior means equal to true value of parameter and the variances can be chosen as, for small value of parameter (a=(0.3,0.5) & fi=0.5) we take small variance for gamma and beta prior and for the large value of parameter (a=0.7 & fi=2.5) we take some large variance for the both prior. An another issue in this simulation study is the choice of To. We take To = (0.5,0.9) for fi=0.5 with different value of a, and To = (3,5) for fi=2.5 with different value of a. On observing the simulation table [1-6] we found that:

In general, as sample size increases the MSE of all estimators decreases. This shows that the estimators are consistent. And this decreasing nature of MSE is also found with increasing To. The average length of asymptotic confidence interval as well as HPD interval decreases as sample size increases.

MSE of Bayes estimates is least in comparison of MLE and MPSE both, and the MSE of MPSE is less that MLE.

Width of the HPD interval is smaller in comparison of width of asymptotic confidence interval and the width of asymptotic confidence interval for MPSE is less than the width for asymptotic confidence interval MLE.

Table 1: Estimates of the parameters, MSE, CI and HPD interval for a=0.3, /=0.5

MLE

To Est.

MSE

CI Length

MPS

Est.

MSE

CI Length

Est.

Bayes "MSE

HPDI Length

30

0.5 0.34704 0.9 0.32891

0.07774

0.04485

0.72461, 0.06792

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

0.65669 0.76512, 0.09032 0.67480

50

0.5 0.32582 0.9 0.31201

0.05735

0.72646, 0.08661

0.63985 0.72890, 0.10531 0.62359

70

0.5 0.31557 0.9 0.29997

0.04712

0.02274

0.71985, 0.09472

0.62513 0.68941, 0.11159 0.57782

0.33273 0.31103

0.06535

0.03862

0.72461, 0.06792

0.65669 0.75007, 0.08840 0.66167

0.31781 0.30214

0.05174

0.02796

0.71290, 0.09112

0.62179 0.71555, 0.10427 0.61129

0.31017 0.29338

0.04380

0.02154

0.70609, 0.09875

0.60734 0.67894, 0.11067 0.56827

0.49504 0.46616

0.04156

0.03210

0.73352, 0.25224

0.48128 0.68793, 0.24349 0.44444

0.47571 0.43824

0.03562

0.69921, 0.24710

0.45211 0.64006, 0.23556 0.40450

0.46208 0.41580

0.03090

0.01768

0.67481, 0.24393

0.43088 0.60281, 0.22852 0.37429

ß = 0.5

0.5 0.51621 0.9 0.50952

0.01738

0.00725

0.80992, 0.36543

0.44448 0.67859, 0.38679 0.29180

50

70

0.5 0.51008

0.9 0.50562

0.5 0.50932

0.9 0.50449

0.01114

0.00415

0.00286

0.73543, 0.37591

0.35951 0.63644, 0.40359

0.23284 0.70286, 0.38420

0.31866 0.61525, 0.41471 0.20054

0.50410 0.50958

0.01401

0.00640

0.76774, 0.36775

0.39998 0.67060, 0.39217 0.27843

0.50452 0.50584 0.50536 0.50428

0.00943

0.00366

0.00674

0.00271

0.93305, 0.38101

0.55204 0.63370, 0.40661

0.22708 0.72334, 0.38939

0.33395 0.61537, 0.41601 0.19936

0.50820 0.52538

0.01338

0.00699

0.69941, 0.35175

0.34765 0.67103, 0.39446 0.27657

0.49198 0.00676

0.51580 0.00388

0.48843 0.00462

0.51197 0.00259

0.63733, 0.36875

0.26858 0.62735, 0.41243

0.21492 0.61332, 0.38087

0.23245 0.60662, 0.42325 0.18337

Table 2: Estimates of the parameters, MSE, CI and HPD interval for a=0.5, /=0.5

MLE

n To Est.

MSE

CI Length

MPS

Est.

MSE

CI Length

Bayes

Est.

MSE

HPDI Length

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

a = 0.5

0.5 0.48074 0.9 0.50786

0.07860

0.04226

0.83626, 0.10341

0.73285 0.85294, 0.17571 0.67723

50

0.5 0.49154

0.9 0.50121

0.9 0.48544

0.9 0.50248

0.05688

0.02726

0.04791

0.01965

0.84875, 0.15769

0.69106 0.79694, 0.23084

0.56610 0.83548, 0.18795

0.64754 0.75472, 0.26570 0.48902

0.45601 0.47783

0.06665

0.03675

0.82580, 0.11771

0.70808 0.83123, 0.17776 0.65346

0.47462 0.48342 0.47320 0.48975

0.05084

0.02516

0.04426

0.01847

0.83208, 0.16627

0.66582 0.77794, 0.22849

0.54944 0.81647, 0.19507

0.62140 0.73761, 0.26445 0.47316

0.52302 0.52332

0.00395

0.00517

0.76440, 0.27504

0.48936 0.74687, 0.29615 0.45072

0.52207 0.51618 0.51835 0.51345

0.00485

0.00582

0.00532

0.00587

0.75113, 0.28546

0.46566 0.71989, 0.30895

0.41094 0.73660, 0.29214

0.44445 0.70190, 0.32155 0.38035

ß = 0.5

30

50

0.5 0.53008

0.9 0.50709

0.5 0.52053

0.9 0.50307

0.01885

0.00447

0.01047

0.00248

0.87626, 0.36571

0.51055 0.65383, 0.39495

0.25888 0.75021, 0.38034

0.36987 0.61212, 0.41396 0.19816

0.5 0.51661 0.9 0.50260

0.00179

0.70762, 0.39148

0.31614 0.59260, 0.42653 0.16608

0.51941 0.50638 0.51528 0.50265

0.01454

0.00415

0.00238

0.88980, 0.36510

0.52470 0.65251, 0.39464

0.25787 0.73297, 0.38130

0.35167 0.71388, 0.41340 0.30048

0.51257 0.50227

0.00643

0.00173

1.18937, 0.39147

0.79790 0.59251, 0.42607 0.16643

0.53692 0.51898 0.52322 0.51130

0.01431

0.00463

0.00794

0.00250

0.74564, 0.36762

0.37801 0.65208, 0.39763

0.25445 0.67951, 0.39025

0.28926 0.61151, 0.41741 0.19410

0.51562 0.50890

0.00486

0.00179

0.64536, 0.40269

0.24267 0.59275, 0.42927 0.16348

n

a = 0.3

30

30

70

70

Table 3: Estimate of the parameters, MSE, CI and HPD interval for a=0.7, fi=0.5

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

To Est.

MLE

"MSE CI

Length

Est.

MPS

"MSE CI

Length

Est.

Bayes

"MSE CI

Length

30

50

70

0.5 0.63980 0.9 0.50786

0.06248

0.04226

0.92610, 0.14204

0.78406 0.91664, 0.26414 0.65250

0.5 0.68032 0.9 0.70544

0.01720

0.92746, 0.24556

0.68190 0.89096, 0.37177 0.51919

0.9 0.69356 0.9 0.70790

0.02819

0.01276

0.91520, 0.31706

0.59813 0.87322, 0.43324

0.59994 0.64827

0.05700 0.02526

0.91185, 0.17602

0.73583 0.89101, 0.28691 0.60410

0.65050 0.67673

0.03410 0.01564

0.90945, 0.27204

0.63741 0.87044, 0.38060 0.4

0.67052 0.68693

0.02584 0.01162

0.89762, 0.33618

0.56143 0.85563, 0.43526 0.42037

0.55456 0.58975

0.01594

0.79820, 0.30191

0.49629 0.80713, 0.36497 0.44216

0.57610 0.61656

0.01872

0.01144

0.80553, 0.33467

0.47086 0.80940, 0.41673 0.39268

0.59264 0.63296

0.01592

0.80882, 0.36402

0.44481 0.80789, 0.45184 0.35605

ß = 0.5

50

70

0.5 0.53196 0.9 0.50076

0.01545

0.93145, 0.37106

0.56039 0.61889, 0.40570 0.21319

0.5 0.51477

0.9 0.50108

0.5 0.51100

0.9 0.50143

0.00624

0.00166

0.00391

0.00115

0.69256, 0.39453

0.29802 0.58866, 0.42670

0.16196 0.64484, 0.41075

0.23409 0.57417, 0.43798 0.13619

0.52256 0.49952

0.01226 0.00257

0.80758, 0.36741

0.44017 0.61928, 0.40339 0.21588

0.51045 0.49997 0.50826 0.50052

0.00563 0.00158 0.00365 0.00110

0.79455, 0.39222

0.40234 0.60593, 0.42482 0.18110 0.64154, 0.40867

0.23287 0.57392, 0.43666 0.13726

0.56326 0.50111

0.01757

0.78578, 0.38201

0.40377 0.61721, 0.39407 0.22314

0.54068 0.00797

0.49808 0.00166

0.53239 0.00491

0.49782 0.00109

0.70099, 0.40394

0.29705 0.58369, 0.41675

0.16695 0.66268, 0.41861

0.24407 0.56874, 0.42956 0.13918

Table 4: Estimates of the parameters, MSE, CI and HPD interval for a=0.3, fi=2.5

MLE

To Est.

MSE

CI Length

MPS

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Est.

MSE

CI Length

Bayes

Est.

MSE

HPDI Length

3 0.33908 5 0.33122

0.06482

0.04056

0.74655, 0.07368

0.67287 0.76794, 0.09554 0.67240

50

3 0.32277 5 0.31184

0.04611

0.02651

0.74687, 0.09224

0.65463 0.71418, 0.10936 0.60482

3 0.31052 5 0.30539

0.03601

0.01948

0.73439, 0.09710

0.63729 0.66915, 0.11946 0.54969

0.32510 0.31172

0.05521

0.03484

0.73790, 0.07829

0.65961 0.75338, 0.09133 0.66206

0.31534 0.30053

0.04189

0.02438

0.73531, 0.09535

0.63996 0.70142, 0.10662 0.59480

0.30563 0.29735

0.03362

0.01840

0.72240, 0.10037 0.62202 0.65943, 0.11694 0.54248

0.48950 0.04125 0.46244 0.03133

0.79820, 0.30191

0.49629 0.67978, 0.24478 0.43500

0.46766 0.43199

0.03350

0.02197

0.68422, 0.24752

0.43670 0.62896, 0.23485 0.39411

0.45009 0.41117

0.01666

0.65550, 0.24180

0.41370 0.59270, 0.22909 0.36360

ß = 2.5

30

50

70

3 2.56213

5 2.52823

3 2.53769

5 2.52399

3 2.54392

5 2.51170

0.33418

0.18019

0.21201

0.09420

0.14758

0.07450

3.79115, 1.82942

1.96174 3.34280, 1.92556

1.41723 3.47349, 1.90637

1.56711 3.15481, 2.02575

1.12907 3.34860, 1.96835

1.38025 3.04064, 2.07754 0.96310

2.53662 2.53977 2.52595 2.52977 2.53228 2.51796

0.25619

0.14785

0.16525

0.12293

0.06407

3.64655, 1.85827

1.78828 3.33731, 1.94855

1.38876 4.80067, 1.93116

2.86951 3.15171, 2.03905 1.11266 3.41558, 1.98083

1.43475 3.04172, 2.08817 0.95355

2.48647 2.58703 2.47033 2.57688 2.47440 2.55707

0.13565

0.12178

0.10858

0.07407

0.07359

0.06053

3.18897, 1.86317

1.32580 3.19483, 2.02135

1.17349 3.04908, 1.94862

1.10045 3.07648, 2.10567

0.97081 2.98561, 2.01011

0.97550 2.99013, 2.14544 0.84470

n

a = 0.7

30

n

30

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

a = 0.3

70

Table 5: Estimates of the parameters, MSE, CI and HPD interval for a=0.5, /=2.5

MLE

To Est.

MSE

CI Length

MPS

Est.

MSE

CI Length

Est.

Bayes

TMS!

HPDI Length

a = 0.5

3 0.49063 5 0.51364

0.06249

0.03845

0.86146, 0.12443

0.73703 0.84534, 0.18898 0.65635

50

3 0.49569 0.04508

5 0.50622 0.02396

3 0.49200 0.03487

5 0.50179 0.01767

0.84606 , 0.18054

0.66552 0.78083, 0.24494

0.53589 0.82882, 0.21163

0.61719 0.73912, 0.27530 0.46382

0.46567 0.48145

0.05350

0.03331

0.84526, 0.13707

0.70819 0.82155, 0.18757 0.63398

0.47960 0.48643 0.48054 0.48773

0.04058

0.02201

0.03235

0.01671

0.82739, 0.18762

0.63978 0.75979, 0.24123

0.51856 0.81042, 0.21597

0.59444 0.72061, 0.27285 0.44775

0.52890 0.52551

0.00457

0.00554

0.76369, 0.28798

0.47572 0.74427, 0.30314 0.44113

0.52462 0.51811 0.51908 0.51249

0.00526

0.00584

0.00542

0.00587

0.74473, 0.29795

0.44678 0.71663, 0.31660

0.40003 0.72744, 0.30437

0.42307 0.69522, 0.32677 0.36845

ß = 2.5

30

50

3 2.60200

5 2.52288

3 2.56577

5 2.51464

0.27134

0.10553

0.14969

0.05957

3.90702, 1.86206

2.04496 3.22612, 1.97815

1.24797 3.42124, 1.96766

1.45359 3.03958, 2.08200 0.95758

70

3 2.55300 5 2.51102

0.09713

0.04241

3.25062, 2.02987

1.22075 2.94791, 2.13973 0.80818

2.56793 2.52547 2.54752 2.51666

0.22632

0.09680

0.13165

0.05730

4.41924, 1.86555

2.55369 3.23023, 1.97960

1.25063 3.35205, 1.97103

1.38102 3.04707, 2.08138 0.96569

2.53941 2.51263

0.08819

0.04122

5.83190, 2.02799

3.80391 2.95148, 2.14001 0.81147

2.56169 2.55712 2.54999 2.54589

0.11156

0.07207

0.07483

0.04696

3.26219, 1.93257

1.32962 3.11970, 2.02915

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

1.09055 3.12237, 2.03016

1.09221 2.99657, 2.11663 0.87995

2.54610 2.53840

0.05616

0.03553

3.04397, 2.08901

0.95496 2.92507, 2.16716 0.75791

Table 6: Estimates of the parameters, MSE, CI and HPD interval for a=0.7, /=2.5

MLE

To Est.

MSE

CI Length

MPS

Est.

MSE

CI Length

Bayes

Est.

MSE

HPDI Length

50

3 0.66784 0.04386

5 0.69938 0.02430

3 0.68838 0.02839

5 0.70956 0.01542

0.93602, 0.18502

0.75101 0.91249, 0.27819

0.63430 0.91716, 0.29125

0.62591 0.88856, 0.38549 0.50307

3 0.69760 5 0.71161

0.01970

0.01148

0.89736, 0.36228

0.53508 0.87161, 0.44849 0.42313

0.62763 0.65117 0.66076 0.67790

0.04023

0.02269

0.02607

0.01394

0.91603, 0.21852

0.69750 0.88475, 0.29866

0.58609 0.89869, 0.31178

0.58691 0.86624, 0.39026 0.47598

0.67665 0.68855

0.01817

0.01042

0.88071, 0.37265

0.50806 0.85369, 0.44483 0

0.57242 0.59734 0.59256 0.62341

0.01912

0.01431

0.01550

0.01003

0.80563, 0.32935

0.47628 0.80928, 0.37808

0.43120 0.80859, 0.36665

0.44193 0.81088, 0.42952 0.38137

0.60782 0.63988

0.01306

0.00781

0.80874, 0.39669

0.41204 0.80919, 0.46494 0.34425

= 2.5

30

50

70

3 2.55373

5 2.49284

3 2.53912

5 2.50142

3 2.52649

5 2.50640

0.14866

0.06864

0.07923

0.04142

0.05159

0.02923

3.46516, 1.92668

1.53848 3.06494, 2.02924

1.03570 3.14974, 2.06092 1.08882 2.92997, 2.13616

0.79381 3.00072, 2.13151

0.86921 2.86355, 2.19413 0.66942

2.52432 2.49091 2.52252 2.49971 2.51528 2.50501

0.13224

0.06396

0.07473

0.04996

0.02801

3.39144, 1.90963

1.48181 3.07687, 2.01978

1.05709 3.13833, 2.04526

1.09306 2.93508, 2.12985

0.80523 2.99912, 2.11911 0.88001 2.86692, 2.18957 0.67735

2.57792 2.47366 2.57224 2.47471 2.56087 2.48004

0.08864

0.04880

0.06356

0.03182

0.04968

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

0.02321

3.25633, 1.96686

1.28947 2.97784, 1.99879

0.97905 3.11481, 2.07595

1.03886 2.86599, 2.09898

0.76701 3.02008, 2.13653

0.88355 2.81090, 2.15919 0.65172

n

30

70

n

30

a = 0.7

70

7. Real Data Analysis

In this section, we use a real data set to illustrate the applicability of the proposed work in real life situation. The data-set recently considered in [4], which is taken from the website of the Bureau of Meteorology of the Australian Government (www.bom.gove.au). It represent the monthely rainfall (in mm) between the period of January 2000 to February 2007 in the rain guege station of Carrol, located in the State of New South Wales on the east coast of Australia. The data are given as:

Table 7: Carrol Data-set (n=83)

12 22.7 75.5 28.6 65.8 39.4 33.1 84 41.6 62.3 52.5 13.9 15.4 31.9

32.5 37.7 9.5 49.9 31.8 32.2 50.2 55.8 20.4 5.9 10.1 44.5 19.7 6.4

29.2 42.5 19.4 23.8 55.2 7.7 0.8 6.7 4.8 73.8 5.1 7.6 25.7 50.7

59.7 57.2 29.7 32 24.5 71.6 15 17.7 8.2 23.8 46.3 36.5 55.2 37.2

33.9 53.9 51.6 17.3 85.7 6.6 4.7 1.8 98.7 62.8 59 76.1 67.9 73.7

27.2 39.5 6.9 14 3 41.6 49.5 11.2 17.9 12.7 0.8 21.1 24.5

To ensure that this data-set is appropriate for the illustration of proposed work, first we draw the Total Time on Test (TTT) plot Fig.[1(a)], which shows that the Hazard rate of considered data-set same as the considered distribution (increasing Hazard rate). The Kolmogorov-Smirnov (K-S) test has been used to verify that this data fitted or not on the considered distribution, and found that the the value of K-S statistic is 0.057005 and p-value is 0.9502 which were shows that the considered data fitted on SMD. Graphical representation of K-S test is shown in Fig[1(b)].

(a) TTT Plot Carrol Data-set (b) empirical cdf and fitted cdfplot

Figure 1: (a) plot shows the increasing hazard rate and (b) show fitted cdf of real data on SMD

In [4] it is already shows that SMD gives best fit for this data-set in comparison of some other lifetime distribution (Weibull, Gamma, log-Normal and exponential distribution), on basis of their AIC and BIC values shown in table[8]. The MLE, MPSE and Bayes estimate of the parameter for this data-set in case of type-1 censoring given in table[9].

Table 8: MLE, AIC and BIC values for the Carrol Data-set

Model MLE AIC BIC

SMD (a, ß) a = 0.4608,/3 = 33.9049 740.3600 741.5220

Weibull (a, b) a = 1.3665, b = 36.9120 744.4891 745.6511

Gamma(a, b) a = 1.5160, b = 22.3838 747.3087 748.4706

Exponential (A) A = 0.0295 753.0520 753.6330

Log-Normal (p, a) p = 3.1597, a = 1.0253 767.1983 768.3602

(a) traceplot and marginal density plot for a

(b) traceplot and marginal density plot for ß Figure 2: Shows that MCMC traceplot and marginal density plot with histogram

Table 9: Estimate of parameter for Carrol Data-set

MLE MPSE Bayes

To m est CI est CI est HPD

25 36 a 0.4079 0.8835, 0.0589 0.8246 0.3912 0.8807, 0.0529 0.8278 0.4328 0.7031, 0.1414 0.5617

ß 33.4852 51.7903, 21.6500 30.1403 33.2816 51.4880, 21.5130 29.9749 33.4929 37.3285, 29.2367 8.0918

50 60 a 0.3060 0.7129, 0.0726 0.6403 0.2985 0.7081, 0.0694 0.6387 0.3565 0.6010, 0.1126 0.4884

ß 36.3427 45.1023, 29.2843 15.8180 36.2114 44.9145, 29.1947 15.7197 36.4217 40.0513, 32.6227 7.4286

75 78 a 0.4266 0.6486 , 0.2307 0.4179 0.4152 0.6382 , 0.2223 0.4160 0.4254 0.6131, 0.2169 0.3962

ß 34.1870 39.8435, 29.3336 10.5099 34.2052 39.8796, 29.3382 10.5415 34.3796 37.6232, 31.0938 6.5295

8. Conclusion

In this article, the considered model SMD is another form of Muth distribution by adding a scale parameter. To estimate the parameters of this distribution, we used classical and Bayesian approach under type-1 censoring scheme. In classical approach MLE and MPSE are obtain for time censored data. For the Bayes estimate of the parameters, beta and gamma prior are considered for shape parameter a and scale parameter fi respectively. The simulation study was also done by using Monte Carlo method of simulation. From the simulation study it is found that the Bayes estimate perform better than classical method (MLE and MPSE) on the basis of their MSE and length of interval estimation.

References

[1] Muth, E. J. (1977). Reliability models with positive memory derived from the mean residual life function. Theory and applications of reliability, 2, 401-436.

[2] Rinne, H. (1981). Estimating the lifetime distribution of private motor-cars using prices of used cars: The Teissier model. Statistiks Zwischen Theorie und Praxis, 172-184.

[3] Leemis, L. M., & McQueston, J. T. (2008). Univariate distribution relationships. The American Statistician, 62(1), 45-53.

[4] Jodra, P., Jimenez-Gamero, M. D., & Alba-Fernandez, M. V. (2015). On the Muth distribution. Mathematical Modelling and Analysis, 20(3), 291-310.

[5] Jodra, P., Gomez, H. W., Jimenez-Gamero, M. D., & Alba-Fernandez, M. V. (2017). The power Muth distribution. Mathematical Modelling and Analysis, 22(2), 186-201.

[6] Chesneau, C., & Agiwal, V. (2021). Statistical theory and practice of the inverse power Muth distribution. Journal of Computational Mathematics and Data Science, 1, 100004.

[7] Saroj, A., Sonker, P. K., & Kumar, M. (2022). Statistical properties and application of a transformed lifetime distribution: inverse muth distribution. Reliability: Theory & Applications, 17(1 (67)), 178-193.

[8] Sonker, P. K., Kumar, M., & Saroj, A. (2023). Stress-strength reliability models on power-Muth distribution. International Journal of System Assurance Engineering and Management, 14(Suppl 1), 173-195.

[9] Almarashi, A. M., & Elgarhy, M. (2018). A new muth generated family of distributions with applications. J. Nonlinear Sci. Appl, 11, 1171-1184.

[10] Alzaatreh, A., Lee, C., & Famoye, F. (2013). A new method for generating families of continuous distributions. Metron, 71(1), 63-79.

[11] Al-Babtain, Abdulhakim A., Ibrahim Elbatal, Christophe Chesneau, and Farrukh Jamal. "The transmuted Muth generated class of distributions with applications." Symmetry 12, no. 10 (2020): 1677.

[12] Almarashi, A. M., Jamal, F., Chesneau, C., & Elgarhy, M. (2021). A new truncated muth generated family of distributions with applications. Complexity, 2021, 1-14.

[13] Kohansal, A., & Bakouch, H. S. (2022). Estimation and prediction on power Muth distribution with progressive censored data: a Bayesian approach. Communications in Statistics-Simulation and Computation, 1-23.

[14] BICER, H. D., & OZTURKER, B. (2021). Estimation procedures on Type-II censored data from a scaled Muth distribution. Sigma Journal of Engineering and Natural Sciences, 39(2), 148-158.

[15] Basu, S., Singh, S. K., & Singh, U. (2017). Parameter estimation of inverse Lindley distribution for Type-I censored data. Computational Statistics, 32, 367-385.

[16] Epstein, B. (1954). Truncated life tests in the exponential case. The Annals of Mathematical Statistics, 555-564.

[17] Sinha S (1986) Reliability and life testing. Wiley Eastern Limited. https://books.google.co.in/books? id=esWaAAAACAAJ

[18] Cohen, A. C. (1965). Maximum likelihood estimation in the Weibull distribution based on complete and on censored samples. Technometrics, 7(4), 579-588.

[19] Cheng, R. C. H., & Amin, N. A. K. (1983). Estimating parameters in continuous univari-ate distributions with a shifted origin. Journal of the Royal Statistical Society: Series B (Methodological), 45(3), 394-403.

[20] Ranneby, B. (1984). The maximum spacing method. An estimation method related to the maximum likelihood method. Scandinavian Journal of Statistics, 93-112.

[21] Cheng, R. C. H., & Traylor, L. (1995). Non?regular maximum likelihood problems. Journal of the Royal Statistical Society: Series B (Methodological), 57(1), 3-24.

[22] Pitman EJ (1979) Some basic theory for statistical inference, vol 7. Chapman and Hall, London

[23] Cheng, R. C. H., & Iles, T. C. (1987). Corrected maximum likelihood in non?regular problems. Journal of the Royal Statistical Society: Series B (Methodological), 49(1), 95-101.

[24] Coolen, F. P. A., & Newby, M. J. (1990). A note on the use of the product of spacings in Bayesian inference. Department of Mathematics and Computing Science, University of Technology.

[25] Shao, Y., & Hahn, M. G. (1999). Maximum product of spacings method: a unified formulation with illustration of strong consistency. Illinois Journal of Mathematics, 43(3), 489-499.

[26] Ghosh, K., & Jammalamadaka, S. R. (2001). A general estimation method using spacings. Journal of Statistical Planning and Inference, 93(1-2), 71-82.

[27] Anatolyev, S., & Kosenok, G. (2005). An alternative to maximum likelihood based on spacings. Econometric Theory, 21(2), 472-476.

[28] Singh, U., Singh, S. K., & Singh, R. K. (2014). A comparative study of traditional estimation methods and maximum product spacings method in generalized inverted exponential distribution. Journal of Statistics Applications & Probability, 3(2), 153.

[29] Metropolis, N., & Ulam, S. (1949). The monte carlo method. Journal of the American statistical association, 44(247), 335-341.

[30] Roberts, G. O., & Smith, A. F. (1994). Simple conditions for the convergence of the Gibbs sampler and Metropolis-Hastings algorithms. Stochastic processes and their applications, 49(2), 207-216.

[31] Chib, S., & Greenberg, E. (1995). Understanding the metropolis-hastings algorithm. The american statistician, 49(4), 327-335.

[32] J. O. Berger, Statistical Decision Theory and Bayesian Analysis, Springer Series in Statistics, Springer, New York, NY, USA, 2nd edition, (1985).

[33] Edwards, W., Lindman, H., & Savage, L. J. (1963). Bayesian statistical inference for psychological research. Psychological review, 70(3), 193.

[34] Chen, M. H., & Shao, Q. M. (1999). Monte Carlo estimation of Bayesian credible and HPD intervals. Journal of computational and Graphical Statistics, 8(1), 69-92.

[35] Box, G. and Tiao, G. (1973). Bayesian Inference in Statistical Analysis. Addison-Wesley, Reading, Massachusetts.

i Надоели баннеры? Вы всегда можете отключить рекламу.