Научная статья на тему 'On MIXTURE OF BURR XII AND NAKAGAMI DISTRIBUTIONS: PROPERTIES AND APPLICATIONS'

On MIXTURE OF BURR XII AND NAKAGAMI DISTRIBUTIONS: PROPERTIES AND APPLICATIONS Текст научной статьи по специальности «Математика»

CC BY
21
11
i Надоели баннеры? Вы всегда можете отключить рекламу.
Область наук
Ключевые слова
Nakagami distribution / Burr XII / mixture distribution / Renyi and beta entropy / MTBF

Аннотация научной статьи по математике, автор научной работы — Hemani Sharma, Parmil Kumar

The Burr XII and Nakagami distributions hold significant importance in both lifetime distribution and wealth distribution analyses. The Burr XII distribution serves as a valuable tool for understanding the distribution of wealth and wages within specific societies, while the Nakagami distribution finds its application in the realm of communication engineering. The incorporation of finite mixture distributions, aimed at accounting for unobserved variations, has gained substantial traction, particularly in the estimation of dynamic discrete choice models. This research delves into the fundamental characteristics of the mixture Burr XII and Nakagami distributions. The study introduces parameter estimation techniques and explores various aspects, including the cumulative distribution function, hazard rate, failure rate, inverse hazard function, odd function, cumulative hazard function, rth moment, moment generating function, characteristic function, moments, mean and variance, Renyi and Beta entropies, mean deviation from mean, and mean time between failures (MTBF). The paper also addresses the estimation of the mixing parameter through a Bayesian approach. To illustrate the effectiveness of the proposed model, two real-life datasets are examined.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «On MIXTURE OF BURR XII AND NAKAGAMI DISTRIBUTIONS: PROPERTIES AND APPLICATIONS»

On MIXTURE OF BURR XII AND NAKAGAMI DISTRIBUTIONS: PROPERTIES AND APPLICATIONS

Hemani Sharma and Parmil Kumar

Department of Statistics, University of Jammu, J&K. hemanisharma124@gmail.com, parmil@yahoo.com

Abstract

The Burr XII and Nakagami distributions hold significant importance in both lifetime distribution and wealth distribution analyses. The Burr XII distribution serves as a valuable tool for understanding the distribution of wealth and wages within specific societies, while the Nakagami distribution finds its application in the realm of communication engineering. The incorporation of finite mixture distributions, aimed at accounting for unobserved variations, has gained substantial traction, particularly in the estimation of dynamic discrete choice models. This research delves into the fundamental characteristics of the mixture Burr XII and Nakagami distributions. The study introduces parameter estimation techniques and explores various aspects, including the cumulative distribution function, hazard rate, failure rate, inverse hazard function, odd function, cumulative hazard function, rth moment, moment generating function, characteristic function, moments, mean and variance, Renyi and Beta entropies, mean deviation from mean, and mean time between failures (MTBF). The paper also addresses the estimation of the mixing parameter through a Bayesian approach. To illustrate the effectiveness of the proposed model, two real-life datasets are examined.

Keywords: Nakagami distribution, Burr XII, mixture distribution, Renyi and beta entropy, MTBF

1. Introduction

Mixture distributions play a pivotal role in modeling complex phenomena where a single underlying distribution cannot fully capture the observed data. These distributions offer a flexible framework by combining multiple component distributions, each representing a distinct pattern or source of variation. In various fields, ranging from statistics and economics to engineering and biology, mixture distributions have gained prominence due to their ability to accommodate diverse data characteristics. By utilizing mixture distributions, researchers can uncover latent subpopulations, account for unobserved factors, and enhance the accuracy of modeling real-world phenomena. The exploration of mixture distributions was instigated by Pearson [14], who initially investigated the blending of two normal distributions. Following Pearson's pioneering work [14], there emerged a significant hiatus in the advancement of mixture distributions. However, this period of inactivity was later reinvigorated by subsequent scholars. Decay [7] built upon Pearson's foundation, while Hasselblad [8] delved extensively into the intricate realm of finite mixture distributions. Numerous practical scenarios involve interpreting observed data as a blend originating from two or more distinct distributions. By leveraging this notion, we have the capability to merge different statistical distributions, thereby forging a novel distribution that inherits the traits of its individual components. This approach holds true particularly when dealing with continuous underlying random variables. An essential characteristic of a mixture distribution is that its cumulative distribution function (CDF) is formed by skillfully combining the CDFs of its constituent distributions through a convex amalgamation. Similarly, the probability density function (PDF) of the mixture can be elegantly expressed as a convex blend, seamlessly

incorporating the individual PDFs of its constituent parts. This versatile mathematical framework not only allows us to effectively handle intricate data but also enables the extraction of meaningful insights by synergistically harnessing the statistical attributes of diverse distributions. Assume T is a continuous random variable with the following probability density function:

k

f (ti) = E Pjfj(f); t > k >0 (!)

i=1

where < pj < 1, j=1,2,3...,k andEk=1 pi = 1. The corresponding c.d.f. is given by

k

F(ti) = E PjFj(t); t > °, k > 0 (2)

i=1

where k is the number of components, the parameters pi, p2,..., pnare called mixing parameters, where pi, represent the probability that a given observation comes from population "i" with density fi(.), and f1 (.),f2(.),...,f3(.), are the component densities of the mixture. When the number of components k=2, a two component mixture and can be written as:

f (t)= p1f1 (t)+ p2f2 (t) (3)

Different research studies have explored various mixture distribution models and their applications. For example, Zaman and colleagues [19] introduced a mixture of chi-square distributions using Poisson elements. Chen and co-authors [4] discussed likelihood inference in finite mixture models, explaining different scenarios for these models. Martin and collaborators [11] utilized a mixture of chi-square distributions to describe the distribution of light in imaging, and they also introduced this distribution to model unstructured light distribution. Daniyel et al. [6] proposed a mixture of Burr II and Weibull distributions and examined their properties. Jaspers and co-workers [10] used a Bayesian approach to estimate mixing weights for multivariate mixture models. Nasiri and Azarian [13] explored estimating mixing proportions for a mixture of two chi-square distributions using various methods and compared their performance. Singh and colleagues [16] introduced a generalized distribution for modeling lifetime data and discussed its statistical properties with real-world applications. Daghestani and team [5] proposed a mixture of Lindley and Weibull distributions, investigated their statistical properties, and estimated their unknown parameters. In this context, this study focuses on the utilization of mixture distributions, specifically exploring the properties and applications of the mixture Burr XII and Nakagami distributions. By delving into their mathematical foundations, parameter estimation methods, and various statistical measures, we aim to provide a comprehensive understanding of how mixture distributions can enhance our ability to capture the intricate nuances present in real-world datasets.

The three-parameter Burr XII distribution was introduced by Titterington et al. [17]. For values of x greater than zero, its cumulative distribution function and probability density function are as follows:

F(x;q,d,b) = 1 - [1 + (X)b]-d;d,b,q > 0 (4)

q

And

f (x; q, d, b) = bdq-bxb-1 [1 + (X)b]-d-1; d, b, q > 0 (5)

q

respectively, where d and b are the shape parameters and q is the scale parameter. It is frequently employed to model income data.

Nakagami [12] introduced the Nakagami distribution, also referred to as the Nakagami-m distribution, to describe radio signal fading. It involves a shape parameter denoted as'm.' If a random variable 'X' conforms to the Nakagami distribution with a shape parameter a > 0.5 and

Hemani Sharma, Parmil Kumar RT&A, No 4 (76) On MIXTURE OF BURR XII AND NAKAGAMI_Volume 18, December 2023

a scale parameter A > 0, its probability density function (p.d.f) can be expressed in the following manner: a

f (x; a, A) = rOy (x2a-1 exp ^Ax2); x > 0, a > 0.5, A > 0 (6)

The corresponding commulative distribution function is given by

F(x) = rar( Jx2,a);x > 0,a > 0.5, A > 0 (7)

where r(x, a) = J0x ta-1e-tdt is the lower incomplete gamma function.

This particular distribution has found applications spanning a diverse range of fields, encompassing hydrology, the transmission of multimedia data across networks, investigations into medical imaging, and the modeling of high-frequency seismogram envelopes.

The primary objective of this research paper is to introduce and explore a novel distribution, which emerges as a fusion of the Burr XII and Nakagami distributions, termed as the Mixture of Burr II and Nakagami Distribution (MBND). The study delves into the statistical properties inherent to this mixture distribution, in addition to the techniques for parameter estimation through Maximum Likelihood Estimation. Furthermore, the paper extends its investigation to include the estimation of the mixing parameter using a Bayesian approach. To validate the efficacy of the proposed model, real-world data is employed, allowing for a comparative assessment against other optimally fitting models.

2. Probability Density Function of MBND

The PDF of the mixture of Burr XII Nakagami distribution (MBND) has the following form

f (x; q, d, b, a, A) = pif1 (x; q, d, b) + p2f2(x; a, A) (8)

Here, with p1 and p2 representing the mixing proportions where their sum equals one (p1 + p2 = 1), f1 (x; q, d, b) signifies the PDF of the Burr XII distribution, and f2 (x; a, A) denotes the PDF of the Nakagami distribution. Combining these individual probability densities results in the mixture density as follows:

f (x;q,d,b,a, A) = p1bdq-bxh-1 [1 + (x)b]-d-1 + p2 JOy ( ^ Va-1exp (-Ax2) (9)

VJ ' T(a)\A where x > 0 and d, b, q >, A > 0, a > 0.5, p1 + p2 = 1

3. Cumulative Probability distribution of MBND

The cumulative distribution function is given by

F(x; q, d, b, a, A) = p1 (x; q, d, b) + p2F2(x; a, A) (10)

Here, with p1 and p2 representing the mixing proportions where their sum equals one (p1 + p2 = 1), f1(x; q, d, b) signifies CDF of the Burr XII distribution, and f2 (x; a, A) denotes the CDF of the Nakagami distribution. Combining these individual probability densities results in the mixture density as follows:

F(x; q, d, b, a, A) = p^1 - [1 + (x)b]"d) + p^-Lr( Ax2, «)) (11)

where x > 0, d, b, q >, A > 0, a > 0.5, pi + p2 = 1.

Figure 1: Plot of the PDF for the MBND

Figure 2: Plot of the CDF for the MBND

4. Properties of Bunami distribution

4.1. Area under the curve

Both the Burr XII and Nakagami distributions constitute complete probability density functions individually, their amalgamation as a mixture will similarly result in a comprehensive probability

density function.

4.2. Reliability function

The reliability function, also called the survival function, describes a feature of a random variable related to how a system might stop working within a certain period.It is defined as the probability that the system will keep working beyond a particular point in time. Mathematically,

R(x) = 1 - F(x)

(12)

By inserting equation (11) into equation (12), the reliability or survival function for the MBND can be expressed in the following manner

R(x) = 1 - (^ - t1 + (x)b]-d) + p2(r!)r(Ax2,a)))

(13)

where x > 0, d, b, q >, A > 0, a > 0.5, pi + p2 = 1.

Figure 3: Plot of the Reliability function for the MBND

4.3. Hazard Function

The hazard function is mathematically defined as the quotient of the probability density function divided by the reliability function, and its expression takes the following form:

h(x)

f (x) R(x)

(14)

Hemani Sharma, Parmil Kumar RT&A, No 4 (76) On MIXTURE OF BURR XII AND NAKAGAMI_Volume 18, December 2023

By substituting equations (9) and (13) into equation (14), the hazard rate for the MBND can be derived, leading to the subsequent definition.

p1bdq-bxb-1 [1 + (I)b]+ p2^ (A) V-1 exp (-ax2) h(x) =-7-:-^-^-J- (15)

1 - (p^1 - [1 + (f)b]-d) + p^rfer(Ax2,a)))

where x > 0, d, b, q >, A > 0, a > 0.5, p1 + p2 = 1.

Figure 4: Plot of the Hazard function of the MBND

4.4. Cumulative Hazard Function

The cumulative hazard function can be defined as

A(x) = - logR(x) (16)

By substituting equation (13) into equation (16), the cumulative hazard function for the MBND can be derived, yielding the subsequent expression.

A(x) = - log where x > 0, d, b, q >, A > 0, a > 0.5, p1 + p2 = 1.

1 - ( K1 - [1+( x )b ]-d ) + P2( ra r( A x2, «)))

(17)

Hemani Sharma, Parmil Kumar RT&A, No 4 (76)

On MIXTURE OF BURR XII AND NAKAGAMI Volume 18, December 2023

6

ill -. 0.3, p2 = 0.7,t :- !.fi = 1.1 = ! a = ft. 4

- - ft - 0 6,^5 - 0.4, ï = 0i.fi = 1. 1- 2.d = 3, = n

- - pi = 05, p2 ■ O.ii.Ii = 05.4 = 2, 1 = 3.« :: i. \ — 2

-pi - 0 4,p2 = 0.6, It = O.fl.if - 3. 3 = 1 rt _ A. \ = 3

4.5

0 1 2 3 4 5 6

X

Figure 5: Plot of Cumulative Hazard function for the MBND

4.5. Reversed Hazard Rate

The reversed hazard rate can be conceptually understood as the result of dividing the probability density function by the cumulative distribution function, as demonstrated by the following formulation:

r(x)=Fi (18)

The reversed hazard rate corresponding to the MBND can be derived by substituting equations (9) and (11) into equation (18), resulting in the following expression.

pxbdq-bxb-1 [1 + (x)b]-d-1 + p2f) x2a-1exp (-ax2)

r(x) =-7-^-^-r- (19)

p\1 - [1 + (x)b]"d) + p^r1)r(fx2,«))

where x > 0, d, b, q >, f > 0, a > 0.5, pi + p2 = 1.

Figure 6: Plot of the Reversed Hazard function for the MBND

4.6. Odds Function

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

The odds function, represented as O(x), is characterized as the proportion of the cumulative distribution function to the reliability function, and it is expressed through the following equation:

O(x) = W) (20)

The odds function corresponding to the MBND can be derived by substituting equations (11) and (13) into equation (20), resulting in the following expression.

pi (l - [1 + (x)b]-d) + P2(rfer(fx2,a))

O(x) =--J--A-(21)

1 - (p^1 - [1 + (x)b]"d) + p^rfer(fx2,«))

where x > 0, d, b, q >, f > 0, a > 0.5, p1 + p2 = 1.

Figure 7: Plot of the Odd Function for the MBND

4.7. rth Moment about Origin

The r moment of the real-valued function can be acquired by applying

Vr = J xrf (x)dx (22)

Vr = /o" x^p1bdq-bxb-1 [1 + (x)c]-k-1 + p2f2y (^ "x2a-1 exp ( Ax2) ) dx (23) ]lr = p1 Jo" xrbdq-cxc-1 [1 + (x)b] -d-1 dx + p2^" xr ^ (j) x2a-1 exp Ax^j dx (24)

j _ piqrr(b + !)r(k - b) + p?r(a + 2)(a\2

4.8. Raw moments about Origin

Putting r=1, 2, 3 and 4 in (25),first four raw moments are

, r( 1 + 1)r(k - 1) r(a + 2) / A \

(25)

ji _ pi q

j2 _ piq

r(d) +p2 Ta \a 2r(2 + 1)r(k - 2) r(a + 1) f A

Mean

r(d)

+ p2"

Ta

j3 _ pi q

3 r(3 + 1)r(k - 3) r(a + § ) / A\ 3

r(d)

p2"

Ta Va

' 4 r(4 + 1)r(k - 4) r(a + 2) f A J4 _ piq4—^--tt7"T\-— + p2^-- I -

r(d)

Ta \ a

(26)

(27)

(28) (29)

a

4.9. Moments about Mean

m = 0 (30)

U2 = u2 - (ui )2 = Variance (31)

№=pi q2r( 2+i;r;k - b)+p2 la+n fa \ - (pi qir( 1+-1)+„2r(a+2) fA \2 n2

r(d) r2 Ta \a) r(d) r2 Ta

' ' ' ' (32)

U3 = U3 - 3U2Hi + (Ui)3

u = „ q3r;3 + l)r(k - 3) „ r(a + 3 )/A\ 3 3

U3 =piq—m-P2—ra~ {a) - 3

pq2r;2 + i)r;k- 2) + r(a + iWA

pi q r^ + U

r(d) fa

p qir;i + i)r(k - i) + „ r(a + 2 v^ 2

Piq ---r p2

+

' qir(i + i)r(k - i) + p r(a + i) f A\ 3

Piq ---r p2

r(d) fa \a

3

(33)

U4 = u3- 4ui u3 + 6(ui)2(u2) - 3(ui)4

3r(b + i)r(k - 3) T(a + 2) f A\ 2 / , r(i + i)r(k - i) r(a + i) f A\ i\

U4 = Pi q3 " r;d) P^^T^ I ~a) - % Piqi F(d) + „2 ra Va ,

3 i 2

Piq3 r(3 + ^ - 3) P2f A\ 3 + 6 L qi r(i + i)r(k - i) + P2 r(a + i) f A \ ^

r(d) r2 Ta \uj r(d) r2 Ta \a

i \ 4

Pi q2 r( 2 + W - 2) + „2 E^ f A \ - 3 pi qir( i + i)r;k - i) + „2 r(a + i) f A \ Z

r(d) r2 Ta \aj r(d) r2 Ta \u,

(34)

4.i0. Measure of Skewness and Kurtosis

Skewness refers to the measure of asymmetry exhibited by a probability distribution. It indicates the extent to which the distribution leans or deviates from being symmetric around its mean. Kurtosis helps characterize the peakedness or flatness of the distribution's curve. Skewness and kurtosis are symbolized as ( and 82 respectively, and their mathematical representation is given as:

(H3 )2 (H2 )3

Using eq.(32) and (33) in the above equation, we can get the value of skewness.

Pi = s (35)

And,

H4

(H2 )

Using eq.(32) and (34) in the equation (36), the expression for the p2 can be obtained.

(82 = (36)

4.ii. Moment Generating Function

The moment generating function is a mathematical function that can be defined as follows:

Mx(t) = E(etx), provided it exists.

where E(etx) = f etxf (x)dx and is given by

^(etX) = I ef^P1 bdq-bxb-1 [1 + (x)b]-d-1 + p2(A) 'x2a-1 exp ( Ax2)) E(etx) = p1 Jo~ etxbdq-bxb-1 [1 + (x)b] -d-1dx + p2 ^" etx^ (j) "x2"-1 exp ( Ax2) dx

E(etx) = p1 £

(qt)j r(b + l)r(fc - b)

j=o

r(d)

+ P2

1 - HI A a

(37)

(38)

4.12. Characteristic Function

The characteristic function is derived from the moment generating function by substituting't' with 'it', and its expression is given as follows:

E(eitx) = P1 £

j=o

fx (t) = E(e'tx)

(qit) r(b + 1)r(k - b) r(d)

+ P2

1 - it\lA a

(39)

(40)

4.13. Mean deviation about Mean

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

The mean deviation represents the average of the absolute differences between data points and the mean of a dataset. It offers a way to estimate the level of variability in a population. The mean deviation about mean can be obtained as follows:

!• TO

51(x) = J |x - v|f (x)dx

where V = E(x). It can be calculated by the following relationship

t V t to

51 (x) = (v - x)f (x)dx + 2 (v - x)f (x)dx

oV t V

51 (x) = 2 J (v - x)f (x)dx

$1 (x) = 2

/■ V

VF(v) + J xf(x)dx

(41)

(42)

$1(x) = 2

p1q

1 r( 1 + 1)r(k - 1) r(a + 2) / A \ 2

r(d)

+ p 2

Ta V a

p^1 - [1 + (x)b+ M ^^V2,a)

q

1 „/a

r(a) A

-( P1bd £(-1)j

j=o

d + 1\ V q

^+1

j J j + b +1

+ p 2

1 .[ATO(-i) (Av2)

2 )i+a-2

r(a)V a £ i! i + a + 1

(43)

—a

TO

—a

TO

V

TO

4.i4. Renyi Entropy

The quantification of uncertainty inherent in random variable X can be computed using Renyi entropy [i5]. This can be articulated using the subsequent equation:

Sr (x) = ^ log (I f (x)dx)

where

(44)

fr (x) =

P^x-[1 + (X)r-1 + P2f2)(f) x^exp^x2)

(45)

By using binomial expansion

The integral will take the form

(a + b)n = £ (^Jan-kbk

E

k=0

n Jo

pi bdq-bxb-1[1 + ( x )b ]-d-1 q

2 f2a-1 (-a 2

P2rwix) x exHXx

dx

Due to their independence, the integral can be treated as the multiplication of the two integrals, leading to the resulting expression for the Renyi entropy as:

Sr (x)

1- r

log

ri (r-n)(b-1)+1 A brd-bnd+r-n-1 bdq-b )r-nq(b-1)(r-n)+1 M b IM b

b

ri (r-n)(b-1)+1 I brd-bnd+r-n-1 11 b + b

+p2

2

n1

(r(a))n V V \n)

2 /1\

(an-n+1)

n 1 )

aan - 2 + 2)

(46)

4.i5. (-Entropy

The (-entropy is formulated as an extension of the Shannon entropy with a single parameter. This concept was initially introduced by Havrda and Charvat [9], and subsequently adapted for use in physical contexts by Tsallis [i8]. The (-entropy can be defined as:

ß ß - 1 ß-entropy for the MBND is given by

/*œ —

1 - /„ fß (x)

dx, for ß = 1.

(47)

H = =— r

ß ß - 1 Jo

x)i]-d-1 fè)(a) x^M-a-2

P1bdq-bxb-1 [1 + (-)b] d 1 + p2) x2 1 exp [ x

By using binomial expansion The integral will take the form

q

(a + b)n = £ (n)an-kbk

k=o V kJ

ß

dx, forß = 1. (48)

ß

E ßn

n=0

p1bdq-bxb-1[1 + (- )b]

x )b ] - d- 1

q

ß-n

p2

r(a) V X

a \ 2«- 1

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

- I x2a 1 exp

x x2,

dx

r

rn

n

œ

r

1

1

a

n

2

œ

0

Due to their independence, the integral can be treated as the multiplication of the two integrals, leading to the resulting expression for the ^-entropy as:

1 - t>

(p1 bdq-b )^-nq(b-1)(^-n)+1

•j-n)(b-1)+1\ rf bfid-bnd+fi-n-1

b JM b

(fi-n)(b-1)+1 + bfid-bnd+fi-n-1

b

b

in—1

+ p 2

(r(a))

a A

(an-n+1)

-s/n^"-2+v n 1,

-) T(an -- + -)

n

(49)

4.16. Mean Time between failures

The Mean Time Between Failures, often abbreviated as MTBF, serves as an indicator of a product's or component's reliability. This metric represents the average interval between successive failures in a process. Its calculation involves determining:

MTBF = j tf (t)dt

(50)

It can be calculated as

MTBF = f <P1 "T^1 + (q>']-d-1 + P2fw(A) ,2"-1ex<Adt (51)

= P1f .Mq-V-1 i1 + )b]-d-1 dx + x^ (A)'t2a-1exp ( A^ (52)

= + p2r(^(A)2 (53)

4.17. Bonferoni and Lorenz curve

The Bonferroni Curve, introduced by Bonferroni [3], serves as a method for analyzing income inequality. It finds utility across various domains, including the examination of economic factors like income and poverty, as well as applications in reliability, demographics, insurance, and medicine.

On the other hand, the Lorenz curve provides a visual representation of the distribution of income or wealth. Though commonly associated with depicting economic inequality, it holds the ability to illustrate uneven distributions in diverse systems. The curvature of the Lorenz curve relative to a baseline, often represented as a straight line, signifies the extent of inequality - the greater the inequality, the farther the curve deviates from this baseline. The following expression is used to calculate the Bonferroni curve:

Bf [F(X)]

1

VF(x) Jo

uf(u)du

(54)

where

j* uf (u)du = Jo\(p1 bdq-bub-1 [1 + (U)b] -d-1 + p2ray (A) u2a-1 exp ( Au2) ) du (55) = jfx p1 ubdq-bub-1 [1 + (u)b] -d-1 dx + Jox p2u^ (j) Xu2x-1 exp (Au^ du (56)

r

1

b

r

1

TO

uf (u)du = p,dq E(-1)j

j=0

d+i

i±j

+1

j ) j + j + 1

+ p2

1 ./A ^ (-i) (A x2)

r(a)

2\i+a-2

A E y V AA a ¿0 i

+ a + 2

(57)

Pi dq Ej=0 (-1)j (d+*) ji+r- + P2 rfeV A E=o

j+i

+i

Bf [F(x)]

(-i) (Ax2)

i+a+

[p, q* +P2 ^ (a) 2) („, (* - [1+(x ? ]-d) + p{ rfe r( Au

Now, Lorenz curve is given as

L(z)

So xf (x)dx

where

J\f (x)du =[x

The final expression for Lorenz curve is

Pibdq-bxb-i[1 + (q)fc]-d-i+P2f2)(f) x2a-iex^-ax2

++1

L(z)

Pidq Ej=o (-i)j C+ + P2 frV a Ei=0 i!T

Piqi r(b+*№-i) + P2 £0^+1) f A \ :

Piq r(d) + P2 r« I a J

(-i)(Az2\

i+a+

a (58)

(59)

dx (60)

(6i)

5. Maximum Likelihood Estimation

Suppose x*, x2, ...xn be the random sample derived from the MBND having the probability density function defined in (9). Therefore, for n observations , the logarithm of the likelihood function is expressed as below

L(x) = n (Pibdq-bxb-i[i + (x)b]-d-i + P2^A) x- "exp

2 /M" 2«-, I -a 2

- I x2a 1 exp ( a x2

(62)

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

The log likelihood function is given as

logL(x) = E¡og(pibdq-bxb-*[1 + (x)b]- d -* + (1- piAj x2» - 1 exp

2 fa

—a

b

0

b

u

;X

n

2

x

[ ( xi ) b]

nlog(pi) + nlog(b) + nlog(d) + -nblog(q) + (b - 1)log E xi + (-d - 1)log E [* + (] + nlog(1 - p,)

i=i i=i q

r^) + aniog[ A) +- 1)logE xi + logE (

i (63)

The equation are in implicit form, so it can be solved using numerical iteration method, such as the Newton-Raphson method via R to obtain the estimates of b, d, q, a, A and p.

6. Bayesian estimation for mixing parameter

To estimate the mixing parameter p1 within the context of MBND, we can adopt the Beta distribution as a suitable prior due to the range constraint p1 € (Q, 1). Let's assume that the prior distribution for p1 is represented by fi(Q, 1).

n(p) _ 9pi

9-1

The likelihood function is given by

n / \ a /

^6-ih , ix^-,-d-i , ^ 2 ( a\ (

A

L(x) _ n (pibdq-bxb-i [i + (-ö-1 + p2^ (A j x2a-i exp

q

And, the posterior distribution of pi is obtained by

L(x)n(p)

n(p|x) _ —

fo L(x)n(p)dp

The marginal distribution is given as

m(x) _ J L(x)n(p)dp

(64)

(65)

(66)

(67)

m( x)

pi-i

n n / / \ —k—i

( pibdq-b )n £ xC-i n (i + ( j)) +(i - pi)n

r(a)

n an n

i_i

i_i

n exp ( -1

m(x) _ fin(x; b,d, q)

9 + n

+ 0f2n (x; a, A)ß(9, n + 1)

Eq.(66) becomes

n(p|x) _

9p?-i(pnf (x; b, d, q) + (1 - pi)nf2n(x; a, A))

dp (68)

(69)

fn (x; b, d, q) 0+n + 9f2n(x; a, A)ß(9, n + 1)

9 pn+9-i f (x; b, d, q)

(1 - pi)n9pi-1 fn(x; a, A)

+

f"(x; b, d, q) ^ + 9fn (x; a, A)ß(9, n + 1) f (x; b, d, q) ^ + 9f2n (x; a, A)ß(9, n + 1)

The Bayesian estimate of p1 under Squared Error Loss Function (SELF) is the mean of the posterior distribution denoted as pB1.

7. Simulation Study

In this part, a Monte Carlo simulation is performed using R for estimating the parameters of mixture of Burr XII and Nakagami distribution (MBND) . The values of n and the parameters are fixed as n = (20,40,70,100) , (b, d, q, a, A) = (1.5,1,1.5,1.5,1) respectively, the mixing parameter p1 is fixed at 0.8 which is estimated using the classical and bayesian approach. The performance of the estimates was evaluated using the Mean Squared Error (MSE). The procedure is replicated 1000 times and obtained results are shown in table 1.

1n

MSE = — £ (a - a)2

i_i

n

2

x

2

0

n

x

9

Table 1: Average values of estimates and their MSE in parentheses

n b d q & X P1 pB1

20 1.9703 1.2695 0.8140 2.4148 0.6744 0.3404 0.4231

(0.8963) (0.5409) (0.7580) (0.7568) (0.7631) (0.4065) (0.3932)

40 1.7349 0.8504 0.9948 2.3071 0.7112 0.5763 0.5892

(0.7130) (0.5140) (0.4177) (0.7200) (0.6878) (0.3896) (0.3624)

70 1.6313 0.9631 1.4108 2.1647 0.8132 0.6296 0.6742

(0.5990) (0.3167) (0.3139) (0.4988) (0.6157) (0.2329) (0.2004)

100 1.3119 0.9726 1.4372 1.9322 0.9542 0.7608 0.7814

(0.2175) (0.1400) (0.1880) (0.2604) (0.2332) (0.1498) (0.1413)

From the table 1, it is noted that

• As the sample size increases, the MSE of the estimates comes down.

• With increase of n, the estimates comes closer to their true values, shows the consistency of the estimates.

• The mixing parameter performs better under the bayesian approach.

8. Real Data Analysis

In this section, we will use two real-world datasets to demonstrate the significance and versatility of the Mixture of Burr XII and Nakagami Distribution (MBND). We will compare the suitability of our proposed model (MBND) with other competing models, namely the Burr XII Lomax Distribution (BLD), Burr II Distribution (BD), and the Length Based Nakagami Distribution (LBND).

To conduct this comparison, we will employ various goodness-of-fit metrics, including the —2l statistic, the Akaike Information Criterion (AIC), the Corrected Akaike Information Criterion (AICc), and the Bayesian Information Criterion (BIC). These criteria aid in selecting the most suitable distribution by identifying the one with the lowest values of AIC, AICc, and BIC. Additionally, we will calculate the Kolmogorov-Smirnov (KS) distance and its associated p-value, providing a further assessment of the goodness of fit.

and

AIC = 2k — 2l BIC = kln(n) — 2f

2k(k +1)

AICc = AIC +

nk1

(70)

(71)

(72)

where l is the maximized log-likelihood, k is the number of parameters, and n is the sample size. Through this comprehensive analysis, we aim to establish the effectiveness of the MBND model in capturing the characteristics of the given datasets and to ascertain its superiority over the alternative distribution models.

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Data Set I: The dataset pertains to the COVID-19 vaccination rate across 46 distinct countries situated in southern Africa. This dataset has been previously subjected to analysis by [2]. The data is as follows: 0.042, 0.205, 0.285, 0.319, 0.464, 0.550, 0.889, 0.895, 0.939, 0.986, 1.000, 1.088, 1.212, 1.244, 1.450, 1.593, 1.844, 2.039, 2.157, 2.167, 2.334, 2.440, 2.657, 3.685, 3.879, 4.493, 4.800, 4.944, 5.155, 5.674, 7.602,10.004,12.238,12.520,12.553,13.063,15.105,15.229,15.629,15.848,18.641, 18.940, 29.885, 58.162, 61.838, 72.286

Figure 8: (a) Ecdf plot for the dataset I (b) q-q plot for the dataset I

data

Figure 9: Plot of the fitted densities for dataset I

Table 2: Maximum Likelihood Estimates of the parameters of the different models.

Model b q d a A

MBND 1.7543 0.8746 3.4638 1.0266 6.3816

BLD 1.8732 4.1136 19.3219 0.03191 13.995

BD 1.0461 32.1767 28.9722 - -

LBND - - - 0.4712 1.6594

Hemani Sharma, Parmil Kumar RT&A, No 4 (76)

On MIXTURE OF BURR XII AND NAKAGAMI_Volume 18, December 2023

Table 3: Goodness of fit measures for the different models.

Model -2f AIC AICC BIC K-S p-value

MBND 110.3650 120.3650 121.9034 129.3983 0.0773 0.9313

MBLD 114.4867 124.4867 126.0251 133.5200 0.0835 0.8864

BD 178.6642 184.6642 185.2496 190.0842 0.5233 5.279e-12

LBND 155.8333 159.8333 160.1190 163.4466 0.1757 0.1096

Data Set II: Almongy et al. [1] analyzed this dataset to showcase the utility of an innovative extended Rayleigh distribution. This distribution was employed to model COVID-19 mortality rate data from Mexico spanning a period of 108 days, starting from March 4th and concluding on July 20th, 2020. The dataset pertains to a raw mortality rate and is characterized by its relatively unrefined nature. The data are as follows: 8.826, 6.105, 10.383, 7.267, 13.220, 6.015,10.855, 6.122, 10.685, 10.035, 5.242, 7.630, 14.604, 7.903, 6.327, 9.391, 14.962, 4.730, 3.215, 16.498, 11.665, 9.284, 12.878, 6.656, 3.440, 5.854, 8.813,10.043, 7.260, 5.985, 4.424, 4.344, 5.143, 9.935, 7.840, 9.550, 6.968, 6.370, 3.537, 3.286, 10.158, 8.108, 6.697, 7.151, 6.560, 2.988, 3.336, 6.814, 8.325, 7.854, 8.551, 3.228, 3.499, 3.751, 7.486, 6.625, 6.140, 4.909, 4.661,1.867, 2.838, 5.392, 12.042, 8.696, 6.412, 3.395,1.815, 3.327, 5.406, 6.182, 4.949, 4.089, 3.359, 2.070, 3.298, 5.317, 5.442, 4.557, 4.292, 2.500, 6.535, 4.648, 4.697, 5.459, 4.120, 3.922, 3.219, 1.402, 2.438, 3.257, 3.632, 3.233, 3.027, 2.352, 1.205, 2.077, 3.778, 3.218, 2.926, 2.601, 2.065, 1.041, 1.800, 3.029, 2.058, 2.326, 2.506, 1.923.

Empirical CDF

8 10 x

6 8 10 12 Theoretical Quantlles

4 r

3 -

2

0

2

4

6

12

14

16

18

0

0

2

4

14

18

Figure 10: (a) Ecdf plot for the dataset II (b) q-q plot for the dataset II

CD Q

C5

O C5

O

O O

MBND BD

MBLD LBND

10

15

data

Figure 11: Plot of the fitted densities for dataset II Table 4: Maximum Likelihood Estimates of the parameters of the different models.

Model b q d Ci X

MBND 5.2318 7.4680 0.9442 2.0159 12.4548

MBLD 3.7382 4.8573 0.9421 0.0090 18.9176

BD 2.3266 8.7808 2.6983 - -

LBND - - - 0.4719 21.1924

Table 5: Goodness of fit measures for the different models.

Model -2f AIC AICC BIC K-S p-value

MBND 530.8075 540.8075 541.3957 554.2181 0.0486 0.9602

MBLD 558.2015 568.2015 568.7097 581.6121 0.1151 0.1144

BD 685.1428 691.1428 691.3735 699.1892 0.5054 2.20e-16

LBND 538.4605 542.4605 542.5748 547.8247 0.3049 3.676e-09

Table 2 and Table 3 provide the Maximum Likelihood Estimates (MLEs) and offer a comparison of the performance between the Mixture of Burr XII and Nakagami Distribution (MBND) and the alternative distributions for the first dataset. Similarly, Table 4 and Table 5 present the MLEs and compare the performance of MBND with other distributions for the second dataset. The findings displayed in Table 3 and 5 demonstrate that MBND achieves the lowest values for AIC, AICc, and BIC, indicating its superiority over several well-established competing models. The claim is

0

further supported by Figures 8 and 10. Additionally, the results presented in Table 3 and 5 are corroborated by examining the Empirical Cumulative Distribution Function (ECDF) plots and Quantile-Quantile (Q-Q) plots for the MBND model in the context of both datasets, showcased in Figure 7 and 9. These visual representations lend support to the outcomes highlighted in Table 3 and 5.

9. Conclusion

In this paper, we presented a mixture of the Burr XII and Nakagami distribution (MBND). We investigate its statistical properties and demonstrate that this new mixture distribution can serve as an alternative model to some existing distributions. For MLEs, simulation schemes are developed that yield less Mean Squared Error (MSE) as sample size increases. Additionally, Bayesian estimates of the mixing parameter are computed, demonstrating their superiority over classical estimates. We utilize real-world datasets to illustrate the efficacy of the MBND model and compare it with other competing distributions. The results of this comparison reveal that the MBND model offers a better fit than the considered distributions. Through this comprehensive analysis, we establish the viability of MBND as a valuable tool for modeling and analyzing various data scenarios.

References

[1] Almetwally E. M., Aljohani H.M., Alghamdi A.S, Almongy, H. M. and Hafez, E.H. (2021). A new extended rayleigh distribution with applications of covid-19 data. Results in Physics, 23, 104012.

[2] Almetwally E. M., HajAhmad, H., Almongy, H. M. and Al-nefaie, A. H. (2022). Modeling of covid 19 vaccination rate using odd lomax inverted nadarajah- haghighi distribution. Plos one, 17(10), e0276181.

[3] Berferroni, C. E. Elmenti di statitica generale. Libreria Seber, Firenze., 1930.

[4] Chen, X., Ponomareva, M. and Tamer, E. Likelihood inference in some finite mixture models. Journal of Econometrics, 182:87-99.

[5] Daghestani, A. F., Sultan, K. S. and Al-Moisheer, A. S. (2021). Mixture of lindley and weibull distributions: Properties and estimation. Journal of Statistics Applications and Probability, 10(2):301—313.

[6] Daniyal, M. and Aleem, M. (2014). On the mixture of burr xii and weibull distributions. journal of statistics applications and probability, 3:251--267.

[7] Decay, M. F. (1964). Modified poisson probability law for a point pattern more than random. (1964). Association of American Geographers Annals, 54:559—565.

[8] Hasselblad, V. (1968). Estimation of mixtures of distribution from exponential family. Journal of American Statistical Association, 64:300—304.

[9] Havrda, J. and Charvat, F. (1967). Quantification method in classification processes: concept of structural a-entropy. Kybernetika, 3:30--35.

[10] Jaspers, S., Komrek, A. and Aerts, M. (2018). Bayesian estimationof multivariate normal mixtures with covariate-dependent mixing weights, with an application in antimicrobial resistance monitoring. Biometrical JournaL, 60:7—19.

[11] Martin, M., Hory, C. and Huchard, M. Chi-square law mixture for an unsupervised time frequency pattern extraction. Proceedings of the Tenth International Congress on Sound and Vibration, Stockholm, Sweden, 7-10 July, Code 62825, 2003.

[12] Nakagami, M. (1960). The m-distribution-a general formulation of intensity distribution of rapid fading. in W. C. Hoffman (ed.) Statistical Methods in Radio Wave Propagations, 3—36.

[13] Nasiri, P. and Azarian, A. A. (2018). Estimating the mixing proportion of mixture of two chi square distributions. International Journal of Applied Mathematics and Information Sciences, 12(4):753--759.

[14] Pearson, k. (1894). Contribution to the mathematical theory of evolution. Philosophical Transactions of the Royal Society of London, 185:71-110,1894.

[15] Renyi, A. On the measure of entropy and information. Proceedings of the fourth Berkely symposium on mathematical statistics and probability, 1:547-561,1961.

[16] Singh, B. P., Singh, S. and Das, U. D. (2021). A general class of new continuous mixture distribution and applications. J. Math. Comput. Sci., 11(1):585—602.

[17] Titterington, D.M., Smith, A.F.M. and Makov, U.E. Statistical Analysis of Finite Mixture Distributions. John Wiley and Sons, New York, 1985.

[18] Tsallis, C. (1968). Possible generalization of boltzmann-gibbs statistics. Journal of Statistical Physics, 52:479--487.

[19] Zaman, M. R., Paul, D. N. R., Akhter, N., Howlader, M. H. and Kabir, M. S. (2006). Chi-square mixture of chi-square distributions. Journal of Applied Sciences, 6:243--246.

i Надоели баннеры? Вы всегда можете отключить рекламу.