Научная статья на тему 'Fast methods for jackknifing inequality indices'

Fast methods for jackknifing inequality indices Текст научной статьи по специальности «Математика»

CC BY
134
25
i Надоели баннеры? Вы всегда можете отключить рекламу.
Журнал
Прикладная эконометрика
Scopus
ВАК
Область наук
Ключевые слова
JACKKNIFE / RESAMPLING / SAMPLING VARIABILITY / INEQUALITY

Аннотация научной статьи по математике, автор научной работы — Karoly L. A., Schröder C.

The jackknife is a resampling method that uses subsets of the original database by leaving out one observation at a time from the sample. The paper develops fast algorithms for jackknifing inequality indices with only a few passes through the data. The number of passes is independent of the number of observations. Hence, the method provides an efficient way to obtain standard errors of the estimators even if sample size is large. We apply our method using micro data on individual incomes for Germany and the US.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Fast methods for jackknifing inequality indices

The jackknife is a resampling method that uses subsets of the original database by leaving out one observation at a time from the sample. The paper develops fast algorithms for jackknifing inequality indices with only a few passes through the data. The number of passes is independent of the number of observations. Hence, the method provides an efficient way to obtain standard errors of the estimators even if sample size is large. We apply our method using micro data on individual incomes for Germany and the US.

Текст научной работы на тему «Fast methods for jackknifing inequality indices»

L. A. Karoly, C. Schroder

Fast methods for jackknifing inequality indices

The jackknife is a resampling method that uses subsets of the original database by leaving out one observation at a time from the sample. The paper develops fast algorithms for jack-knifing inequality indices with only a few passes through the data. The number of passes is independent of the number of observations. Hence, the method provides an efficient way to obtain standard errors of the estimators even if sample size is large. We apply our method using micro data on individual incomes for Germany and the US. Keywords: jackknife; resampling; sampling variability; inequality. JEL classification: C81; C87; D3.

1. Introduction

When examining time-series changes in inequality or cross country differences in inequality, the measured changes are sometimes small. The delta method, bootstrapping and jackknifing are popular methods for performing statistical inference on inequality indices. The delta method is based on the asymptotic distribution of the index, whereas the jackknife and the bootstrap are distribution free resampling methods. The bootstrap samples with replacement from the original sample1. The jackknife uses subsets of the original database by leaving out one observation at a time from the sample2.

The delta method, if applicable to an inequality index, has as advantages that it derives the asymptotic distribution using the central limit theorem, and that the computational burden to estimate the asymptotic variance consistently is low. However, the delta method is not free of limitations. For example, one difficulty arises if cross-sections of a panel survey are used as the empirical database, as it is the case for available micro datasets including the Luxembourg Income Study (LIS)3, probably the most frequently used database for distributional analyses worldwide. Then for testing for inter-temporal changes in inequality indices the inter-temporal covariance structure of incomes must be considered. Further difficulties arise in the following contexts (see (Heshmati, 2004), and references therein for details): correlated data, correlations introduced in error terms4, panel attrition, and inequality decomposition5.

1 For the validity of the bootstrap technique in inequality analyses see (Biewen, 2002).

2 For the theoretical justification for the jackknife and other related resampling techniques see (Efron, 1982).

3 http://www.lisdatacenter.org/.

4 Modarres and Gastwirth (2006) show that regression estimates of the standard error for the Gini index can be biased «as it does not account for the correlation introduced in the error terms once the data are sorted» (p. 387).

5 Decomposition analyses impose restrictions on the functional forms of regression models (see (Fields, Yoo, 2000),

or (Morduch, Sicular, 2002)).

Aforementioned limitations of the delta method explain the interest in resampling methods6. The central advantage of the jackknife over other resampling methods such as the bootstrap is that it allows the replication of results. A disadvantage of standard jackknife procedures is that for large sample sizes the computational burden is substantial. This is because there are as many subsets as there are observations in the sample, and for each subset the jackknife statistic needs to be computed. This paper offers a solution. We provide fast algorithms, requiring only a few passes through the data, for jackknifing several popular inequality indices: coefficient of variation, variance of the logarithms, mean log deviation, Theil index, and Atkinson index7. Since the number of passes is independent of the number of observations, even for large samples the computational burden remains small.

To get an idea of the computational burden see Fig. 1. It charts the computer time in minutes for a standard procedure for jackknifing inequality indices as a function of sample size8. Computer time increases exponentially in sample size, and for a sample size of about 80 000 cases it already exceeds four hours. Since many comparative inequality analyses rely on data from several points in time, countries and income concepts, computing the jackknife for all results can easily take days or weeks. This is a serious limitation, especially for researchers who use data stored on external servers (e. g., the Luxembourg Income Study) and face limited processing power.

Section 2 explains our jackknife algorithm. Section 3 provides the results from an empirical application relying on data from the Luxembourg Income Study. Section 4 concludes. Derivations of all the formulas are provided in an Appendix.

The jackknife offers a conceptually simple way to estimate the precision of a statistic, see the pioneering (Tukey, 1958; Efron, 1982; Efron, Gong, 1983; Wolter, 1985). In the context of inequality measurement, we have a random sample of N observations on income, y = (y1,yN) and sampling weights, w1,w2,...,wN. Let d = d(y) denote our measure of inequality. Let d( f) = d( y1, y2,..., yi-1, yM,..., yN) denote the jackknife estimate of the same measure of inequality for the subset where the z-th observation has been deleted.

Following Wolter (1985), the jackknife estimate of the standard error of 9 is

6 For example, Biewen (2002) shows the validity of the bootstrap in a number of aforementioned contexts. Modarres and Gastwirth (2006) recommend bootstrap and jackknife as reliable methods for random samples as opposed to other methods that assume independence between errors when they are dependent (as it is the case for the Gini coefficient). Bhattacharya (2007) suggests techniques of asymptotic inference of the Gini coefficient based on process theory and the delta method. The resulting variance formula, however, is rather difficult to implement (Davidson, 2009).

7 Algorithms for the Gini coefficient are provided in (Karagiannis, Kovacevic, 2000) and (Yitzhaki, 1991). Karoly (1989) derives jackknife procedures for calculating the between- and within-group inequality components of the variance of the logarithms, the mean log deviation, and the Theil index. Ogwang (2000) shows that it is also possible to obtain standard errors for the Gini index from OLS regression. Giles (2004) extends the regression-based approach to test hypotheses regarding the sensitivity of the Gini coefficient to changes in the data using seemingly unrelated regressions.

8 We have used the STATA software package inequal7.ado on the following hardware: 64-bit system; 8 GB RAM; Core 2 Duo CPU, 3GHz. STATA code is available in the authors' working paper (Karoly, Schröder, 2014).

2. Efficient jackknife procedures for inequality indices

number of obs.

Fig. 1. Computer time and sample size

_ 1 v N

with w = 9 Computing the jackknife standard error estimate relies on the values of

d(i), one jackknife statistic per subset. For large samples the computational burden to derive equation (1) seems to be large. However, as we will outline below, for standard inequality indices deriving the N values of requires just a few passes through the data. Hereby, the number of passes is independent of the number of sample observations, N.

The procedure is detailed below by means of the Theil index, and the variance of logarithms. The general idea of the procedure is to write the jackknife estimates as a function of statistics from the overall sample (i. e., as a function of 6, N, arithmetic or geometric mean) and a subset-specific correction factor that can be derived with a single run through the data. The procedure can be adapted to other inequality indices including indices of the generalized entropy class, and indices based on the variance or social-welfare functions (e. g. the Atkinson index). We will make use of the following notation and definitions.

/ 1 V* N

1. wi denotes the normalized weight, wt =®y ^^ . Accordingly, ^ wt = N.

_ _ 1 V*N

2. y denotes the arithmetic mean of income, y = — > w.y. .

3. y* denotes the geometric mean of income, y* = exp ^ ^ W ln (yi)). The natural logarithm of the geometric mean is denoted x = ln (y*) = N ^ ,_Wxi with xt = ln (yt).

9 An alternative method is to compute the squared differences between the jackknife statistics and their mean (see, for example, (Yitzhaki, 1991)).

2.1. Efficient jackknife procedure for the Theil index

The Theil index from the sample is

QT —

1

. ^ wtyt ln (y) - ln (y).

N y \ i—i /

The Theil index for the subset where the i-th observation has been deleted is

/ \

Or,n =

1

Г(г)_ (N - w ))

Swyjln(yj) -ln())'

with y(i) denoting the arithmetic mean of income from the subset

У(' )

Ny - wiy

N - w;

The first step is to write 9T(i} in terms of 9T. Initially, from (3):

1

T(i) LLL "J- "T

I N \

втT) = (N - мл )) ^

2 мУln (y )

м y

( N - wt )y

ln (y )- ln (У( i)).

Rewriting equation (2) gives

2!=!ln (y )=&+ln (y )]Ny,

and substituting (6) and (4) into (5) gives

- - Ny (+ln(y))-wy^yl-in

'Ny- wtyts

'T (i )

Ny - ЩУ>

Ny - ЩУ,

N - w

i

(2)

(3)

(4)

(5)

(6)

(7)

Equation (7) reveals that QT (i} can be expressed as a function of three statistics from the full sample, N, y, and 6T, and characteristics of the observation that is left out, wi and y .. Thus, after having calculated N,y, and 6T for the full sample, to compute all the jackknife statistics вт (j),..., вт ( N ) takes a single pass through the data.

2.2. Efficient jackknife procedure for the variance of logarithms

Applying Bessel's correction10, the variance of the logarithms from the sample is

^VL =

1

N

Sw, ln i . , '

/ \2 л

У

1

N -

1S w -x)2.

(8)

10

Bessel's correction, the division in the variance formula by N — 1 instead of by N, secures unbiasedness.

N

N

The variance of the logarithms for the subset where the i-th observation has been deleted is

1

(i > N - 2

S wj (i >(

j(i> ( xj X(i>

with x,... =-\NX - xw ], and with w....

' N - w. J '

(N - w.)/(N-1)

) , (9)

denoting re-weighted norma-

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

ls u>

Si

£ -j

lized weights. By means of the re-weighting the average of Wj(i} over the subset where the i-th observation has been deleted equals unity. So, the analogue of the term ——- in (8) in (9) is

——^. Substituting the definition of Wj(i} in (9) gives

в _ (N-1) VL( 0 (N - 2)(N - )

S (xj- xa))2

Initially, from (8):

dVr =-^ w. (x. - x ) +--w. (x, - x ) .

VL N N -1 '

Substituting x = -1 [(VV - wi )} + xiwi J in (11) gives

dvr =-^ w\x.--|"(n- w.)x(1) + x.w. 1| +--w. (x. -x) =

VL n^ 1 NL (1) 11J) N-\ '

(10)

(11)

=-^ w.

N-1 £ j

\2

N _ Wi _ wt

X- X/J-, + X/J-, X-

J N (i) n (i) N '

\

(12)

1 / _\2

+--W' (X - X )

N- 1 ii

/

Equation (12) can be rewritten as 1

dVL =

N -

7S w( - xw)2 +TN—7)2 wj[(xj - xw 2(N)(x»- x')

+

+--S w,. | — xin---x,

-1

N-1 ' N (i) N "'

1 ( -V

+--W' ( - x 2

N- 1 ii

(13)

(N - 2) (N - )

The C-term on the right hand side of (13) can be rewritten as 0VL(i)----2-.

The D-term is zero since (N —1)

2 2w

D = N— In 2wj -)-)=rN-^TN()- - ) =0. (14)

w

The E-term after some algebra becomes

E = 1 , W'2 У w, (x - x, )2 = 1 7 W ч2 (TV - w, )(x - x, )2 = N-1 (N - wt fit Л J N-1 (n - Wj 2 J

1 W2 /_ 42

(15)

—(x - X f

N-1 N - w;

Substituting (14)-(15) in (13), the variance of the logarithms for the sample becomes

(N- 2)(N - w,) 1 w2 / N2 w. , N2 dVL =dVW)--/ A x2 '' +--(x -xt) +—^( -x) . (16)

VL VL(,) (n-1)2 N-1N - Wt N-1 ' '

After some algebra, (16) becomes

(N- 2){N- wt) + Nwj (_ )2 (17)

Solving (17) with respect to 9VL(i} gives the desired expression for the jackknife estimator of

the variance of the logarithms:

л (N-l)2 Nwi (N-1)

'VL( i) "VL

(N - 2)(n - w) (n - wt )2 (N - 2)

(x -Xi)2. (18)

Equation (18) is the analogue of the jackknife estimator of the Theil index in equation (7): dVL(^ can be expressed as a function of statistics from the full sample (N, x, and 9VL) and the characteristics of the observation that is left out, wi and xi. Thus, after having calculated N, x, and 9VL for the full sample, computing 9VL,..., 9VL{N} takes a single pass through the data.

2.3. Efficient jackknife procedure for other inequality indices

Similar derivations as those explained in Sections 2.1 and 2.2 can be made for other inequality indices. Formulas for an efficient computation of the Atkinson index, dA (with inequality aversion parameter e = 1 and £ = 2), the mean log deviation, 9MLD, and the (coefficient of variation, 9CV, are as follow:

exp

вл^а) -1-"

ln (/ )-^^

N - щ v ' N - w

(Ny wy)/(N - щ)

(19)

О^) =1-(N - wt)/

Ny — wiyi N

4

wt (Ny -

■wyt)

■wi)

(20)

' MLD (i )

N Гд i Wiln (yi )

Г0MLD - ln (У )J +- + ln

'Ny- ytwS

N - w

N - w

iv

N - w

i

(N -l)2

Nwi (N -l)

\0.5

'CV (i)

(N- 22(n - Wi ) (n- wi )2 (N - 2)

(N - Wi у (ny - yWi)

(y - yi )2

(21)

(22)

U)

Sí £

-j

Derivations of the formulas can be found in the Appendix. Again, after having calculated some basic statistics from the full sample, computing all the jackknife indices takes only a single pass through the data.

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

в

V

3. Empirical application

We have calculated the above inequality indices and their associated jackknife confidence intervals for distributions of disposable household incomes in the US and in Germany from the Luxembourg Income Study (LIS) database. For 40 countries and several years, the LIS provides representative micro-level information on private households' incomes and their demographics.

Our computations rely on the LIS household-level datasets. Household disposable income is our income concept. Household disposable income covers labor earnings, property income, and government transfers in cash minus income and payroll taxes. The Luxembourg Income Study spends enormous efforts in providing harmonized (standardized) income data, both in terms of conceptual content (variable definitions are comparable across datasets) and in terms of coding structure. Further information on the LIS data is provided online11.

To adjust household incomes for differences in needs, we have deflated household disposable income by means of the square root equivalence scale. The square root equivalence scale is the number of household members to the power of 0.5. This gives the needs-adjusted equivalent income of the household. Household units are weighted by the frequency weights12 times the number of household members. Our weighting procedure accommodates the principle of normative individualism that considers any person as important as any other. The so derived distribution depicts differences in living standards, captured by differences in equivalent incomes, among individuals (Bönke, Schröder, 2012).

We have removed household observations with missing information or with negative values of disposable income. Moreover, to avoid outlier-driven biases of inequality estimates, we use trimmed data with the one percent observations with the highest and with the lowest incomes being discarded.

It has taken a few seconds to obtain all the results presented in Table 1. The Table is split in two panels. The upper panel provides the results for the US, the lower panel provides the results

11 For details see http://www.lisdatacenter.org/our-data/lis-database/documentation/.

12 Household frequency weights are provided from the national data providers. For example, the German frequency weights come from the German data provider, the Socio-economic Panel. These weights make the sample representative of the total (covered) national population (usually, collective or institutionalized households are left out).

for Germany13. In the US, the results cover the period 1991-2010; in Germany, the results cover the period 1994-2010. For every country-period combination, the Table provides the point estimates of the inequality indices along with their upper and lower bounds of 95 percent confidence intervals, CI? and CI? , derived from the jackknife statistics.

Table 1. Inequality indices

Country Year Atkinson (e = 1) Atkinson (e = 2) Mean log deviation

' .1 ah: 1 CI? ' A2 CI? ' A2 CI? 'MLD л 'MLD CI? 'MLD

US 1991 0.162 0.166 0.169 0.329 0.337 0.345 0.177 0.181 0.186

1997 0.177 0.181 0.185 0.348 0.357 0.366 0.195 0.199 0.204

2000 0.173 0.177 0.180 0.340 0.348 0.356 0.190 0.194 0.199

2004 0.179 0.183 0.186 0.361 0.371 0.380 0.197 0.202 0.206

2007 0.185 0.188 0.191 0.363 0.370 0.377 0.204 0.208 0.212

2010 0.193 0.197 0.201 0.402 0.411 0.421 0.215 0.219 0.224

DE 1994 0.088 0.095 0.102 0.175 0.188 0.200 0.093 0.100 0.107

2000 0.088 0.093 0.098 0.174 0.185 0.196 0.092 0.098 0.103

2004 0.098 0.106 0.114 0.184 0.203 0.222 0.103 0.112 0.121

2007 0.102 0.111 0.120 0.193 0.210 0.226 0.107 0.117 0.127

2010 0.103 0.110 0.117 0.198 0.212 0.225 0.109 0.116 0.124

Theil index Variance of logs Coefficient of variation

CI?T 0T a?T CI' 'VL 'VL CI? 'VL CI' 'CV 'CV CI? 'CV

US 1991 0.158 0.161 0.165 0.396 0.408 0.419 0.574 0.581 0.587

1997 0.180 0.184 0.189 0.422 0.435 0.447 0.637 0.646 0.654

2000 0.177 0.181 0.185 0.410 0.421 0.432 0.633 0.643 0.653

2004 0.178 0.182 0.185 0.439 0.452 0.464 0.625 0.633 0.640

2007 0.188 0.192 0.196 0.445 0.456 0.466 0.653 0.661 0.669

2010 0.189 0.192 0.196 0.494 0.508 0.522 0.639 0.646 0.652

DE 1994 0.090 0.097 0.104 0.191 0.207 0.222 0.437 0.456 0.475

2000 0.090 0.095 0.099 0.190 0.203 0.216 0.439 0.451 0.463

2004 0.103 0.111 0.119 0.204 0.226 0.248 0.480 0.500 0.519

2007 0.108 0.118 0.129 0.213 0.234 0.254 0.493 0.522 0.551

2010 0.107 0.114 0.122 0.221 0.238 0.254 0.482 0.504 0.525

Note. Data from Luxembourg Income Study.

We comment on the US first. An examination of the statistics shows a significant rise of inequality over the observation period: the point estimate of the Theil index increases from 0.161 in 1991 to 0.192 in 2010, and the confidence intervals are clearly distinct: [0.158; 0.165] vs.

13 The LIS data for Germany are based on the German Socio-Economic Panel Study (SOEP).

[0.189; 0.196]. However, some inter-temporal changes in inequality for this sample are not sta- .¡в tistically significant (e. g. 1997-2000; 2000-2004; 2004-2007). |

For Germany, we also see a significant rise of inequality over the observation period. This is « due to a prominent rise of inequality between 2000 and 2004. The inter-temporal comparisons ^ before the rise (1994-2000) and after the rise (2004-2007 and 2007-2010) indicate no signifi- | cant changes in inequality. *

Comparing inequality levels in the US and Germany there is significantly more inequality j in the US. The result holds for all six inequality indices and all the observed points in time14.

Finally, as empirical illustration, we compare the jackknife confidence intervals for the Theil index with the normal 95% confidence intervals from the bootstrap and from asymptotic variances15. The results are summarized in Table 2.

Table 2. Alternative confidence intervals for the Theil index

Country Year Jackknife Bootstrap Asymptotic

CI'I CIeT CI'I Cll ci'I CIeT

US 1991 0.158 0.165 0.159 0.163 0.159 0.163

1997 0.180 0.189 0.182 0.187 0.182 0.187

2000 0.177 0.185 0.178 0.184 0.178 0.184

2004 0.178 0.185 0.179 0.184 0.179 0.184

2007 0.188 0.196 0.189 0.195 0.189 0.195

2010 0.189 0.196 0.190 0.195 0.190 0.195

DE 1994 0.090 0.104 0.093 0.102 0.093 0.101

2000 0.090 0.099 0.091 0.098 0.091 0.098

2004 0.103 0.119 0.107 0.116 0.107 0.115

2007 0.108 0.129 0.113 0.124 0.113 0.124

2010 0.107 0.122 0.110 0.119 0.110 0.119

Note. Data from Luxembourg Income Study.

The general message from Table 2 is that the confidence intervals from the three methods are very close. Indeed, confidence intervals from the bootstrap and from asymptotic variances, for many years, coincide. Inter-temporal and cross-country comparisons are insensitive to the chosen methods. The inter-temporal comparisons indicate a significant rise in inequality for the US between 1991 and 1997 and again between 2004 and 2007 (Germany: between 2000 and 2004). The cross-country comparisons indicate that inequality is significantly higher in the US.

14 We have executed our empirical analysis using the alternative formulation of the standard error introduced in footnote 4. It did not change our conclusions since confidence intervals changed very tittle.

15 The bootstrap confidence intervals have been obtained with the STATA ado-package ineqerr with 1000 repetitions. The confidence intervals from asymptotic variances have been obtained with the STATA ado-package geivars, with the underlying formulas coming from (Cowell, 1989). Results for other inequality indices can be provided on request.

4. Conclusion

This paper has outlined a procedure to obtain jackknife estimates for several inequality indices with only a few passes through the data. The number of passes is independent of the number of observations: After having computed some statistics from the overall sample, computing all the jackknife indices takes only a single pass through the data. Hence, the method provides an efficient way to get standard errors of the estimators even if sample size is large.

We have applied our method using data from the Luxembourg Income Study to evaluate the statistical significance of inter-temporal inequality in Germany and the US, and also to evaluate cross country differences in inequality levels.

References

Bhattacharya D. (2007). Inference on inequality from household survey data. Journal of Econometrics, 137, 674-707.

Biewen M. (2002). Bootstrap inference for inequality, mobility and poverty measurement. Journal of Econometrics, 108, 317-342.

Bönke T., Schröder C. (2012). Country inequality rankings and conversion schemes. Economics: The Open-Access, Open-Assessment E-Journal, 6, 2012-2028.

Cowell F. A. (1989). Sampling variance and decomposable inequality measures. Journal of Econometrics, 42, 27-41.

Davidson R. (2009). Reliable inference for the Gini index. Journal of Econometrics, 150, 30-40.

Efron B. (1982). The jackknife, the bootstrap and other resampling plans. Society for Industrial and Applied Mathematics, Philadelphia PA.

Efron B., Gong G. (1983). A leisurely look at the bootstrap, the jackknife, and cross-validation. The American Statistician, 37, 36-48.

Fields G. S., Yoo G. (2000). Falling labour income inequality in Korea's economic growth: Patterns and underlying causes. Review of Income and Wealth, 46, 139-159.

Giles D. (2004). Calculating a standard error for the Gini coefficient: Some further results. Oxford Bulletin of Economics and Statistics, 66, 425-433.

Heshmati A. (2004). A review of decomposition of income inequality. IZA Discussion Paper, 1221.

Karagiannis E., Kovacevic M. (2000). Practitioners corner — a method to calculate the jackknife variance estimator for the Gini coefficient. Oxford Bulletin of Economics and Statistics, 62, 119-122.

Karoly L. A. (1989). Computing standard errors for measures of inequality using the jackknife. Unpublished manuscript.

Karoly L. A., Schröder C. (2014). Fast methods for jackknifing inequality indices. SOEPpapers onMul-tidisciplinary Panel Data Research No. 643, DIW Berlin.

Modarres R., Gastwirth J. L. (2006). A cautionary note on estimating the standard error of the Gini index of inequality. Oxford Bulletin of Economics and Statistics, 68, 385-390.

Morduch J., Sicular T. (2002). Rethinking inequality decomposition, with evidence from rural China. The Economic Journal, 112, 93-106.

Ogwang T. (2000). A convenient method of computing the Gini index and its standard error. Oxford Bulletin of Economics and Statistics, 62, 123-129.

Tukey J. W. (1958). Bias and confidence in not quite large samples. Annals of Mathematical Statistics, 29, 614-623.

Wolter K. (1985). Introduction to variance estimation. Springer, New York.

Yitzhaki S. (1991). Calculating jackknife variance estimators for parameters of the Gini method. Journal of Business and Economic Statistics, 9, 235-239.

Appendix

Derivation of jackknife formulas

Mean log deviation (Entropy 0)

1 N

1 ^ N 0 MLD = N Zi=i wln

y

1 N

y) N

2<=1 wiln )+ln (y).

в

1

MLD (i)

N - w^i*

S^WlnU) +ln(y).

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

From (2mld ): в

MLD (i) N - wt в

1 ^j^Wiln (yj) + wln (y) 1 [Sli Wln (у)

, wln(y ) , 1 (- ) + ^-+ln W)),

N - w

, Wln (У ) , 1 (- )

+ ~-+ln Uo).

N - w

) N - _ Substituting -N[0MLD - ln(y)] = 2NiWi ln(y>) from (MLD) gives

~ 1 {-N[e^ -ln(y)]] + ^ + lnfe„).

MLD(i)

N - w;

_ Ny-yw Substituting y(j) by ——-— gives

'MLD (i)

N - w

N N - w

[««. - ln (y)] + + ln

N - w

N - w

((MLD ) (2mld )

(3MLD )

(4MLD )

(5MLD )

(6MLD )

is

V)

Si £

-J

' /

Atkinson Index

The general form of the Atkinson index is 0 A = 1 —

N (y/y )1-

1

1-£

. Below we de-

rive the jackknife formulas for two prominent case of the inequality aversion parameter, e.

Inequality aversion parameter e = 1

в = 1-^ = 11 У

exp

SWi ln ( y i )

N i=1

y

(1")

exp

в a (i ) -1 —

SWj ln (y. )

N w j*i 1 У 1 '

y(i)

(2A1)

ln (y. )w. ln (y. )w.

Expansion of the term in brackets in the numerator with-'—L--'—L, and substitu-

, - , Ny- yWi ■

Hon of y{i) by —- gives

N — w;

N - w N - w.

exp

в a (i ) -1--

N i N

N - w [ N

У Nw ln (y ))-

Ny - wyj N - W;

(3A1)

Substitution of the term — ^ _ wi ln (y.) (log of the geometric mean of income from the full sample) by ln (y*) gives N ' 1

exp

0 4 о ) =1-

ln(y )-

N - щ v ' N - w

Ny - wiyi N - W;

(4A1)

Inequality aversion parameter e = 2

0A =1-

N

SN

=1 ^

y

(i) -1-

Уг

N - W

2

w.

y» yj

(1A2 )

(2a2 )

Expansion of the denominator with wt — — wi —^ — and rewriting the sum gives

y yi y yi

0a2 (i) -1-

N - w

4 )

У

Sw.

y У

j

, y(t) У y(i) У

+ wi —---w —--

У У1 У У1

(3А 2 )

ft _1__N - w

UA~ (i) 1 /— _ \

2 ( y(i) y

y

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

yi

w

i

y» y

y yi

(4a2 )

2

From вА = 1-

N

wty/yt

the sum in the denominator gives

N - w

it follows that ^ wty/yl = N^( - 0a), and replacement of ;§

« U)

о

^wa -w......N. (5a2 ) a

y 1-в

A

'чУц )V \ yi I

4 -J

_ Ny- yw

Finally, substitution of y(i) by -— gives

(; N — w,

Oa (i) =1---Y . (бАг )

A2 (i) Ny- wy N _ w {Ny - wy ) V ;

y(N - w )1 -в, yt {N - w)

A2

Variance and Coefficient of Variation

1 "

» = -¡h 2 ( - y )2, (1F )

JV 1 ¡=1

_ (N-1) ■ V (i} (N - 2)(N - w )

0v(O wj (yj - У(г>)2 . )

Rewriting of 6V gives

1 ^ I \2 1

1S ( - y«)) + N —(y - y )2. (3V )

Substituting y = -1 [(n — wi )y(i) + yiwi ] and reorganizing in analogy to the variance of the logarithms gives

ev = C + D + E + N— Wi {yt - y )2, (4V )

C 1 V ( - Y « (N- 2)(N- w) ( ,

C=n— Vwj(--Y = —(N-Y—, (5 '

D = n— N % Wj ( - y0)(()- y') = (N (()- y' ) w (- ^)) =0, (6V)

E = ■

1

N-1

Analogously to 0VL we can rewrite (7V) as

i

1 W. / \2

E =--^ (y - y ) .

N-1 N - w Substitution (5V), (6V), and (8V) in (4V) gives (N - 2)(N - wt) 1

_ _(-_ )2 + Wi ( - -)2

^ (N-1)2 ' N-1 N - wt ( -i)+ N-1-)

Analogously to 0VL we can rewrite (9V) as

(N - 2){N- wt) , Мщ ( )2

= (n-i)2 + (n-)n-w7((-).

Solving (l 0V) for dV(i} gives

„ (N-1)2 Nwt (N-1)

V (i) WV

(N- 2)(N - wt) (N - Wi )2 (N - 2) The coefficient of variation is defined as

ecr =(?,- T'/y.

(y - yi 22.

Hence:

Qcv (i) -(V (i) ) /y(i) .

Substitution of

( N -l)2

V (i) WV

Nwi (N -l) (N - 2)(N - Wi) (n - Wi )2 (N - 2)

(y - y,)2

and of

y(i)

(N - Wi)

'CV (i)

(Ny-ywi) in (2CV) gi

gives

' (N-l)2___Nwi (N-l) (_- у'""'5

Uv (N - 2)(n - w ) (N - w )2 (N - 2)( _ )

(N-WT) (N_~ _W)

(8V )

(9V )

(10V )

(11V )

Г ) (2CV )

Г)

1

i Надоели баннеры? Вы всегда можете отключить рекламу.