7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
View/Download PDF

Translate this page into:

31 (
4
); 1167-1174
doi:
10.1016/j.jksus.2019.01.015

Goodness-of-fit testing for the Cauchy distribution with application to financial modeling

Department of Statistics, Hakim Sabzevari University, P.O. Box 397, Sabzevar, Iran
Department of Statistics, University of Isfahan, Isfahan 81746-73441, Iran

⁎Corresponding author. mahdizadeh.m@live.com (M. Mahdizadeh),

Disclaimer:
This article was originally published by Elsevier and was migrated to Scientific Scholar after the change of Publisher.

Peer review under responsibility of King Saud University.

Abstract

This article deals with goodness-of-fit test for the Cauchy distribution. Six new tests based on Kullback-Leibler information are proposed, and shown to be consistent. Monte Carlo evidence indicates that the tests have satisfactory performances against symmetric alternatives. An empirical application to quantitative finance is provided.

Keywords

Entropy
Fat-tailed distributions
Financial returns
1

1 Introduction

A Cauchy random variable with location parameter μ R and scale parameter σ > 0 , denoted by X C ( μ , σ ) , has probability density function

(1)
f 0 ( x ; μ , σ ) = π σ 1 + x - μ σ 2 - 1 , x R , and cumulative distribution function
(2)
F 0 ( x ; μ , σ ) = 1 2 + 1 π tan - 1 x - μ σ , x R .

Siméon Denis Poisson discovered the Cauchy distribution in 1824, long before its first mention by Augustin-Louis Cauchy. Early interest in the distribution focused on its value as a counterexample which demonstrated the need for regularity conditions in order to prove important limit theorems (see Stigler, 1974). Thanks to this special nature, the Cauchy distribution is sometimes considered as a pathological case. However, it can be used as a model for describing a wealth of phenomena. This is exemplified in the sequel.

This probability law describes the energy spectrum of an excited state of an atom or molecule, as well as an elementary particle resonant state. It can be shown quantum mechanically that whenever one has a state which decays exponentially with time, the energy width of the state is described by the Cauchy distribution (Roe, 1992). Winterton et al. (1992) showed that the source of fluctuations in contact window dimensions is variation in contact resistivity, and the contact resistivity is distributed as a Cauchy random variable. Kagan (1992) pointed out that the Cauchy distribution describes the distribution of hypocenters on focal spheres of earthquakes. An application of this distribution to study the polar and non-polar liquids in porous glasses is given by Stapf et al. (1996). Min et al. (1996) found that Cauchy distribution describes the distribution of velocity differences induced by different vortex elements. An example in the context of quantitative finance is provided in Section 4.

Many statistical procedures, employed in the above mentioned applications, assume that the random mechanism generating the data follows the Cauchy distribution. A parametric procedure usually hinges on the assumption of a particular distribution. It is, therefore, of utmost importance to assess the validity of the assumed distribution. This is accomplished by performing a goodness-of-fit test. In this article, we suggest six tests of fit for the Cauchy distribution. They are modifications of a test based on Kullback-Leibler (KL) information criterion, previously studied by Mahdizadeh and Zamanzade (2017). Information theory deals with stochastic processes as sources of information, or as models of communication channels; see, for example, Stone (2015). It is known to be a powerful tool in the study of communication and control in the animal and the machine (Wiener, 1961). Being an essential part of probability theory, information theory is also closely related to statistical inference (Kullback, 1997). Vinga (2014) and Bensadon (2016) provide some applications in biological sequence analysis, and machine learning, respectively. A large body of literature has grown around developing goodness-of-fit tests using the information-theoretic measures such as the entropy and the KL distance. This approach has been successfully applied for many distributions, including normal, uniform, exponential, inverse Gaussian and Laplace, among others. See for example Vasicek (1976), Dudewicz and van der Meulen (1981), Grzegorzewski and Wieczorkowski (1999), Mudholkar and Tian (2002), Choi and Kim (2006), Al-Omari and Haq (2016), Al-Omari and Zamanzade (2017), Al-Omari and Zamanzade (2018), Mahdizadeh (2017a,b), Zamanzade and Mahdizadeh (2017a,b).

Section 2 is given to a review of the existing tests. The new goodness-of-fit tests are presented in Section 3. Power properties of these tests are assessed by means of Monte Carlo simulations. The results are reported in Section 4. To illustrate the suggested procedures, a real data set is analyzed in Section 5. We end in Section 6 with a summary.

2

2 Review of the existing goodness-of-fit tests

Given a random sample X 1 , , X n from a population having a continuous density function f ( x ) , consider the problem of testing H 0 : f ( x ) = f 0 ( x ; μ , σ ) for some μ R and σ > 0 , where f 0 ( x ; μ , σ ) is given (1). The alternative hypothesis is H 1 : f ( x ) f 0 ( x ; μ , σ ) for any μ R and σ > 0 .

The Cauchy distribution is a peculiar distribution due to its heavy tail and the difficulty of estimating its parameters (see Johnson et al., 1994). First, the method of moment estimation fails since the mean and variance of the Cauchy distribution do not exist. Second, the maximum likelihood estimates of the parameters are very complex. We therefore estimate μ and σ by the median and the half-interquartile range which are attractive estimators because of their simplicity. Suppose X ( 1 ) X ( n ) are the sample order statistics, and ξ p ( 0 < p < 1 ) is the sample p th quantile. Then, the two estimators are given by

(3)
μ ̂ = ( X ( n / 2 ) + X ( n / 2 + 1 ) ) / 2 if n is even X ( ( n + 1 ) / 2 ) Otherwise , and
(4)
σ ̂ = 1 2 ( ξ 0.75 - ξ 0.25 ) .
Suppose F 0 ( x ; μ , σ ) is defined as in (2). The best-known statistic for tests of fit is that of Kolmogorov-Smirnov given by
(5)
KS = max i = 1 , , n max i n - F 0 ( X ( i ) ; μ ̂ , σ ̂ ) , F 0 ( X ( i ) ; μ ̂ , σ ̂ ) - i - 1 n .
Another powerful test, especially for small sample sizes, is based on the Anderson-Darling statistic defined as
(6)
A 2 = - 2 n i = 1 n ( i - 0.5 ) log F 0 ( X ( i ) ; μ ̂ , σ ̂ ) + ( n - i + 0.5 ) log 1 - F 0 ( X ( i ) ; μ ̂ , σ ̂ ) - n .
The famous Cramér-von Mises statistic,
(7)
W 2 = i = 1 n F 0 ( X ( i ) ; μ ̂ , σ ̂ ) - i - 0.5 n 2 + 1 12 n ,
leads to an important goodness-of-fit test. The above three test are based on weighted distance between true and empirical distribution functions. Gürtler and Henze (2000) proposed a test based on the empirical characteristic function Ψ n ( t ) = 1 n j = 1 n exp ( itY j ) of the standardized data Y j = ( X j - μ ̂ ) / σ ̂ , j = 1 , , n . The test statistic, D n , λ = n - Ψ n ( t ) - e - | t | 2 e - λ | t | d t , is the weighted L 2 distance between Ψ n and the characteristic function of the standard Cauchy distribution, where λ denotes a fixed positive weighting parameter. Large values of D n , λ imply rejection of H 0 . After some algebra, an alternative representation of D n , λ is derived as
(8)
D n , λ = 2 n j = 1 n k = 1 n λ λ 2 + ( Y j - Y k ) 2 - 4 j = 1 n 1 + λ ( 1 + λ ) 2 + Y j 2 + 2 n 2 + λ .
Remark 1

In practice, we use λ = 5 which leads to a powerful test according to the simulation results reported by Gürtler and Henze (2000).

Recently, Mahdizadeh and Zamanzade (2017) proposed four new tests of fit for the Cauchy distribution. The first three of them are modifications of the tests introduced by Zhang (2002). The corresponding test statistics are

(9)
Z K = max i = 1 , , n ( i - 0.5 ) log i - 0.5 nF 0 ( X ( i ) ; μ ̂ , σ ̂ ) + ( n - i + 0.5 ) log n - i + 0.5 n ( 1 - F 0 ( X ( i ) ; μ ̂ , σ ̂ ) ) ,
(10)
Z A = - i = 1 n log F 0 ( X ( i ) ; μ ̂ , σ ̂ ) n - i + 0.5 + log 1 - F 0 ( X ( i ) ; μ ̂ , σ ̂ ) i - 0.5 ,
and
(11)
Z C = i = 1 n log 1 / F 0 ( X ( i ) ; μ ̂ , σ ̂ ) - 1 ( n - 0.5 ) / ( i - 0.75 ) - 1 2 .
The fourth test utilizes the KL distance (see Kullback (1997)) between f and f 0 given by
(12)
D ( f , f 0 ; μ , σ ) = - f ( x ) log f ( x ) f 0 ( x ; μ , σ ) d x = - H ( f ) - - f ( x ) log f 0 ( x ; μ , σ ) d x ,
where H ( f ) is Shannon’s entropy of f defined as H ( f ) = - - f ( x ) log f ( x ) d x . The KL distance is the most common information criterion utilized for assessing model discrepancy. It is the expectation of the logarithm of the ratio of the probability density functions of two models, one being a fitted model and the other being the reference model, where the expectation is taken with respect to the reference model. Thus, the KL distance is a measure of the information loss in the fitted model relative to that in the reference model. It is well known that D ( f , f 0 ; μ , σ ) 0 and the equality holds if and only if f ( x ) = f 0 ( x ; μ , σ ) , almost surely. Therefore, D ( f , f 0 ; μ , σ ) can be regarded as a measure of disparity between f and f 0 .

Constructing a test based on (12) entails estimating the unknown quantities. The non-parametric estimation of H ( f ) has been studied by many authors. Vasicek (1976) introduced a simple estimator which has been widely used in developing tests of fit. His estimator is given by

(13)
HV m , n = 1 n i = 1 n log n 2 m X ( i + m ) - X ( i - m ) , where m (called window size) is a positive integer less than or equal to n / 2 , X ( 1 ) X ( n ) are order statistics based on a random sample of size n , X ( i ) = X ( 1 ) for i < 1 , and X ( i ) = X ( n ) for i > n . Vasicek (1976) showed that (13) is a consistent estimator of the population entropy. In particular, HV m , n p H ( f ) as m , n and m / n 0 , where p denotes convergence in probability. Also, - f ( x ) log f 0 ( x ; μ , σ ) d x can be estimated by
(14)
1 n i = 1 n log ( f 0 ( X i ; μ ̂ , σ ̂ ) ) ,
which is consistent by virtue of law of large numbers. Mahdizadeh and Zamanzade (2017) suggested to use
(15)
D ^ 1 = exp - HV m , n - 1 n i = 1 n log ( f 0 ( X i ; μ ̂ , σ ̂ ) )
as the final test statistic. Large values of D ^ 1 provide evidence against the null hypothesis.

It is difficult to derive the null distributions of (5)–(11) and (15) analytically. Monte Carlo simulations were then employed to determine critical values of a generic test statistic, say T. To this end, 50,000 samples were generated from C ( 0 , 1 ) for each sample size n = 10 , 20 , 30 , 50 , 100 , 200 . The estimators (3) and (4) were computed from any sample, and plugged into T. Finally, 1 - α quantile of the resulting values was determined which will be denoted by T 1 - α . The composite null hypothesis is rejected at level α if the observed value of T exceeds T 1 - α .

3

3 The proposed new tests

In this section, we introduce six new testing procedures for the Cauchy distribution. To clarify motivation of these tests, we first examine the entropy estimator component of statistic (15). It is worth noting that H ( f ) can be expressed as H ( f ) = 0 1 log d d p F - 1 ( p ) d p . Vasicek (1976) used the above representation to propose his nonparametric entropy estimator. In doing so, the involved derivative at each sample point ( X ( i ) , i / n ) is estimated by d i = X ( i + m ) - X ( i - m ) 2 m / n , where the order statistics and window size are defined as in Section 2. Now, HV m , n is simply defined to be the mean of logarithm of d i ’s for i = 1 , , n . Clearly, d i is not a correct formula when i m or i n - m + 1 . To fix this problem, the denominator and/or the numerator of d i should be adjusted. It is also possible to employ a fully different approach for entropy estimation. In the following, some improved entropy estimators are reviewed. These estimators are then incorporated in (15) to come up with new tests, which are expected to be more powerful.

Bowman (1992) studied the estimator

(16)
HB n = - 1 n i = 1 n log f ̂ ( X i ) , where f ̂ x = 1 nh j = 1 n K x - X j h , and K ( . ) is a symmetric kernel function which is chosen to be the standard normal density function. The bandwidth h is selected based on the normal optimal smoothing formula, h = 1.06 s n - 1 / 5 , where s is the sample standard deviation.

Van Es (1992) considered estimation of functionals of a probability density and entropy in particular. He proposed the following estimator

(17)
HVE m , n = 1 n - m i = 1 n - m log n + 1 m X ( i + m ) - X ( i ) + i = m n 1 i - log n + 1 m , where m is a positive integer less than n.

Ebrahimi et al. (1994) suggested two improved entropy estimators. The first one is equal to that of Vasicek plus a constant. This implies that the test based on this estimator is equivalent to D ^ 1 . So it is not included in this study. The second estimator is given by

(18)
HE m , n = 1 n i = 1 n log n d i m ( Y ( i + m ) - Y ( i - m ) ) , where d i = 1 + i + 1 m - i m 2 1 i m 2 m + 1 i n - m 1 + n - i m + 1 n - m + 1 i n , the Y ( i ) ’s are Y ( i - m ) = a + i - 1 m ( X ( 1 ) - a ) 1 i m Y ( i ) = X ( i ) m + 1 i n - m Y ( i + m ) = b - n - i m ( b - X ( n ) ) n - m + 1 i n , and a and b are constants to be determined such that P ( a X b ) 1 . For example, when F (the population distribution function) has a bounded support, a and b are lower and upper bound, respectively (for uniform(0,1) distribution, a = 0 and b = 1 ); if F is bounded below (above), then a ( b ) is lower (upper) support, a = x ¯ - ks ( b = x ¯ + ks ) , where x ¯ = 1 n i = 1 n x i , s 2 = 1 n - 1 i = 1 n ( x i - x ¯ ) 2 , and k is a suitable number say 3 to 5 (for exponential distribution, a = 0 and b = x ¯ + ks ); in the case that F has no bound on its support, a and b may be chosen as a = x ¯ - ks and b = x ¯ + ks .

Correa (1995) proposed another entropy estimator defined as

(19)
HC m , n = - 1 n i = 1 n log j = i - m i + m ( X ( j ) - X ( i ) ) ( j - i ) n j = i - m i + m ( X ( j ) - X ( i ) ) 2 , where X ( i ) = 1 2 m + 1 j = i - m i + m X ( j ) . Yousefzadeh and Arghami (2008) introduced the following entropy estimator
(20)
HY m , n = i = 1 n F ^ y X ( i + m ) - F ^ y X ( i - m ) j = 1 n F ^ y X ( j + m ) - F ^ y X ( j - m ) log X ( i + m ) - X ( i - m ) F ^ y X ( i + m ) - F ^ y X ( i - m ) ,
where for i = 2 , , n - 1 , F ^ y X ( i ) = n - 1 n ( n + 1 ) i + 1 n - 1 + X ( i ) - X ( i - 1 ) X ( i + 1 ) - X ( i - 1 ) , and F ^ y X ( 1 ) = 1 - F ^ y X ( n ) = 1 n + 1 . Alizadeh Noughabi (2010) developed an entropy estimator using kernel density estimator. His estimator is defined as
(21)
HA m , n = - 1 n i = 1 n log f ̂ X ( i + m ) - f ̂ X ( i - m ) 2 ,
where f ̂ ( x ) is just as given in (16).
Remark 2

In the all new entropy estimators which employ spacings of the order statistics, it is assumed that m is an integer satisfying 1 m n / 2 , unless otherwise stated.

The test statistics obtained by replacing HV m , n in (15) with HB n , HVE m , n , HE m , n , HC m , n , HY m , n and HA m , n will be denoted by D ^ 2 , D ^ 3 , D ^ 4 , D ^ 5 , D ^ 6 and D ^ 7 , respectively. Again, Monte Carlo approach is adopted to compute critical values of the resulting tests. To calculate test statistics based on the KL distance (with the exception of D ^ 2 ), the window size m corresponding to a given sample size must be selected in advance. In entropy estimation based on spacings, choosing optimal m for given n is still an open problem. For each n, the window size having smallest critical value tends to yield greater power. For sample sizes 10, 20, 30, 50, 100, and 200, window sizes producing the minimum critical values for different tests are given in Table 1. Table 2 contains 0.05 critical points of the tests considered in this study. For the KL distance based tests, the above mentioned optimal window sizes are used. These thresholds will be used in the next section to study the power properties.

Table 1 The optimal window sizes for the tests of size 0.05 based on the KL distance.
Statistic
n D ^ 1 D ^ 3 D ^ 4 D ^ 5 D ^ 6 D ^ 7
10 2 9 5 2 5 5
20 4 19 10 4 10 10
30 8 29 15 11 15 15
50 20 49 25 23 25 25
100 45 99 50 49 50 50
200 96 199 100 100 100 100
Table 2 0.05 critical points of the tests.
n
Statistic 10 20 30 50 100 200
KS 0.270 0.196 0.163 0.128 0.091 0.065
A 2 0.919 0.983 1.026 1.037 1.057 1.056
W 2 0.129 0.138 0.140 0.141 0.143 0.143
D n , λ 0.152 0.152 0.159 0.160 0.162 0.162
Z K 1.890 2.508 2.881 3.231 3.648 4.008
Z A 3.755 3.615 3.541 3.461 3.389 3.346
Z C 12.423 15.940 17.834 19.787 22.240 24.692
D ^ 1 2.088 1.464 1.244 0.940 0.576 0.327
D ^ 2 1.274 1.158 1.103 1.042 0.967 0.896
D ^ 3 1.332 0.975 0.763 0.526 0.302 0.163
D ^ 4 0.842 0.740 0.653 0.531 0.379 0.251
D ^ 5 1.757 1.263 1.042 0.734 0.410 0.213
D ^ 6 1.367 1.109 0.924 0.689 0.426 0.245
D ^ 7 1.117 0.865 0.740 0.614 0.522 0.467

The entropy estimators mentioned in this section are consistent. In proving this result for the estimators dependent on the window size, it is assumed that m / n 0 as m and n . See pertinent references for more details. The next proposition attends to optimal property of the tests based on the KL distance.

Proposition 1

The tests based on D ^ i , i = 1 , , 7 , are consistent.

Proof

Let X 1 , , X n be a random sample of size n from a population with density function f 0 ( x ; μ , σ ) given in (1). It is easy to see that for any μ R and σ > 0 ,  □

1 n i = 1 n log f 0 ( X i ; μ , σ ) a . s . E log f 0 ( X i ; μ , σ ) . We may now conclude that 1 n i = 1 n log f 0 ( X i ; μ ̂ , σ ̂ ) a . s . E log f 0 ( X i ; μ , σ ) , because μ ̂ and σ ̂ , defined in (3) and (4), are strongly consistent estimators. Let H n be a typical entropy estimator. From consistency of H n , we have H n p - E log f 0 ( X i ; μ , σ ) . Putting these together, it follows that under the null hypothesis D ^ i p 1 . Now, the result follows from the fact that D ^ i p d ( > 1 ) under the alternative hypothesis. 

4

4 Power comparisons

In this section, performances of the proposed tests are evaluated via Monte Carlo experiments. Toward this end, we considered nine families of alternatives:

  • t distribution with n degrees of freedom denoted by t n .

  • Normal distribution with mean μ and variance σ 2 denoted by N ( μ , σ 2 ) .

  • Logistic distribution with mean μ and variance π 2 σ 2 / 3 denoted by Lo ( μ , σ 2 ) .

  • Laplace distribution with mean μ and variance 2 σ 2 denoted by La ( μ , σ 2 ) .

  • Gumbel distribution with mean μ + σ γ (where γ is Euler’s constant) and variance π 2 σ 2 / 6 denoted by Gu ( μ , σ 2 ) .

  • Beta distribution with mean α / ( α + β ) denoted by Be ( α , β ) .

  • Gamma distribution with mean α β and variance α β 2 denoted by Ga ( α , β ) .

  • Mixture of the normal and Cauchy distributions with mixing probability p denoted by NC ( p , 1 - p ) . The distribution mixes N(0,1) and C(0,1) with weights p and 1 - p , respectively.

  • Tukey distribution with parameter h denoted by Tu ( h ) . It is distribution of the random variable Z exp { Zh 2 / 2 } with Z N ( 0 , 1 ) .

The members selected from the above families are t 3 , t 5 , N(0,1), Lo(0,1), La(0,1), Gu(0,1), Be(2,1), Ga(2,1), NC(0.3,0.7) and Tu(1). For each alternative, 50,000 samples of sizes n = 10 , 20 , 30 , 50 were generated, and the power of each test was estimated by the percentages of samples entering the rejection region. Tables 3–6 present the estimated powers of the fourteen tests of size 0.05, given in Sections 2 and 3, for different sample sizes (the results for n = 100 , 200 are provided as Supplementary material). To provide enough space for the outputs, the reference to parameters of the distributions is only made in the case of t distribution. For each alternative, power entry associated with the best test among D ^ i ’s is in bold. In addition, the highest power value from the other tests is in italic.

Table 3 Power comparison for the tests of size 0.05 against several alternative distributions for n = 10 .
Alternative
Statistic t 3 t 5 N Lo La Gu Be Ga NC Tu
KS 0.028 0.028 0.031 0.028 0.028 0.048 0.097 0.086 0.040 0.061
A 2 0.013 0.013 0.016 0.012 0.013 0.024 0.051 0.044 0.037 0.069
W 2 0.028 0.028 0.033 0.028 0.028 0.047 0.085 0.076 0.040 0.063
D n , λ 0.005 0.004 0.003 0.003 0.008 0.003 0.002 0.003 0.032 0.067
Z K 0.012 0.012 0.013 0.011 0.012 0.024 0.057 0.048 0.041 0.068
Z A 0.016 0.016 0.021 0.016 0.016 0.033 0.079 0.060 0.041 0.067
Z C 0.008 0.010 0.014 0.010 0.010 0.020 0.054 0.036 0.042 0.068
D ^ 1 0.114 0.145 0.202 0.159 0.103 0.210 0.423 0.282 0.061 0.046
D ^ 2 0.177 0.230 0.329 0.257 0.158 0.295 0.502 0.322 0.074 0.038
D ^ 3 0.178 0.232 0.330 0.260 0.162 0.296 0.515 0.326 0.074 0.038
D ^ 4 0.192 0.252 0.356 0.280 0.172 0.304 0.496 0.311 0.077 0.037
D ^ 5 0.123 0.157 0.220 0.173 0.110 0.225 0.441 0.298 0.063 0.045
D ^ 6 0.168 0.219 0.312 0.245 0.153 0.286 0.512 0.325 0.072 0.038
D ^ 7 0.156 0.201 0.287 0.224 0.142 0.277 0.506 0.335 0.069 0.039
Table 4 Power comparison for the tests of size 0.05 against several alternative distributions for n = 20 .
Alternative
Statistic t 3 t 5 N Lo La Gu Be Ga NC Tu
KS 0.042 0.049 0.063 0.052 0.040 0.127 0.343 0.279 0.043 0.060
A 2 0.028 0.036 0.059 0.042 0.027 0.095 0.247 0.183 0.036 0.072
W 2 0.039 0.048 0.065 0.052 0.038 0.105 0.231 0.192 0.039 0.061
D n , λ 0.035 0.059 0.122 0.073 0.031 0.126 0.365 0.184 0.030 0.078
Z K 0.030 0.038 0.059 0.042 0.026 0.137 0.417 0.333 0.041 0.070
Z A 0.094 0.140 0.261 0.167 0.076 0.313 0.688 0.492 0.053 0.059
Z C 0.061 0.098 0.195 0.118 0.049 0.218 0.565 0.343 0.047 0.068
D ^ 1 0.354 0.501 0.739 0.573 0.334 0.733 0.974 0.852 0.084 0.036
D ^ 2 0.441 0.606 0.812 0.675 0.404 0.765 0.965 0.798 0.098 0.030
D ^ 3 0.416 0.592 0.840 0.678 0.428 0.716 0.968 0.717 0.086 0.032
D ^ 4 0.453 0.639 0.873 0.728 0.456 0.711 0.947 0.669 0.091 0.031
D ^ 5 0.365 0.520 0.761 0.593 0.351 0.745 0.975 0.852 0.085 0.035
D ^ 6 0.410 0.581 0.826 0.666 0.416 0.742 0.977 0.779 0.086 0.032
D ^ 7 0.301 0.412 0.611 0.465 0.274 0.712 0.966 0.893 0.084 0.035
Table 5 Power comparison for the tests of size 0.05 against several alternative distributions for n = 30 .
Alternative
Statistic t 3 t 5 N Lo La Gu Be Ga NC Tu
KS 0.058 0.071 0.106 0.078 0.046 0.247 0.661 0.546 0.048 0.060
A 2 0.050 0.077 0.146 0.092 0.040 0.224 0.560 0.401 0.036 0.075
W 2 0.055 0.072 0.113 0.082 0.047 0.191 0.445 0.348 0.043 0.061
D n , λ 0.123 0.225 0.455 0.283 0.100 0.417 0.811 0.519 0.031 0.087
Z K 0.066 0.099 0.189 0.116 0.047 0.417 0.864 0.776 0.046 0.073
Z A 0.245 0.392 0.669 0.474 0.203 0.731 0.974 0.897 0.065 0.054
Z C 0.172 0.300 0.564 0.371 0.141 0.587 0.933 0.764 0.050 0.068
D ^ 1 0.584 0.791 0.974 0.880 0.633 0.962 1 0.988 0.086 0.030
D ^ 2 0.674 0.855 0.975 0.914 0.659 0.960 1 0.964 0.105 0.026
D ^ 3 0.602 0.811 0.986 0.906 0.690 0.918 1 0.907 0.080 0.030
D ^ 4 0.655 0.861 0.993 0.941 0.738 0.920 0.999 0.887 0.087 0.029
D ^ 5 0.614 0.819 0.983 0.905 0.670 0.965 1 0.983 0.088 0.029
D ^ 6 0.604 0.811 0.983 0.901 0.678 0.947 1 0.963 0.083 0.030
D ^ 7 0.434 0.596 0.841 0.673 0.412 0.935 1 0.997 0.087 0.031
Table 6 Power comparison for the tests of size 0.05 against several alternative distributions for n = 50 .
Alternative
Statistic t 3 t 5 N Lo La Gu Be Ga NC Tu
KS 0.095 0.137 0.253 0.151 0.063 0.583 0.976 0.928 0.054 0.058
A 2 0.142 0.261 0.517 0.316 0.099 0.619 0.955 0.847 0.040 0.079
W 2 0.096 0.148 0.281 0.169 0.069 0.421 0.828 0.689 0.047 0.064
D n , λ 0.400 0.644 0.906 0.740 0.328 0.874 0.997 0.937 0.043 0.099
Z K 0.228 0.385 0.701 0.462 0.168 0.933 1 0.998 0.058 0.075
Z A 0.622 0.853 0.988 0.924 0.603 0.995 1 1 0.088 0.051
Z C 0.514 0.774 0.970 0.866 0.482 0.974 1 0.996 0.063 0.069
D ^ 1 0.815 0.965 1 0.996 0.939 1 1 1 0.083 0.029
D ^ 2 0.917 0.992 1 0.998 0.947 1 1 1 0.115 0.022
D ^ 3 0.803 0.960 1 0.996 0.949 0.995 1 0.993 0.078 0.030
D ^ 4 0.852 0.979 1 0.999 0.970 0.997 1 0.995 0.086 0.028
D ^ 5 0.820 0.967 1 0.997 0.945 1 1 1 0.083 0.029
D ^ 6 0.813 0.964 1 0.996 0.943 0.999 1 1 0.080 0.029
D ^ 7 0.786 0.943 0.999 0.979 0.827 1 1 1 0.094 0.027

It is observed that no single test is uniformly most powerful. We note, however, that the tests based on the KL distance are generally more powerful than the other tests. Compare the bold and italic entries for each alternative. Given a distribution and sample size, difference of the italic entry from the bold one is reported in Table 7. The values are sizable for symmetric distributions like t 3 , t 5 , N(0,1), Lo(0,1), La(0,1) and Gu(0,1). All of the tests perform poorly when the parent distribution is either NC(0.3,0.7) or Tu(1), and increasing the sample size does not give rise to marked improvement in power.

Table 7 Power differences between the best test among D ^ i ’s and the best of other tests.
Alternative
n t 3 t 5 N Lo La Gu Be Ga NC Tu
10 0.164 0.224 0.323 0.252 0.144 0.256 0.418 0.249 0.035 −0.023
20 0.359 0.499 0.612 0.561 0.380 0.452 0.289 0.401 0.045 −0.042
30 0.429 0.469 0.324 0.467 0.535 0.234 0.026 0.100 0.040 −0.056
50 0.295 0.139 0.012 0.075 0.367 0.005 0 0 0.027 −0.069

With the exception of sample size 10, Z A is generally the best among KS, A 2 , W 2 , D n , λ , Z K , Z A and Z C tests. Moreover, it can be seen that either D ^ 2 or D ^ 4 has mostly the best performance among D ^ i ’s.

5

5 Example

Heavy-tailed distributions, like Cauchy, are better models for financial returns because the normal model does not capture the large fluctuations seen in real assets. Nolan (2014) provides an accessible introduction to financial modeling using such distributions.

The stock market return is the return that we obtain from stock market by buying and selling stocks or get dividends by the company whose stock you hold. The stock market price is usually modeled by lognormal distribution, that is to say stock market returns follow the Gaussian law. The feature of stock market return distribution is a sharp peak and heavy tails. The Gaussian distribution clearly does not enjoy these attributes. So the Cauchy distribution may be a potential model. The German Stock Index (DAX) is the major stock market index in Germany which contains the stocks of 30 largest German companies trading on the Frankfurt Stock Exchange. The DAX evaluates the Prime Standard of those 30 major German companies trading on the Frankfurt Stock Exchange. We now apply the fourteen goodness-of-fit tests to a real dataset containing 30 returns of closing prices of the DAX. The data are observed daily from January 1, 1991, excluding weekends and public holidays. The data (rounded up to seven decimal places) are given in Table 8, which are obtained from datasets package in R statistical software. The Cauchy Q-Q plot appears in Fig. 1. The corresponding histogram, superimposed by a Cauchy density function, is also included. The location and scale parameters estimated from the data are μ ̂ = 0.0009629174 and σ ̂ = 0.003635871 .

Table 8 Scores for 30 returns of closing prices of DAX.
0.0011848 −0.0057591 −0.0051393 −0.0051781 0.0020043 0.0017787
0.0026787 −0.0066238 −0.0047866 −0.0052497 0.0004985 0.0068006
0.0016206 0.0007411 −0.0005060 0.0020992 −0.0056005 0.0110844
−0.0009192 0.0019014 −0.0042364 0.0146814 −0.0002242 0.0024545
−0.0003083 −0.0917876 0.0149552 0.0520705 0.0117482 0.0087458
The Cauchy Q-Q plot of the 30 returns, and the corresponding histogram along with fitted Cauchy density.
Fig. 1
The Cauchy Q-Q plot of the 30 returns, and the corresponding histogram along with fitted Cauchy density.

The values of all statistics are computed (see Table 9), and compared with the corresponding critical values in Table 2. By using any test, the null hypothesis that the data follow the Cauchy distribution is not rejected at 0.05 significance level.

Table 9 Observed values of the different statistics.
KS A 2 W 2 D n , λ Z K Z A Z C
0.126 0.498 0.076 0.051 1.343 3.346 5.761
D ^ 1 D ^ 2 D ^ 3 D ^ 4 D ^ 5 D ^ 6 D ^ 7
0.661 0.844 0.255 0.302 0.386 0.358 0.461

6

6 Conclusion

This article concerns goodness-of-fit test for the Cauchy distribution. Six tests based on the KL information criterion are developed, and shown to be consistent. A simulation study is carried out to compare the performances of the new tests with their contenders. In doing so, five sample sizes and nine families of alternatives are considered. It emerges that the new tests are powerful against many symmetric distributions. The proposed procedures are finally applied on real data example.

Acknowledgments

We thank the reviewers for their constructive remarks that helped us to improve this article significantly.

References

  1. , . A new estimator of entropy and its application in testing normality. J. Stat. Comput. Simul.. 2010;80:1151-1162.
    [Google Scholar]
  2. , , . Entropy estimation and goodness-of-fit tests for the inverse Gaussian and Laplace distributions using pair ranked set sampling. J. Stat. Comput. Simul.. 2016;86:2262-2272.
    [Google Scholar]
  3. , , . Goodness-of-fit tests for Laplace distribution using ranked set sampling. Revista Investigacion Operacional. 2017;38:366-376.
    [Google Scholar]
  4. , , . Goodness-of-fit tests for logistic distribution based on Phi-divergence. Electron. J. Appl. Stat. Anal.. 2018;11:185-195.
    [Google Scholar]
  5. , . Applications of Information Theory to Machine Learning. Université Paris-Saclay; . (Ph.D. thesis)
  6. , . Density based tests for goodness-of-fit. J. Stat. Comput. Simul.. 1992;40:1-13.
    [Google Scholar]
  7. , , . Testing goodness-of-fit for laplace distribution based on maximum entropy. Statistics. 2006;40:517-531.
    [Google Scholar]
  8. , . A new estimator of entropy. Commun. Stat.: Theory Methods. 1995;24:2439-2449.
    [Google Scholar]
  9. , , . Entropy-based tests of uniformity. J. Am. Stat. Assoc.. 1981;76:967-974.
    [Google Scholar]
  10. , , , . Two measures of sample entropy. Stat. Probab. Lett.. 1994;20:225-234.
    [Google Scholar]
  11. , , . Entropy based goodness-of-fit test for exponentiality. Commun. Stat.: Theory Methods. 1999;28:1183-1202.
    [Google Scholar]
  12. , , . Goodness-of-fit tests for the Cauchy distribution based on the empirical characteristic function. Ann. Inst. Stat. Math.. 2000;52:267-286.
    [Google Scholar]
  13. , , , . Continuous Univariate Distributions. Vol vol. 1. New York: Wiley; . second ed.
  14. , . Correlations of earthquake focal mechanisms. Geophys. J. Int.. 1992;110:305-320.
    [Google Scholar]
  15. , . Information Theory and Statistics. New York: Dover Publications; .
  16. , . On testing uniformity using an information-theoretic measure. Commun. Stat.: Simul. Comput.. 2017;46:6173-6196.
    [Google Scholar]
  17. , . Test of fit for the Rayleigh distribution in ranked set sampling. J. Stat. Manage. Syst.. 2017;20:901-915.
    [Google Scholar]
  18. , , . New goodness-of-fit tests for the Cauchy distribution. J. Appl. Stat.. 2017;44:1106-1121.
    [Google Scholar]
  19. , , , . Levy stable distributions for velocity and velocity difference in systems of vortex elements. Phys. Fluids. 1996;8:1169-1180.
    [Google Scholar]
  20. , , . An entropy characterization of the inverse Gaussian distribution and related goodness-of-fit test. J. Stat. Planning Inference. 2002;102:211-221.
    [Google Scholar]
  21. , . Financial modeling with heavy-tailed stable distributions. WIREs Comput. Stat.. 2014;6:45-55.
    [Google Scholar]
  22. , . Probability and Statistics in Experimental Physics. New York: Springer; .
  23. , , , , , . Proton and deuteron field-cycling NMR relaxometry of liquids confined in porous glasses. Colloids Surf.: A. 1996;115:107-114.
    [Google Scholar]
  24. , . Cauchy and the witch of Agnesi: an historical note on the Cauchy distribution. Biometrika. 1974;61:375-380.
    [Google Scholar]
  25. , . Information Theory: A Tutorial Introduction. Sebtel Press; .
  26. , . Estimating functionals related to a density by class of statistics based on spacings. Scand. J. Stat.. 1992;19:61-72.
    [Google Scholar]
  27. , . A test of normality based on sample entropy. J. R. Stat. Soc. B. 1976;38:54-59.
    [Google Scholar]
  28. , . Information theory applications for biological sequence analysis. Briefings Bioinf.. 2014;15:376-389.
    [Google Scholar]
  29. , . Cybernetics or Control and Communication in the Animal and the Machine (second ed.). New York: MIT Press; .
  30. , , , . On the source of scatter in contact resistance data. J. Electron. Mater.. 1992;21:917-921.
    [Google Scholar]
  31. , , . Testing exponentiality based on type II censored data and a new CDF estimator. Commun. Stat.: Simul. Comput.. 2008;37:1479-1499.
    [Google Scholar]
  32. , , . Entropy estimation from ranked set samples with application to test of fit. Colombian J. Stat.. 2017;40:223-241.
    [Google Scholar]
  33. , , . Goodness of fit tests for Rayleigh distribution based on Phi-divergence. Colombian J. Stat.. 2017;40:279-290.
    [Google Scholar]
  34. , . Powerful goodness-of-fit tests based on the likelihood ratio. J. R. Stat. Soc. B. 2002;64:281-294.
    [Google Scholar]

Appendix A

Supplementary data

Supplementary material associated with this article can be found, in the online version, at https://doi.org/10.1016/j.jksus.2019.01.015.

Supplementary data

The following are the Supplementary data to this article:

Show Sections