7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
View/Download PDF

Translate this page into:

Original article
32 (
1
); 172-182
doi:
10.1016/j.jksus.2018.04.005

Kumaraswamy generalized Kappa distribution with application to stream flow data

Department of Statistics, Government College University, Faisalabad, Pakistan

⁎Corresponding author. tahir.nawaz@gcuf.edu.pk (Tahir Nawaz),

Disclaimer:
This article was originally published by Elsevier and was migrated to Scientific Scholar after the change of Publisher.

Peer review under responsibility of King Saud University.

Abstract

Recently in literature many families of distributions have been introduced to study the skewness, kurtosis and to explore the shape of the distribution more intensely. These families of distribution have wider applicability in variety of fields. In this paper, we introduce a five-parameter distribution, called the Kumaraswamy generalized Kappa distribution which extends the three-parameter Kappa distribution. The new distribution is more flexible and is applicable in the study of the highly-skewed data. Some mathematical properties of the proposed distribution are studied that includes the explicit expression for generating functions, moments, inequality indices, and entropies. The maximum likelihood estimates are computed using the numerical procedure. An application of the Kumaraswamy generalized Kappa distribution is illustrated using a real data set on stream flow.

Keywords

Entropy
Inequality indices
Kumaraswamy distribution
Kappa distribution
Maximum likelihood
1

1 Introduction

Many families of distribution have been recently proposed which are more flexible and have wider applicability ranging from survival analysis, reliability engineering, and related fields. Classical distributions do not provide adequate fits to the real data which are highly skewed. To overcome this drawback numerous methods of introducing additional shape parameters, and generating new families of distributions are available in the statistical literature.

Some well-known generators are the Marshall and Olkin (1997), exponentiated generalized (exp-G) class of distributions based on Lehmann-type alternatives suggested by Gupta et al. (1998), generalized-exponential (GE) also known as exponentiated exponential (EE) distributions introduced by Gupta and Kundu (1999), beta-generated distributions proposed by Eugene et al. (2002) and Jones (2004), Kumaraswamy generalized (Kum-G) distribution suggested by Cordeiro and de Castro (2011), McDonald generalized (Mc-G) distribution introduced by Alexander et al. (2012), gamma-generated type-1 distributions proposed by Zografos and Balakrishnan (2009) and Amini et al. (2014), gamma-generated type-2 distributions by Ristic and Balakrishnan (2012) and Amini et al. (2014), exponentiated generalized (exp-G) distribution introduced by Cordeiro et al. (2013) and odd Weibull-generated distribution proposed by Bourguignon et al. (2014). The induction of one, two or three more shape parameters to the base-line distribution increases the chances to investigate skewness and vary tail weights. The earlier mentioned generalizations also help to deduce sub-model of generalized distributions that greatly enhances the applicability.

There is no general class of distribution to model skewed data in every practical situation. Mielke (1973) presented a class of asymmetric positively skewed distributions, known as the Kappa distribution, for explaining and examining rainfall data and weather modification. Mielke and Johnson (1973) presented the maximum likelihood estimates and the likelihood ratio tests for the three-parameter Kappa distribution. The Kappa distribution has obtained attention from the hydrologic experts. Conventionally, the log normal and gamma distributions are fitted to precipitation data but these distributions have their own limitations due to non-existence of closed forms of the cdfs and quantile functions. The class of Kappa distribution have closed algebraic expressions that can easily be analysed.

Let X be a three-parameter Kappa random variable, then the pdf and cdf of the Kappa distribution are given by

(1.1)
f ( x ) = α θ β x β θ - 1 α + x β α θ - ( α + 1 ) α and
(1.2)
F ( x ) = x β α θ α + x β α θ 1 α
α , β , θ > 0 and x > 0 .

Parameter β is scale while parameters α and θ are shape.

In this paper [deduced from Hussain (2015)], we have extended the three parameter Kappa distribution, namely the Kumaraswamy generalized Kappa (KGK) distribution by introducing additional shape parameters. The paper is organized as follows. In Section 2 we have defined KGK distribution, its sub-models and the behaviour of its pdf. The mathematical properties such as: quantile functions, moments, and generating functions are derived in Section 3. The Reliability properties are presented in Section 4. Some inequality indices are given in Section 5. Entropies are discussed in Section 6. Maximum likelihood estimation of the parameters of the distribution is presented in Section 7. The application on real life data set of stream flows amount is provided in Section 8. Finally, Section 9 concludes the manuscript.

2

2 Kumaraswamy generalized Kappa distribution

Cordeiro and de Castro (2011) defined a general class of general distributions for double-bounded random process based on Kumaraswamy (1980) which is known as the Kumaraswamy generalized (Kum-G) distribution.Let X be a random variable, the cdf of Kum-G class of distribution is given by

(2.1)
F Kum - G ( x ) = 1 - [ 1 - { F ( x ) } a ] b where a and b are additional shape parameters to generalized distribution, which govern skewness and tail weights. The probability density function ( pdf ) corresponding to (2.1) is
(2.2)
f Kum - G ( x ) = abf ( x ) { F ( x ) } a - 1 [ 1 - { F ( x ) } a ] b - 1
Inserting (1.2) in (2.1), the cdf of the Kumaraswamy generalized Kappa (KGK) distribution is given by
(2.3)
F KGK ( x ) = 1 - 1 - x β α θ α + x β α θ a α b ; x > 0
The pdf corresponding to (2.3), will be
(2.4)
f KGK ( x ) = ab α θ β x β θ - 1 α + x β α θ - ( α + 1 ) α x β α θ α + x β α θ a - 1 α 1 - x β α θ α + x β α θ a α b - 1 ; x > 0
with a , b , α , β , θ > 0 , where a , b , α and θ are shape parameters while β is scale parameter.

2.1

2.1 Special sub-models

The KGK has the following distributions as special sub-models.

  • Exponentiated Kappa Lehmann type II (EK-L2) distribution: If a = 1 , the KGK reduces to the exponentiated Kappa (Lehmann type II) distribution with parameters b , α , β and θ .

  • Exponentiated Kappa Lehmann type I (EK-L1) distribution: If b = 1 , the KGK reduces to the exponentiated Kappa (Lehmann type I) distribution with parameters a , α , β and θ .

  • Three-parameter Kappa distribution: If a = b = 1 , the KGK reduces to three-parameter Kappa distribution with parameter α , β and θ .

  • Two-parameter Kappa distribution: If a = b = θ = 1 , the KGK reduces to two-parameter Kappa distribution with parameter α and β .

  • One-parameter Kappa distribution: If a = b = θ = β = 1 , the KGK reduces to one-parameter Kappa distribution with parameter α .

2.2

2.2 A useful representation of KGK density function

Using the binomial expansion, ( 1 - ω ) b - 1 = j = 0 ( - 1 ) j b - 1 j ω j , where b is real non-integer and | ω | < 1 .

The KGK density given in (2.4) in more simplified form can be expressed as

(2.5)
f KGK ( x ) = ab j = 0 ( - 1 ) j b - 1 j α θ β x β θ a ( j + 1 ) - 1 α + x β α θ - a α ( j + 1 ) - 1 If a is an integer, Eq. (2.5) reveals that KGK density function is equals to Kappa density function multiplied by an infinite power series of W j .
(2.6)
f KGK ( x ) = W j α θ β x β θ a ( j + 1 ) - 1 α + x β α θ - a α ( j + 1 ) - 1
where W j = ab j = 0 ( - 1 ) j b - 1 j .

Graphical representation of the density function for various values of the parameters selected arbitrarily provided in Fig. 1. It can be observed from the Fig. 1 that increasing the value of the shape parameters ( a , b , α , θ ) the peakedness of the density function tends to increase. Similarly, the increase in the value of the scale parameter β shifts the density function away from the origin.

Plots of density function of KGK distribution by varying the parameter values.
Fig. 1
Plots of density function of KGK distribution by varying the parameter values.

3

3 Statistical properties

In this section, various statistical properties of the KGK distribution, viz. Quantile function, median, random number generation, mode, moments, moment generating function (mgf), characteristic function (cf), mean deviation from mean and mean deviation from median.

3.1

3.1 Quantile function and random number generation

The qth quantile, x q , of the KGK distribution can be obtained by 1 - 1 - x q β α θ α + x q β α θ a α b = p , Now, for solving x q , gives us quantile function of the Kum generalized Kappa (KGK) distribution.

(3.1)
x q = β α 1 α θ 1 - ( 1 - p ) 1 b 1 a θ 1 - 1 - ( 1 - p ) 1 b α a - 1 α θ , 0 < p < 1 The median ( Q 2 ) of the KGK distribution will be
(3.2)
Q 2 = x 0.50 = β α 1 α θ 1 - ( 1 - 0.50 ) 1 b 1 a θ 1 - 1 - ( 1 - 0.50 ) 1 b α a - 1 α θ
The lower quartile ( Q 1 ) and upper quartile ( Q 3 ) of the KGK distribution are Q 1 = x 0.25 = β α 1 α θ 1 - ( 1 - 0.25 ) 1 b 1 a θ 1 - 1 - ( 1 - 0.25 ) 1 b α a - 1 α θ and Q 3 = x 0.75 = β α 1 α θ 1 - ( 1 - 0.75 ) 1 b 1 a θ 1 - 1 - ( 1 - 0.75 ) 1 b α a - 1 α θ By the inversion method, the random numbers from KGK distribution can be generated as 1 - 1 - ( x β ) α θ α + x β α θ a α b = u ; u U ( 0 , 1 ) After simplification, we get
(3.3)
x u = β α 1 α θ 1 - ( 1 - u ) 1 b 1 a θ 1 - 1 - ( 1 - u ) 1 b α a - 1 α θ

3.2

3.2 Mode

The mode of the KGK is obtained as logf ( x ) = log ab α θ β + ( θ a - 1 ) log x β - α + a α log α + x β α θ + ( b - 1 ) × log 1 - ( x β ) α θ α + x β α θ a α The first derivative of logf ( x ) for the KGK distribution is x logf ( x ) . So, the modes of the KGK distribution are the roots of the following equation x logf ( x ) = 0 which gives

(3.4)
θ a - 1 x - θ ( α + a ) x β α θ - 1 β α + x β α θ - a ( b - 1 ) α θ x β a θ - 1 β α + x β α θ a α + 1 - α x β a θ - x β θ ( α + a ) = 0 There may be more than one root to Eq. (3.4). If x = x 0 is a root of Eq. (3.4) then it corresponds to a local maximum or local minimum or a point of inflection depending on whether 2 x 2 logf ( x ) < or > or = 0 , respectively.

3.3

3.3 rth moment

The r th moment for KGK random variable X is given by μ r = E ( X r ) = - x r f KGK ( x ) dx using (2.6), we have μ r = 0 x r W j α θ β x β θ a ( j + 1 ) - 1 α + x β α θ - a α ( j + 1 ) - 1 dx Let Z = x β α θ then dx = β Z 1 α θ α θ Z dZ After simplification, we get μ r = W j β r α a α ( j + 1 ) + 1 0 Z r α θ + a α ( j + 1 ) - 1 1 - 1 - 1 + Z α - 1 a α ( j + 1 ) + 1 dZ Let W = 1 - 1 + Z α - 1 and dZ = α ( 1 - W ) 2 dW

After simplification, we get μ r = W j β r α r α θ - 1 0 1 W r α θ + a α ( j + 1 ) - 1 [ 1 - W ] 1 - r α θ - 1 dW Using beta function,

(3.5)
μ r = W j β r α r α θ - 1 B r α θ + a α ( j + 1 ) , 1 - r α θ ; r = 1 , 2 , 3 , 4 where B ( a , b ) = 0 1 U a - 1 [ 1 - U ] b - 1 dU is the beta function.

The mean and the variance of the KGK distribution are, respectively, given by

(3.6)
mean = W j β α 1 α θ - 1 B 1 α θ + a α ( j + 1 ) , 1 - 1 α θ
(3.7)
Variance = W j β 2 α 2 α θ - 1 B 2 α θ + a α ( j + 1 ) , 1 - 2 α θ - W j β α 1 α θ - 1 B 1 α θ + a α ( j + 1 ) , 1 - 1 α θ 2
where W j = ab j = 0 ( - 1 ) j b - 1 j .

3.4

3.4 Moment generating function and characteristic function

The moment generating function and characteristic function of the KGK distribution are given by

(3.8)
M ( t ) = E ( e tx ) = r = 0 t r r ! μ r = r = 0 t r r ! W j β r α r α θ - 1 B r α θ + a α ( j + 1 ) , 1 - r α θ ; r = 1 , 2 , 3 , 4 and
(3.9)
φ ( t ) = E ( e itx ) = r = 0 ( it ) r r ! μ r = r = 0 ( it ) r r ! W j β r α r α θ - 1 B r α θ + a α ( j + 1 ) , 1 - r α θ ; r = 1 , 2 , 3 , 4
where W j = ab j = 0 ( - 1 ) j b - 1 j .

4

4 Reliability properties

In this section, we have derived the reliability properties such as survival (or reliability) function, hazard rate, reversed hazard rate, cumulative hazard rate, mean residual life and mean waiting time for the Kumaraswamy generalized Kappa distribution.

A life time random variable t is said to have KGK ( a , b , α , β , θ ) distribution, when its cdf and pdf are represented as follows

(4.1)
F KGK ( t ) = 1 - 1 - t β α θ α + t β α θ a α b ; t > 0 and
(4.2)
f KGK ( t ) = ab α θ β t β θ a - 1 α + t β α θ - ( α + a ) α 1 - t β α θ α + t β α θ a α b - 1
t > 0 , a , b , α , β , θ > 0 .

The survival (or reliability) function, S ( t ) , is given by

(4.3)
S ( t ) = 1 - F ( t ) = 1 - t β α θ α + t β α θ a α b The failure rate or hazard rate function, h ( t ) , for KGK distribution is given by
(4.4)
h ( t ) = f ( t ) S ( t ) = ab α θ β t β θ a - 1 α + t β α θ - ( α + a ) α 1 - t β α θ α + t β α θ a α - 1
Various shape of the hazard function for different parameter values are presented in Fig. 2.
Plots of hazard rate of KGK distribution by varying the parameter values.
Fig. 2
Plots of hazard rate of KGK distribution by varying the parameter values.

It is observable from Fig. 2 that the hazard rate of the KGK distribution tends to increase initially, and then after reaching a certain level it starts decreasing. This indicates that the KGK distribution can be useful to model first increasing then decreasing hazard rate.

The cumulative hazard rate function, H ( t ) , and reversed hazard rate, r ( t ) , of KGK distribution are given respectively as

(4.5)
H ( t ) = 0 t h ( t ) dt = - lnS ( t ) = - bln 1 - t β α θ α + t β α θ a α and
(4.6)
r ( t ) = f ( t ) F ( t ) = ab α θ β t β θ a - 1 α + t β α θ - ( α + a ) α 1 - { t β α θ α + t β α θ } a α b - 1 1 - 1 - t β α θ α + t β α θ a α b
The mean residual life (MRL) or life expectancy at a given time t measures the expected remaining life time of an individual of age t . It is defined as
(4.7)
m ( t ) = 1 S ( t ) t tf ( t ) dt - t
which can also be written as m ( t ) = 1 S ( t ) E ( t ) - 0 t tf ( t ) dt - t Now consider t tf ( t ) dt = t tW j α θ β t β θ a ( j + 1 ) - 1 α + t β α θ - a α ( j + 1 ) - 1 dt After simplification and using incomplete beta function, we get
(4.8)
= W j β α 1 α θ - 1 B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ
where W = 1 - 1 + Z α - 1 , Z = t β α θ and B W ( a , b ) = 0 t t a - 1 [ 1 - t ] b - 1 dt is the incomplete beta function.

Using the results of (4.8) in (4.7), the MRL of the KGK distribution obtained as

(4.9)
m ( t ) = W j β α 1 α θ - 1 B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ - B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 1 - t β α θ α + t β α θ a α b - t The mean waiting time (MWT) of an item failed in an interval [ 0 , t ] is defined as
(4.10)
μ ¯ ( t , θ ) = t - 1 F ( t ) 0 t tf ( t ) dt
The MWT of the KGK distribution is given as
(4.11)
μ ¯ ( t , θ ) = t - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 1 - 1 - t β α θ α + t β α θ a α b - 1

5

5 Inequality measures

In this section, inequality measures such as Gini index, Lorenz curve, Bonferroni curve, Zenga index, Atkinson index, Pietra index and generalized entropy for the KGK distribution have been obtained.

5.1

5.1 Gini index

The most well-known inequality index is Gini index, suggested by Gini (1914), is defined as

(5.1)
G = 1 E ( x ) 0 F ( x ) { 1 - F ( x ) } dx Consider
(5.2)
F ( x ) { 1 - F ( x ) } = 1 - x β α θ α + x β α θ a α b - 1 - x β α θ α + x β α θ a α 2 b
Now integrating and applying binomial expansion on (5.2) for KGK distribution, we get 0 F ( x ) { 1 - F ( x ) } dx = 0 W i x β α θ α + x β α θ a α i dx - 0 W k x β α θ α + x β α θ a α k dx where W i = i = 0 ( - 1 ) i b i and W k = k = 0 ( - 1 ) k 2 b k .

Now simplifying further, we have

(5.3)
0 F ( x ) { 1 - F ( x ) } dx = W i 0 x β α θ α + x β α θ a α i dx - W k 0 x β α θ α + x β α θ a α k dx Using transformation, Z = x β α θ such that dx = β α θ Z 1 α θ - 1 dZ we may write,
(5.4)
= β α θ W i α a α i 0 Z a α i + 1 α θ - 1 1 - 1 - ( 1 + Z α ) - 1 a α i dZ - W k α a α k 0 Z a α k + 1 α θ - 1 1 - 1 - 1 + Z α - 1 a α k dZ
Let W = 1 - 1 + Z α - 1 then dZ = α ( 1 - W ) 2 dW .Then (5.4) becomes
(5.5)
0 = β α 1 α θ - 1 θ W i 0 1 W a α i + 1 α θ - 1 [ 1 - W ] - 1 α θ - 1 dW - W k 0 1 W a α k + 1 α θ - 1 [ 1 - W ] - 1 α θ - 1 dW
Using beta function, we get
(5.6)
0 F ( x ) { 1 - F ( x ) } dx = β α 1 α θ - 1 θ W i B a α i + 1 α θ , - 1 α θ - W k B a α k + 1 α θ , - 1 α θ
Substituting the results of (5.6) and (3.6) in (5.1), the Gini index for KGK distribution obtained as
(5.7)
G = W i B a α i + 1 α θ , - 1 α θ - W k B a α k + 1 α θ , - 1 α θ θ W j B 1 α θ + a α ( j + 1 ) , 1 - 1 α θ
where W i = i = 0 ( - 1 ) i b i , W k = k = 0 ( - 1 ) k 2 b k , W j = ab j = 0 ( - 1 ) j b - 1 j and E ( x ) = W j β α 1 α θ - 1 B 1 α θ + a α ( j + 1 ) , 1 - 1 α θ is the mean of KGK distribution.

5.2

5.2 Lorenz curve

Lorenz (1905) provided a curve, L ( p ) , which is defined as

(5.8)
L ( p ) = 1 μ 0 x xf ( x ) dx having a cdf, F ( x ) , with a finite mean μ .

The L ( p ) for KGK distribution is given as

(5.9)
L ( p ) = B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ

5.3

5.3 Bonferroni curve

The curve suggested by Bonferroni (1930) based on partial means for inequality measure can be determined through the relation

(5.10)
BC ( p ) = L ( p ) F ( x ) Using (5.9) and (2.3), we obtain BC ( p ) for KGK distribution as
(5.11)
BC ( p ) = 1 - 1 - x β α θ α + x β α θ a α b - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ

5.4

5.4 Zenga index

Zenga (1984, 1990) provided the following income inequality index

(5.12)
Z = 1 - μ ( x ) - μ ( x ) + where
(5.13)
μ ( x ) - = 1 F ( x ) 0 x xf ( x ) dx
and
(5.14)
μ ( x ) + = 1 1 - F ( x ) x xf ( x ) dx
which can also be written as
(5.15)
μ ( x ) + = 1 1 - F ( x ) μ - 0 x xf ( x ) dx
For KGK distribution, we obtain
(5.16)
μ ( x ) - = W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 1 - 1 - x β α θ α + x β α θ a α b - 1
and
(5.17)
μ ( x ) + = μ - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 1 - x β α θ α + x β α θ a α - b
Now, from (5.16) and (5.17), the Zenga index obtained for KGK distribution is
(5.18)
Z = 1 - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ μ - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 1 - x β α θ α + x β α θ a α - b - 1 - 1

5.5

5.5 Atkinson index

Atkinson (1970) presented a family of subjective indices, is defined as

(5.19)
A F ( θ , ε ) = 1 - 1 μ 0 x 1 - ε dF ( x ) 1 1 - ε which can also be written as
(5.20)
A F ( θ , ε ) = 1 - 1 μ { μ 1 - ε } 1 1 - ε
Using (3.5), we get
(5.21)
μ 1 - ε = W j β 1 - ε α 1 - ε α θ - 1 B 1 - ε α θ + a α ( j + 1 ) , 1 - 1 - ε α θ
Using (3.6) and (5.21), the Atkinson index obtained for KGK distribution is
(5.22)
A F ( θ , ε ) = 1 - W j β 1 - ε α 1 - ε α θ - 1 B 1 - ε α θ + a α ( j + 1 ) , 1 - 1 - α θ 1 1 - ε W j β α 1 α θ - 1 B 1 α θ + a α ( j + 1 ) , 1 - 1 α θ

5.6

5.6 Pietra index

Pietra (1915) offered an index, which is also known as Schutz index or half the relative mean deviation is defined as

(5.23)
P X = 1 2 μ 0 | x - μ | dF ( x ) = MD x ¯ 2 μ The Pietra index for KGK distribution is
(5.24)
P X = μ F ( μ ) - W j β α 1 α θ - 1 B W a α ( j + 1 ) + 1 α θ , 1 - 1 α θ W j β α 1 α θ - 1 B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ

5.7

5.7 Generalized entropy (GE)

Cowell (1980) and Shorrocks (1980) introduced the generalized entropy (GE) index and is defined as

(5.25)
GE F ( θ , δ ) = 1 δ ( δ - 1 ) 0 x μ δ - 1 f ( x ) dx ; δ 0 , 1 which can also be written as
(5.26)
GE F ( θ , δ ) = 1 δ ( δ - 1 ) μ δ μ δ - 1
where μ δ is the δ th moment about origin.

Using (3.5), we get

(5.27)
μ δ = W j β δ α δ α θ - 1 B a α ( j + 1 ) + δ α θ , 1 - δ α θ Now, inserting the results of (5.27) and (3.6) in (5.26), the GE obtained for KGK distribution is
(5.28)
GE F ( θ , δ ) = 1 δ ( δ - 1 ) W j β δ α δ α θ - 1 B a α ( j + 1 ) + δ α θ , 1 - δ α θ W j β α 1 α θ - 1 B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ δ - 1
If δ = 2 in (5.28), we obtain another inequality measures (squared coefficient of variation)
(5.29)
E 2 ( x ) = CV 2 2 = 1 2 W j β 2 α 2 α θ - 1 B a α ( j + 1 ) + 2 α θ , 1 - 2 α θ W j β α 1 α θ - 1 B a α ( j + 1 ) + 1 α θ , 1 - 1 α θ 2 - 1

6

6 Entropies

The most essential entropies of X KGK ( a , b , α , β , θ ) are Shannon entropy, Rényi entropy, β -entropy. Here we consider two entropy measures: the Rényi entropy, β -entropy. Entropies and kurtosis measures play the same role in comparing the shape of various densities and measuring heaviness of tails.

6.1

6.1 Rényi entropy

Rényi (1961) provided an extension of the Shannon entropy which is defined as:

(6.1)
I R ( γ ) = 1 1 - γ log f r ( x ) dx where γ > 0 and γ 1 .

Applying binomial expansion on KGK pdf given by (2.4), we get f γ ( x ) in more simple form

(6.2)
f γ ( x ) = ab α θ β γ j = 0 ( - 1 ) j γ ( b - 1 ) j x β a θ ( γ + j ) - γ α + x β α θ - a α + 1 γ - aj α The integral in (6.1) becomes
(6.3)
ab α θ β γ j = 0 ( - 1 ) j γ ( b - 1 ) j 0 x β a θ ( γ + j ) - γ α + x β α θ - a α + 1 γ - aj α dx
By using transformation W = x β α θ such that dx = β α θ W 1 α θ - 1 dW (6.3) reduces to
(6.4)
0 f γ ( x ) dx = ( ab ) γ θ β γ - 1 α 1 α θ - γ α θ - 1 j = 0 ( - 1 ) j γ ( b - 1 ) j 0 1 Z a α ( γ + j ) - γ α θ + 1 α θ - 1 [ 1 - Z ] γ + γ α θ - 1 α θ - 1 dz
Using beta function, (6.4) reduces to
(6.5)
0 f γ ( x ) dx = ( ab ) γ θ β γ - 1 α 1 α θ - γ α θ - 1 j = 0 ( - 1 ) j γ ( b - 1 ) j B a α ( γ + j ) - γ α θ + 1 α θ , γ + γ α θ - 1 α θ
Substituting the result of (6.5), then (6.1) becomes, the Rényi entropy for KGK distribution is given by
(6.6)
I R ( γ ) = 1 1 - γ log ( ab ) γ θ β γ - 1 α 1 - γ α θ - 1 j = 0 ( - 1 ) j γ ( b - 1 ) j B a α ( γ + j ) - γ α θ + 1 α θ , γ + γ α θ - 1 α θ

6.2

6.2 β -entropy

Havrda and Charvát (1967) presented β -entropy and later Tsallis (1988) applied it to physical problems. Furthermore, one-parameter generalization of the Shannon entropy is β -entropy which can lead to models or statistical results that are different from those acquired by using the Shannon entropy.

For a continuous random variable X having pdff ( x ) , the β -entropy is defined by

(6.7)
I β ( γ ) = 1 β - 1 [ 1 - f β ( x ) dx ] where β > 0 and β 1 .

Using the result of (6.5), we get

(6.8)
0 f β ( x ) dx = ( ab ) β θ β β - 1 α 1 α θ - β α θ - 1 j = 0 ( - 1 ) j β ( b - 1 ) j B a α ( β + j ) - β α θ + 1 α θ , β + β α θ - 1 α θ Hence, the expression for β -entropy is given as
(6.9)
I β ( γ ) = 1 β - 1 1 - ( ab ) β θ β β - 1 α 1 α θ - β α θ - 1 j = 0 ( - 1 ) j β ( b - 1 ) j B a α ( β + j ) - β α θ + 1 α θ , β + β α θ - 1 α θ

7

7 Estimation of parameters of KGK distribution

This section considers the estimation of the model parameters of the KGK distribution by using the method of maximum likelihood.

We assume that X follows the KGK distribution and let Θ = ( a,b, α , β , θ ) be the parameter vector of interest. The log-likelihood function = L ( Θ ) for a random sample x 1 , x 2 , , x n is given by

(7.1)
= nlog ( ab α θ β - θ a ) + ( θ a - 1 ) i = 1 n logx i - 1 + a α i = 1 n log α + x i β α θ + ( b - 1 ) i = 1 n log 1 - x i β α θ α + x i β α θ a α Let Z i = x i β α θ α + x i β α θ 1 α . Then, we can write as
(7.2)
= nlog ( ab α θ β - θ a ) + ( θ a - 1 ) i = 1 n logx i - 1 + a α i = 1 n log α + x i β α θ + ( b - 1 ) i = 1 n log [ 1 - Z i a ]
The components of the score vector U ( Θ ) are given by U a = n a ( 1 - a θ log β ) + θ i = 1 n logx i - 1 α i = 1 n log α + x i β α θ + ( 1 - b ) i = 1 n logZ i Z i - a - 1 U b = n b + i = 1 n log [ 1 - Z i a ] U α = n α + a α 2 i = 1 n log α + x i β α θ - 1 + a α i = 1 n α + x i β α θ - 1 - θ 1 + a α × i = 1 n Z i α log x i β + a ( 1 - b ) i = 1 n Z i a - 1 Z i α 1 - Z i a U β = - na θ β + ( a + α ) θ β i = 1 n Z i α + a ( 1 - b ) i = 1 n Z i a - 1 Z i β 1 - Z i a U θ = n θ ( 1 - a θ log β ) + a i = 1 n logx i - ( a + α ) i = 1 n Z i α log x i β + a ( 1 - b ) i = 1 n Z i a - 1 Z i θ 1 - Z i a To find the MLEs of the five parameters a , b , α , β and θ . Setting U a , U b , U α , U β and U θ to zero and solving them simultaneously. For interval estimation of the model parameters, we require the 5 × 5 observed information matrix J ( Θ ) = { U s , t } where s , t = a , b , α , β , θ given in Appendix A. To construct approximate confidence intervals for a , b , α , β and θ the multivariate normal N 5 ( 0 , J ( Θ ̂ ) - 1 ) distribution can be used under standard regularity conditions. Here, J ( Θ ̂ ) is the total observed information matrix evaluated at Θ ̂ . Then, 100 ( 1 - α ) % confidence interval for a , b , α , β and θ are given by a ̂ ± Z α / 2 σ a ̂ , b ̂ ± Z α / 2 σ b ̂ , α ̂ ± Z α / 2 σ α ̂ , β ̂ ± Z α / 2 σ β ̂ and θ ̂ ± Z α / 2 σ θ ̂ , respectively, where Z α / 2 is the quantile ( 1 - α / 2 ) of the standard normal distribution, and σ ( · ) ’s denote the diagonal elements of J ( Θ ̂ ) - 1 corresponding to a , b , α , β and θ .

To check KGK distribution is strictly superior to the Kappa distribution for a given data set the likelihood ratio (LR) statistic can be used. Then, the test of H 0 : a = b = 1 versus H 1 : H 0 is not true is equivalent to compare the KGK and Kappa distributions and the LR statistic becomes w = 2 { ( a ̂ , b ̂ , α ̂ , β ̂ , θ ̂ ) - ( 1 , 1 , α ̂ , β ̂ , θ ̂ ) } , where a ̂ , b ̂ , α ̂ , β ̂ and θ ̂ are the MLEs under H 1 and α ̂ , β ̂ and θ ̂ are the estimates under H 0 .

8

8 Empirical illustration

In this section, real data set is examined for illustration purpose. The MLEs are calculated and the measures of goodness of fit are used to compare the proposed model Kumaraswamy generalized Kappa distribution with the other competing models. We computed log-likelihood ( · ) , AIC (Akaike information criterion), BIC (Bayesian information criterion) and CAIC (consistent Akaike information criterion): AIC = - 2 ( · ) + 2 p , BIC = - 2 ( · ) + plogn , and CAIC = - 2 ( · ) + 2 pn n - p - 1 where ( · ) signifies the log-likelihood function examined at the maximum likelihood estimates, p is the number of parameters, and n is the sample size. We also used conventional goodness-of-fit tests in order to check which distribution fits better to this data set. We look at the Cramer-von Mises ( W ) and Anderson-Darling ( A ) statistics. In general, the smaller the values of these statistics, the better the fit to the data.

The following data set from Mielke and Johnson (1973) consists of Stream flow amounts (1000 acre-feet) for 35 year (1936–70) at the U.S. Geological Survey (USGS) gaging station number 9-3425 for April 1–August 31 of each year:

192.48, 303.91, 301.26, 135.87, 126.52, 474.25, 297.17, 196.47, 327.64, 261.34, 96.26, 160.52, 314.60, 346.30, 154.44, 111.16, 389.92, 157.93, 126.46, 128.58, 155.62, 400.93, 248.57, 91.27, 238.71, 140.76, 228.28, 104.75, 125.29, 366.22, 192.01, 149.74, 224.58, 242.19, 151.25.

Table 1 presents the estimated values of the MLEs of the parameters of the distributions and their standard errors. Estimated values for AIC, BIC, CAIC, A and W against all the fitted distributions are provided in Table 2. The histogram of the data set and the estimated pdfs and cdfs for the fitted models are displayed in Fig. 3. It is evident from the lower value of the statistics A and W for KGK distribution in comparison to the rest of the fitted distributions that it is the best fit distribution in the given circumstances. These findings are further supplemented by the estimated pdf and cdf plots given in Fig. 3 which clearly depicts that KGK distribution is a better fitted model as compared to others.

Table 1 MLEs and their standard errors (in parentheses) for Stream flow amounts (1000 acre-feet).
Distribution a b α β θ
KGK 37.6739
(563.6182)
4.8300
(9.7572)
0.1678
(1.0243)
17.3661
(219.7929)
7.2191
(44.7869)
EK (Lehmann type-I) 58.6622
(1716.8957)

0.2164
(1.6786)
32.0895
(373.3277)
11.7289
(91.2334)
EK (Lehmann type-II)
8.3742
(36.9606)
0.0629
(0.1273)
515.3789
(1623.1231)
16.8224
(26.8103)
Kappa3

0.0457
(0.2369)
161.5442
(15.9801)
56.7357
(287.0343)
Kappa2

10.7420
(5.9393)
312.1797
(46.5958)

Table 2 The ( · ) , AIC, BIC, CAIC, A , W values for Stream flow amounts (1000 acre-feet).
Distribution ( · ) AIC BIC CAIC A W
KGK 206.4671 422.9343 430.7110 425.0033 0.4555 0.0788
EK (Lehmann type-I) 207.0059 422.0117 428.2331 423.3450 0.4931 0.0794
EK (Lehmann type-II) 206.6339 421.2678 427.4892 422.6011 0.4867 0.0844
Kappa3 207.0037 420.0074 424.6735 420.7816 0.4919 0.0799
Kappa2 214.1449 432.2899 435.4006 432.6649 1.0876 0.1802
Plot of estimated pdf and cdf for stream flow amounts.
Fig. 3
Plot of estimated pdf and cdf for stream flow amounts.

9

9 Conclusion

In this paper, we propose a five-parameter Kumaraswamy generalized Kappa distribution. Algebraic expressions for various properties of the proposed distribution are provided. The method of maximum likelihood is employed to estimate the parameters along with an empirical illustration expressing the application by using stream flow amount data set. The proposed distribution is compared with some similar existing distributions by using Cramer-von Mises ( W ) and Anderson-Darling ( A ) statistics as measure of goodness of fit. It is concluded that KGK distribution is good competitive model for stream flow data set. The proposed distribution KGK has the potential to attract wider application in various areas such as hydrology, reliability and survival analysis.

Acknowledgment

The authors are grateful to the anonymous learned reviewers for their valuable suggestions which have greatly helped in improving the presentation of the manuscript.

References

  1. , , , , . Generalized beta-generated distributions. Comput. Stat. Data Anal.. 2012;56(6):1880-1897.
    [CrossRef] [Google Scholar]
  2. , , , . Log-gamma-generated families of distributions. Statistics. 2014;48(4):913-932.
    [CrossRef] [Google Scholar]
  3. , . On the measurement of th inequality. J. Econ. Theory. 1970;2(3):244-263.
    [CrossRef] [Google Scholar]
  4. , . Elementi Di Statistica Generale. Firenze: Libreria Seeber; .
  5. , , , . The Weibull-G family of probability distributions. J. Data Sci.. 2014;12(1):53-68. www.jds-online.com/files/JDS-1210.pdf·2013-10-21
    [Google Scholar]
  6. , , . A new family of generalized distributions. J. Stat. Comput. Simul.. 2011;81(7):883-898.
    [CrossRef] [Google Scholar]
  7. , , , . The exponentiated generalized class of distributions. J. Data Sci.. 2013;11(1):1-27.
    [Google Scholar]
  8. , . On the structure of additive inequality measures. Rev. Econ. Stud.. 1980;47(3):521-531. Oxford University Press
    [Google Scholar]
  9. , , , . Beta-normal distribution and its applications. Commun. Statist. Theory Methods. 2002;31(4):497-512.
    [CrossRef] [Google Scholar]
  10. Gini, C., 1914. Sulla Misura della concentrazione e della variabilit’a dei caratteri. Attidel regio Istituto Venteto di SS. LL.AA.,a.a. 1913-14, tomo LXXIII, parte II, pp. 1203–1248.
  11. , , , . Modeling failure time data by Lehmann alternatives. Commun. Statist. Theory Methods. 1998;27(4):887-904.
    [CrossRef] [Google Scholar]
  12. , , . Generalized exponential distribution. Austral. New Zealand J. Statist.. 1999;41(2):173-188.
    [CrossRef] [Google Scholar]
  13. , , . Quantification method in classification processes: concept of structural alpha-entropy. Kybernetika. 1967;3(1):30-35.
    [Google Scholar]
  14. Hussain, S., 2015. Properties, Extension and Application of Kappa Distribution. Unpublished M.Phil thesis submitted to Government College University, Faisalabad, Pakistan.
  15. , . Families of distributions arising from the distributions of order statistics. Test. 2004;13(1):1-43.
    [CrossRef] [Google Scholar]
  16. , . A generalized probability density function for double-bounded random processes. J. Hydrol.. 1980;46(1–2):79-88.
    [CrossRef] [Google Scholar]
  17. , . Methods of measuring the concentration of wealth. Publ. Am. Statist. Assoc.. 1905;9(70):209-219.
    [CrossRef] [Google Scholar]
  18. , , . A new method for adding a parameter to a family of distributions with applications to the exponential and weibull families. Biometrika. 1997;84(3):641-652.
    [CrossRef] [Google Scholar]
  19. , . Another family of distributions for describing and analyzing precipitation data. J. Appl. Meteorol.. 1973;12(2):275-280.
    [Google Scholar]
  20. , , . Three parameter Kappa distribution maximum likelihood estimations and likelihood ratio tests. Monthly Weather Rev.. 1973;101:701-707. ftp://ftp.library.noaa.gov/docs.lib/htdocs/rescue/mwr/101/mwr-101-09-0701.pdf
    [Google Scholar]
  21. Pietra, G., 1915. Delle relazioni tra indici di variabilità, note I e II. Atti del Reale Istituto Veneto di Scienze, Lett ere ed Arti, a.a. 1914–15, LXXIV, 775–804.
  22. , . On Measures of Entropy and Information. University of California Press 1961:547-561.
    [Google Scholar]
  23. , , . The gamma-exponentiated exponential distribution. J. Statist. Comput. Simul.. 2012;82(8):1191-1206.
    [CrossRef] [Google Scholar]
  24. , . The class of additively decomposable inequality measures. Econometrica. 1980;48(3):613-625.
    [Google Scholar]
  25. , . Possible generalization of Boltzmann-Gibbs statistics. J. Statist. Phys.. 1988;52(1–2):479-487.
    [CrossRef] [Google Scholar]
  26. Zenga, M., 1984. Proposta per un indice di concentrazione basto sui rapporti fraquantili di populizone e quantili di reddito. Giornale degli Economisti e Annali di Economia Nuova Serie, Anno 43, No. 5/6 (Maggio-Giugno 1984), pp. 301–326. http://www.jstor.org/stable/23246074.
  27. , . Concentration curves and concentration indexes derived from them. In: , , eds. Income and Wealth Distribution, Inequality and Poverty. Studies in Contemporary Economics. Berlin, Heidelberg: Springer; . p. :94-110.
    [Google Scholar]
  28. , , . On families of beta- and generalized gamma-generated distributions and associated inference. Statist. Method.. 2009;6(4):344-362.
    [CrossRef] [Google Scholar]

Appendix A

The elements of the 5 × 5 observed information matrix J ( Θ ) = { U s , t } where s , t = a , b , α , β , θ are given by U aa = - n a 2 + ( 1 - b ) i = 1 n { logZ i } 2 Z i a [ Z i - a - 1 ] 2 , U ab = - i = 1 n logZ i Z i - a - 1 U a α = 1 α 2 i = 1 n log α + x i β α θ - 1 α i = 1 n α + x i β α θ - 1 - θ α i = 1 n Z i α log x i β + a ( 1 - b ) × i = 1 n [ Z i - a - 1 ] - 2 Z i - a - 1 Z i α logZ i + ( 1 - b ) i = 1 n [ Z i - a - 1 ] - 1 Z i - 1 Z i α U a β = θ β i = 1 n Z i α - n + a ( 1 - b ) i = 1 n [ Z i - a - 1 ] - 2 Z i - a - 1 Z i β logZ i + ( 1 - b ) i = 1 n [ Z i - a - 1 ] - 1 Z i - 1 Z i β U a θ = - nlog β + i = 1 n logx i - i = 1 n Z i α log x i β + a ( 1 - b ) × i = 1 n [ Z i - a - 1 ] - 2 Z i - a - 1 Z i θ logZ i + ( 1 - b ) i = 1 n [ Z i - a - 1 ] - 1 Z i - 1 Z i θ U bb = - n b 2 , U b α = - a i = 1 n Z i a - 1 Z i α [ 1 - Z i a ] - 1 , U b β = - a i = 1 n Z i a - 1 Z i β [ 1 - Z i a ] - 1 U b θ = - a i = 1 n Z i a - 1 Z i θ [ 1 - Z i a ] - 1 U α α = - n α 2 - 2 a α 3 i = 1 n log α + x i β α θ + 2 a α 2 i = 1 n α + x i β α θ - 1 + a α 2 i = 1 n θ Z i α log x i β + 1 + a α i = 1 n α + x i β α θ - 2 + 1 + a α i = 1 n θ Z i 2 α x i β - α θ log x i β + a θ α 2 × i = 1 n Z i α log x i β - θ 1 + a α i = 1 n [ Z i α logZ i + α Z i α - 1 Z i α ] log x i β + a ( 1 - b ) i = 1 n ( 1 - Z i a ) { Z i a - 1 Z i α + ( a - 1 ) Z i a - 2 ( Z i α ) 2 } + aZ i 2 a - 2 ( Z i α ) 2 ( 1 - Z i a ) 2 U α β = - a θ α β i = 1 n Z i α - ( a + α ) θ β i = 1 n Z i 2 α x i β - α θ - ( a + θ ) i = 1 n α Z i α - 1 Z i β log x i β - Z i α β + a ( 1 - b ) i = 1 n ( 1 - Z i a ) Z i a - 1 Z i α β + ( a - 1 ) Z i a - 2 Z i α Z i β + aZ i 2 a - 2 Z i α Z i β ( 1 - Z i a ) 2 U α θ = ( a + α ) i = 1 n Z i 2 α x i β - α θ log x i β - i = 1 n Z i α log x i β - θ ( a + α ) u m i = 1 n Z i α - 1 Z i θ log x i β + a ( 1 - b ) i = 1 n ( 1 - Z i a ) { Z i a - 1 Z i α θ + ( a - 1 ) Z i a - 2 Z i α Z i θ } + aZ i 2 a - 2 Z i α Z i θ ( 1 - Z i a ) 2 U β β = na θ β 2 - ( a + α ) θ β 2 i = 1 n Z i α + α θ ( a + α ) β i = 1 n Z i α - 1 Z i β + a ( 1 - b ) i = 1 n ( 1 - Z i a ) { Z i a - 1 Z i β + ( a - 1 ) Z i a - 2 ( Z i β ) 2 } + aZ i 2 a - 2 ( Z i β ) 2 ( 1 - Z i a ) 2 U β θ = - na β + ( a + α ) β i = 1 n Z i α + α θ ( a + α ) β i = 1 n Z i α - 1 Z i θ + a ( 1 - b ) i = 1 n ( 1 - Z i a ) { Z i a - 1 Z i β θ + ( a - 1 ) Z i a - 2 Z i β Z i θ } + aZ i 2 a - 2 Z i β Z i θ ( 1 - Z i a ) 2 U θ θ = - n θ 2 ( 1 - a θ log β ) - na θ log β - α ( a + α ) i = 1 n Z i α - 1 Z i θ log x i β + a ( 1 - b ) i = 1 n ( 1 - Z i a ) { Z i a - 1 Z i θ + ( a - 1 ) Z i a - 2 ( Z i θ ) 2 } + aZ i 2 a - 2 ( Z i θ ) 2 ( 1 - Z i a ) 2 where Z i α = 1 α 2 Z i log α + x i β α θ - 1 α Z i α + 1 x i β - α θ - θ α Z i α + 1 log x i β Z i β = θ β { Z i α + 1 - Z i } Z i θ = log x i β [ Z i - Z i α + 1 ] Z i α = x i β - α θ + θ log x i β 2 α 2 Z i α + 1 - 1 + 1 α Z i α Z i α - 1 α Z i α + 1 logZ i + Z i α α 2 - 2 α 3 Z i log α + x i β α θ + θ α Z i α + 1 x i β - α θ log x i β Z i α β = 1 α 2 Z i β log α + x i β α θ - θ β Z i α + 1 x i β - α θ - 1 + 1 α Z i α Z i β × x i β - α θ + θ log x i β Z i α θ = 1 α 2 Z i θ log α + x i β α θ + θ α Z i α + 1 x i β - α θ log x i β - 1 + 1 α Z i α Z i θ x i β - α θ + θ log x i β Z i β = θ β 2 Z i [ 1 - Z i α ] + θ β Z i β [ ( α + 1 ) Z i α - 1 ] Z i β θ = 1 β [ Z i α + 1 + θ ( α + 1 ) Z i α Z i θ + θ Z i θ + Z i ] Z i θ = log x i β Z i θ [ 1 - ( α + 1 ) Z i α ]

Show Sections