1. Introduction
Shannon [Reference Shannon29] introduced the concept of entropy, which is widely used in the fields of probability, statistics, information theory, communication theory, economics, physics, and so forth. In information theory, entropy is a measure of the uncertainty associated with a random variable. For more details, see [Reference Cover and Thomas6].
Suppose that X and Y are two nonnegative random variables representing time to failure of two systems with probability density functions (p.d.f.s) f(x) and g(x), respectively. Suppose that F(x) and G(Y) are failure distributions or cumulative density functions (c.d.f.s) and that $\bar{F}(x) $ and $\bar{G}(x) $ are survival functions (SFs) of X and Y, respectively. Shannon’s measure of entropy in [Reference Shannon29] and Kerridge’s measure of inaccuracy in [Reference Kerridge12] associated with the random variable X are given by
and
respectively, where “log” means the natural logarithm. In the case where $g(x) = f(x)$, then Eq. (2) reduces to Eq. (1).
The measure of information and inaccuracy are associated as $H(\,f, g) =H(\,f ) + H(\,f|g)$, where $H(\,f|g)$ represents the relative information measure of X about Y (see [Reference Kullback13]), defined as
1.1. Extropy
Recently, Lad et al. [Reference Lad, Sanfilippo and Agro15] proposed an alternative measure of uncertainty of a random variable called extropy. The term extropy, as an antonym to entropy, had been used earlier in academic literature as well. Extropy was defined by Max More, and it is “the extent of a living or organizational system’s intelligence, functional order, vitality, energy, life, experience, and capacity and drive for improvement and growth”. Extropy is not a rigorous technical term defined in the philosophy of science. In a metaphorical sense, it is simply expressed as the opposite of entropy. The extropy of the random variable X is defined by [Reference Lad, Sanfilippo and Agro15] as follows:
The properties of this measure such as the maximum extropy distribution and statistical applications were presented by [Reference Lad, Sanfilippo and Agro15].
The characterization results and symmetric properties of the extropy of order statistics and record values were derived in [Reference Qiu22]. Two estimators for the extropy of an absolutely continuous random variable were introduced by [Reference Qiu and Jia21]. Qiu et al. [Reference Qiu, Wang and Wang23] explored an expression of the extropy of a mixed system’s lifetime. Yang et al. [Reference Yang, Xia and Hu32] studied the relations between extropy and variational distance and determined the distribution that attains the minimum or maximum extropy among these distributions within a given variation distance from any given probability distribution. Also, for more details, see [Reference Pakdaman and Haashempour19, Reference Pakdaman and Hashempour20] and references therein. Jahanshahi et al. [Reference Jahanshahi, Zarei and Khammar11] introduced an alternative measure of uncertainty of nonnegative random variable X, which they called the cumulative residual extropy (CRJ), given as
They studied some properties of the aforementioned information measure. The measure (3) is not applicable to a system having survived for some units of time. Hence, Sathar and Nair [Reference Sathar and Nair27] proposed a dynamic version of CRJ (called dynamic survival extropy) to measure the residual uncertainty of lifetime random variable X as follows:
Analogous to CRJ, Nair and Sathar [Reference Nair and Sathar17] introduced a cumulative extropy useful in measuring the inactivity of a system. It is a dual concept of the CRJ and is suitable to measure the uncertainty on past lifetimes of the system. It is defined as
In analogy with Eq. (4), [Reference Nair and Sathar17] proposed a dynamic cumulative extropy for the past lifetime, defined by
Recently, [Reference Mohammadi and Hashempour16] proposed interval weighted cumulative residual extropy (WCRJ) and weighted cumulative past extropy (WCPJ), respectively, as follows:
and
They proposed some properties and nonparametric estimators of these measures.
1.2. Record values
Record values can be viewed as order statistics from a sample whose size is determined by the values and the order of occurrence of the observations. Suppose that $X_1, X_2, \ldots $ are a sequence of independent and identically distributed (i.i.d.) random variables with continuous c.d.f. F(x) and p.d.f. f(x). The order statistics of a sample is defined by the arrangement of $X_1, \ldots, X_n$ from the smallest to the largest, denoted as $X_{1:n}, \ldots, X_{n:n}$. An observation ${X_k}$ will be called an upper record value if its value exceeds that of all previous observations. Thus, X k is an upper record if $X_k \gt X_i$ for every i < k. An analogous definition can be given for a lower record value. Suppose that T(k) is the time at which the kth record value is observed. Since the first observation is always a record value, we have $T(1) = 1, \ldots, T( k +1) =\min \{i : X_i \gt X_{T( k)} \}$, where T(0) is defined to be 0. The sequence of upper record values can thus be defined by $ U_k= X_{T( k)},\ k= 1, 2, \ldots$. Let $S( k) = T( k+ 1)-T( k)$ denote the inter-record time between the kth record value and the $(k+1)$th record value, and let the kth record value $X_{T(k)}$ be denoted by X k for simplicity. It can be easily shown that the p.d.f. of the kth upper record values X k and kth lower record values X k are given as
and
where $(k-1)!=\Gamma(k)$ denotes the gamma function defined as $\Gamma(k)=\int^\infty_0 x^{k-1}\,{\rm e}^{-x}\,{\rm d}x.$
We refer the reader to [Reference Arnold, Balakrishnan and Nagaraja2] and [Reference Ahsanullah1] and references therein for more details on the theory and applications of record values. The information measures for record values have been investigated by several authors, including [Reference Baratpour, Ahmadi and Arghami4, Reference Ebrahimi, Soofi and Zahedi7, Reference Kumar14, Reference Raqab and Awad24].
A question is how we can determine the amount of inaccuracy contained in a set of record values from a sequence of i.i.d. random variables. Based on this idea, in this relevance, we suggest a measure of inaccuracy between a parent random variable and the distributions of nth record value based on extropy. This paper is organized as follows: In Section 2, we determine the amount of inaccuracy contained in a set of record values from a sequence of i.i.d. random variables and study a characterization result based on this measure. In Sections 3 and 4, the measure of inaccuracy for some distributions has been studied. In Section 5, nonparametric estimators for the proposed measures are also obtained. Further, in Section 6, we use a Monte Carlo simulation study to evaluate our estimators. Finally, Section 7 investigates the behavior of the proposed estimators for a real dataset.
2. A new measure of inaccuracy for record statistics
In this section, we introduce and study a new measure of uncertainty between the distributions of nth record value and parent random variable based on extropy.
Definition 2.1. Let X and Y be nonnegative continuous random variables with p.d.f.s f(x) and g(x) and c.d.f.s F(x) and G(x), respectively. Then the “extropy-inaccuracy” measure between the distributions X and Y is defined as
In what follows, let f(x) and g(x) be density functions of the lifetime random variables X and Y, respectively. It is assumed that f(x) is the actual density function corresponding to the observation and g(x) is the density function assigned by the experimenter. Also, the discrimination information based on extropy and inaccuracy between density functions f(x) and g(x) can be defined by
for more details, see [Reference Raqab and Qiu25].
The extropy-inaccuracy measure can be viewed as a generalization of the idea of extropy. This measure is a useful tool for the measurement of errors in experimental results. In fact, the extropy inaccuracy measure can be expressed as the sum of an uncertainty measure and discrimination measure between two distribution. When an experimenter states the probability of various events in an experiment, the statement can lack precision in two ways: one resulting from incorrect information (e.g., mis-specifying the model) and the other from vagueness in the statement (e.g., missing observation or insufficient data). All estimation and inference problems are concerned with making statements, which may be inaccurate in either or both of these ways. The extropy-inaccuracy measure can take account for these two types of errors.
This measure has its application in statistical inference and estimation. Also, some concepts in reliability studies for modeling lifetime data such as failure rate and mean residual life function can be described using extropy-inaccuracy measure. In lifetime studies, the data are generally truncated. Hence, there is scope for extending information theoretic concepts to ordered situations and record values. Motivated by this, we extend the definition of inaccuracy to the extropy-inaccuracy measure. Further, we also look into the problem of characterization of probability distributions using the functional form of these measures. Also, the identification of an appropriate probability distribution for lifetimes is one of the basic problems encountered in reliability theory. Although several methods such as probability plots, the goodness of fit procedures, etc., are available in the literature to find an appropriate model followed by the observations, they fail to provide an exact model. A method to attain this goal can be to utilize an extropy-inaccuracy measure.
According to [Reference Nath18], we provide the following basic properties of extropy-inaccuracy measure in Eq. (5):
1. In case $f(x) = g(x)$, then $ J(\,f, g) $ reduces to extropy and achieves its minimum value.
2. The extropy inaccuracy measure is symmetric, that is, $ J(\,f, g)= J(g,f)$.
3. Finiteness of extropy inaccuracy measure: A necessary condition for $ J(\,f, g) $ to be finite is that F and G should be absolutely continuous with respect to each other, that is, $ F\equiv G $. It is obvious that $ J(\,f, g)\leq 0 $ .
4. $J(\,f|h) -J(\,f|g) =J(\,f, h) -J(\,f, g)$.
5. Convergence of extropy inaccuracy measure: Let there be a sequence of p.d.f.s $\{g_n(x)\}_{n=1}^{\infty}$ such that $ \lim_{n\rightarrow \infty} g_n(x)= f(x) $ almost everywhere. If $ \vert g_n(x) \vert \leq M $, for all $ n \in N $, where N is the set of positive integers, then
$$ \vert g_n(x) f(x) \vert \leq M N^{\prime}, $$where $ f(x)\leq N^{\prime} $ and $ \lim_{n\rightarrow \infty} g_n(x){\ }f(x)= f^2(x) $ almost everywhere, so that by Lebesgue convergence theorem,$$ \lim_{n\rightarrow \infty} J(\,f,g_n)= J(\,f). $$
The inaccuracy and extropy-inaccuracy measures are not competing but are rather complementary. However, the properties of symmetry, finiteness, and simplicity in calculations can be considered as the advantages of extropy-inaccuracy measure over Kerridge’s inaccuracy measure. The most important advantage of extropy is that it is easy to compute, and it will therefore be of great interest to explore its important potential applications in developing goodness-of-fit tests and inferential methods.
In Example 2.2, the behaviors of the extropy-inaccuracy measure and Kerridge’s inaccuracy measure are compared for the exponential distribution.
Example 2.2. We consider the exponential distribution with mean values of 0.5 for f(x) and λ −1 for g(x). The Kerridge’s inaccuracy measure $H(\,f,g)$ and extropy-inaccuracy measure $J(\,f,g)$ can be obtained as
Plots of these two measures are shown in Figure 1. It is observed that the Kerridge’s inaccuracy measure $H(\,f,g)$ does not have a monotone behavior with respect to the parameter lambda, while our proposed inaccuracy measure $J(\,f,g)$ has a monotone behavior with respect to the parameter lambda. In Figure 1, it can be seen that with the increase in the parameter lambda, the value of $J(\,f,g)$ decreases uniformly because the distance between f(x) and g(x) density functions increases. Also, when lambda tends to zero, the value of $H(\,f,g)$ will be infinite, while that of $J(\,f,g)$ is finite and nonpositive.
In the following result, we obtain the measure of inaccuracy between the distributions of nth record value and parent random variable.
Proposition 2.3. Let X be a nonnegative continuous random variable with p.d.f. f(x) and c.d.f. F(x). Then
Proof. We have
where $J(\,f_n)$ and $J(\,f_n^u |\, f )$ denote the measure of extropy and the measure of relative information based on extropy, respectively. Also, the measure of information and inaccuracy is associated as $J(\,f )= J(\,f_n^u, f)-J(\,f_n^u |\, f)$.
In rest of the section, we show that the inaccuracy measure defined above characterizes the distribution function of parent random variable X uniquely. To prove this, we use the following lemma.
Lemma 2.4 A complete orthogonal system for the space L 2 is given by the sequence of Laguerre functions
where $L_n(x)$ is the Laguerre polynomial defined as the sum of coefficients of ${\rm e}^{-x}$ in the nth derivative of $x^n \,{\rm e}^{-x}$, that is,
The completeness of Laguerre functions in L 2 means that if $f \in L_2$ and $\int^\infty_0 f(x)L_n(x) {\rm e}^{-x/2}=0,$ then f is zero almost everywhere.
Theorem 2.5. Suppose that X 1 and X 2 are two nonnegative random variables with p.d.f.s $f_1(x)$ and $f_2(x)$ and continuous c.d.f.s $F_1(x)$ and $F_2(x)$, respectively. Then F 1 and F 2 belong to the same family of distributions but for change in location if and only if
where $f_{n,1}$ and $f_{n,2}$ are the p.d.f.’s of nth record value for the parent distributions f 1 and f 2, respectively.
Proof. We prove the sufficiency part. For all $n \geq 1$, suppose $ J(\,f^u_{n,1}(x), f_1(x))= J(\,f^u_{n,2}(x), f_2(x))$. Then
Moreover,
Using substitutions $u_1=-\log(\bar{F_1}(x))$ and $u_2=-\log(\bar{F_1}(x))$, we obtain
Thus, we have
It follows from Lemma 2.4 that
where $L_n(u)$ is the Laguerre polynomial given in Lemma 2.4. Using Lemma 2.4, we have
Substituting $w=1-{\rm e}^{-u}$ in the above expression, we get
It is easy to show that $(F^{-1})^{\prime} (w)={\rm d}(F^{-1}(w))/{\rm d}w=1/f((F^{-1}(w)) $. Thus, we have $ (F^{-1}_1)^{\prime} (w)=(F^{-1}_2)^{\prime} (w) \ \text{for all } w \in (0, 1).$ Therefore, $F^{-1}_1 (w)=F^{-1}_2(w)+b,$ where b is a constant. The necessary part is obvious.
In what follows, we prove another result to show the effect of monotone transformations on the inaccuracy measure defined in Eq. (7). In this context, we prove the following theorem.
Theorem 2.6. Assume that X is a nonnegative and continuous random variable with p.d.f. f(x) and c.d.f. F(x). Suppose $Y=\psi(X)$, where ψ is a strictly monotonically increasing and differentiable function with derivative $\psi^{\prime}(x)$ and that G(y) and g(y) denote the distribution and density functions of Y, respectively. Assume that X n denotes the nth record value associated with X with p.d.f. f n and that Y n denotes the nth record value associated with Y with p.d.f. g n. Then
Proof. The p.d.f. of $Y=\psi(X)$ is $g(y)=f\left(\psi^{-1}\left(y\right)\right)/ \psi^{\prime}\left(\psi^{-1}\left(y\right)\right).$ Therefore,
This gives
By changing $x=\psi^{-1}(y)$, we get
The proof is completed.
Corollary 2.7. Let X be a nonnegative absolutely continuous random variable and let us define $Y = bX +c$, where b and c are constants with b > 0. Then
Proof. The p.d.f. of $Y=bX+c$ is $g(y)=b f(\frac{y-c}{b}).$ Therefore,
By changing $x=\frac{y-c}{b}$, we get
The proof is completed.
Thus, the inaccuracy measure defined in Eq. (7) is invariant under scale but not under location transformation.
3. Measure of inaccuracy for some distributions
In this section, for some continuous distributions for a random variable X, the expressions for the proposed measure read as follows:
1. Exponential distribution
Suppose that a random variable X is exponentially distributed over x > 0 with mean $1/\lambda \gt 0$. Then, $J(\,f^u_n, f)=-\frac{\lambda}{2^{n+1}}$. We observe that for a fixed value of n, the inaccuracy of the nth record value for the exponential distribution decreases with increase in the value of the parameter λ > 0. Similarly, if λ is fixed, then $J(\,f^u_n, f)$ increases with an increase in the sample size. In this case, we obtain $ ~\Psi_n=J(\,f^u_{n+1}, f)-J(\,f^u_n, f)=\frac{\lambda}{2^{n+2}},\ \text{for all } n $, that is, the nth inaccuracy differential depends on the sample size.
In Figure 2, the functions $J(\,f^u_n, f)$ and $\Psi_n$ are, respectively, shown in the left panel and the right panel for the exponential distribution with respect to the parameter n for some selected values of $\lambda=1, 5, 10$. It is observed that $J(\,f^u_n, f)$ is nondecreasing and $\Psi_n$ is nonincreasing with respect to n for fixed values of λ. Note that $J(\,f^u_n, f)$ is nonincreasing and that $\Psi_n$ is nondecreasing in λ.
2. Uniform distribution
Let a nonnegative random variable X be uniformly distributed over (a, b), a < b, with density and distribution functions, respectively, given by
$$ f (x) =\frac{1}{b-a} $$and$$ \quad F(x) =\frac{x-a}{b-a}, \quad~a\lt x \lt b. $$Then
\begin{eqnarray*} J\left(\,f^u_n, f\right)&=&-\frac{1}{2\Gamma(n)(b-a)^2}\int^b_a \left[ -\log\left(\frac{x-a}{b-a}\right) \right]^{n-1}\,{\rm d}x \\ &=& -\frac{1}{2(b-a)}. \end{eqnarray*}Suppose that $\Psi_n=J(\,f^u_{n+1}, f)-J(\,f^u_n, f) $ is the nth inaccuracy differential, that is, the change in inaccuracy in observing the record value from the nth to the $(n + 1)$th. In the case of uniform distribution, $\Psi_n=J(\,f^u_{n+1}, f)-J(\,f^u_n, f)=0$, that is, for the uniform distribution, the measure of inaccuracy $J(\,f^u_n, f)$ remains constant for all n.
3. Weibull distribution
Assume that a nonnegative random variable X has a Weibull distribution over x > 0 with p.d.f. given by
$$f (x) =\lambda \delta x^{\delta-1} \, \exp\{-\lambda x^\delta \}, \quad \lambda,\delta \gt 0,$$where λ and δ are scale and shape parameters, respectively. Then the SF is$$ \quad \bar{F}(x) =\exp \{-\lambda x^\delta \}. $$We have
(8) \begin{eqnarray} J(\,f^u_n, f)=- \frac{\delta \lambda^{-\delta}}{2^{n+2-\delta^{-1}}}\frac{\Gamma(n+1-\delta^{-1})}{\Gamma(n)}. \end{eqnarray}In particular, for δ = 1, Eq. (8) reduces to the nth record inaccuracy for the exponential distribution. The nth inaccuracy differential is given as follows:
\begin{eqnarray*} \Psi_n=\frac{\Gamma(n+1-\delta^{-1})}{\Gamma(n+1)} \frac{\delta\lambda^{-\delta}(n+\delta^{-1}-1)}{2^{n+3-\delta^{-1}}}. \end{eqnarray*}In Figure 3, the functions $J(\,f^u_n, f)$ and $\Psi_n$ are, respectively, shown in the left panel and right panel for the Weibull distribution with respect to the parameter n for some selected values of $\lambda=1, 3$ and $\delta=1, 3$. It is observed that $J(\,f^u_n, f)$ is nondecreasing and that $v\Psi_n$ is nonincreasing with respect to n for fixed values of λ and δ. Note that $J(\,f^u_n, f)$ is nonincreasing and that $\Psi_n$ is nondecreasing in δ. Also, $J(\,f^u_n, f)$ is nondecreasing and $\Psi_n$ is nonincreasing in λ.
4. Measure of inaccuracy for F β distributions
In this section, we introduce the inaccuracy measure for nth lower record value for some F β distributions. The F β distributions have been widely studied in statistics because of their wide applicability in the modeling and analysis of lifetime data. This model is flexible enough to accommodate both monotonic and nonmonotonic failure rates even though the baseline failure rate is monotonic.
Let X and Y be nonnegative continuous random variables with p.d.f.s f(x) and g(x) and c.d.f.s F(x) and G(x), respectively. Also, they satisfy the proportional reversed hazard rate model with proportionality constant β > 0, if
where F(x) is the baseline distribution and G(x) can be considered as some reference distributions (see [Reference Gupta and Nanda9]).
Remark 4.1. Let $X_1, X_2, \ldots, X_m$ be i.i.d. random variables with c.d.f. F(x), representing the lifetime of components in an m-component parallel system. Then the lifetime of the parallel system is given by $Z= \max \{X_1, X_2, \ldots, X_m \}$ with c.d.f. G(x) given by Eq. (9).
1. Power function (PF) distribution
A random variable X is said to have PF distribution if its c.d.f. and p.d.f. are given by
$$ G(x)=\left( \frac{x}{\theta} \right)^\beta,\quad~ x \in (0, \theta), ~\theta \gt 0, ~\beta \gt 0, $$and$$ g(x)=\frac{\beta}{\theta} \left( \frac{x}{\theta} \right)^{\beta-1}. $$We get
$$ J(g^l_n, g)=- \frac{\beta}{2\theta (2-\frac{1}{\beta})^n}, $$and$$ J(g^l_{n+1}, g)=- \frac{\beta}{2\theta (2-\frac{1}{\beta})^{n+1}}. $$Moreover,
$$\Psi_n = J(g^l_{n+1}, g) - J(g^l_n, g)=\frac{\beta-1}{2\theta}\left( \frac{\beta}{2\beta-1} \right)^{n+1}, $$which means that the difference between inaccuracy measures of two consecutive record values from the PF distribution depends on n. Putting β = 1, it is reduced to the uniform distribution, and also $\Psi_n =0$.In Figure 4, the functions $J(g^l_n, g)$ and $\Psi_n$ are, respectively, shown in the left panel and right panel, for the PF distribution with respect to the parameter n for some selected values of $\theta=1, 4$ and $\beta=2, 4$. It is observed that $J(g^l_n, g)$ is nondecreasing and that $\Psi_n$ is nonincreasing with respect to n for fixed values θ and β. Note that $J(g^l_n, g)$ is nondecreasing and that $\Psi_n$ is nonincreasing in θ. Also, $J(g^l_n, g)$ and $\Psi_n$ are not monotone in β.
2. Gompertz–Verhulst Exponentiated (GVE) distribution
GVE distribution is defined by the c.d.f. and p.d.f. given as follows:
$$ G(x)=\left( 1-\rho \exp\{-\lambda x\}\right)^\beta$$and
$$ g(x)=\beta\left( 1-\rho\, \exp\{-\lambda x\}\right)^{\beta-1}\rho \lambda\, \exp\{-\lambda x\}.$$Then,
$$ J(g^l_n, g)=-\frac{\lambda \beta}{2} \left[\frac{1}{(2-\frac{1}{\beta})^n}-{\frac{1}{2^n}} \right]$$and$$ J(g^l_{n+1}, g)=-\frac{\lambda \beta}{2} \left[\frac{1}{(2-\frac{1}{\beta})^{n+1}}-{\frac{1}{2^{n+1}}} \right].$$Thus, we have
$$\Psi_n = J(g^l_{n+1}, g) - J(g^l_n, g)= \frac{\lambda \beta}{2} \left[ \frac{\beta-1}{(2\beta-1)(2-\frac{1}{\beta})^{n}}-\frac{1}{2^{n+1}} \right], $$which means that the difference between inaccuracy measures of two consecutive record values from GVE distribution depends on n.In Figure 5, the functions $J(g^l_n, g)$ and $\Psi_n$ are, respectively, shown in the left panel and right panel for the GVE distribution with respect to the parameter n for some selected values of $\lambda=2, 4$ and $\beta=2, 4$. It is observed that $J(g^l_n, g)$ and $\Psi_n$ are not monotone with respect to n for fixed values λ and β. Note that $J(g^l_n, g)$ is nonincreasing and that $\Psi_n$ is nondecreasing in λ. Also, $J(g^l_n, g)$ is nondecreasing and $\Psi_n$ is not monotone in β.
The exponentiated exponential (EE) distribution introduced by [Reference Gupta and Kundu8] has some interesting physical interpretations. A random variable X is said to have the EE distribution if its c.d.f. and p.d.f. are given by
\begin{equation*} G(x)= \left[ 1-\exp \left(-\lambda x\right) \right]^\beta,\quad x \gt 0,\ \beta \gt 0,\ \lambda \gt 0. \end{equation*}In particular, the exponential distribution is obtained for β = 1. The EE distribution is the special case of the GVE distribution. When ρ = 1, we get $J(g^l_n, g)$ as same as that of GVE distribution.
5. Nonparametric estimators
In this section, we provide two methods for nonparametric estimations of the inaccuracy of nth upper (lower) record values. Recently, similar methods have been used for inaccuracy estimation in [Reference Hashempour and Mohammadi10] and [Reference Viswakala and Sathar31]. Let $X_{1},\dots,X_{n}$ be a random sample taken from a population with density function $ f(\cdot)$ and c.d.f. $ F(\cdot) $. Then a nonparametric estimator of the measure of inaccuracy between the distributions of nth upper record value and parent random variable can be achieved by
where $ \widehat{F}(\cdot) $ and $ \widehat{f}(\cdot) $ denote the estimators of $ F(\cdot) $ and $ f(\cdot)$, respectively.
We use the empirical and kernel methods to estimate the c.d.f. F(x). In the empirical method, the c.d.f. F(x) is estimated by $ \widehat{F}_n(x)=\frac{1}{n} \sum_{i=1}^{n} I(X_i \leq x) $. We design another estimator based on the kernel-smoothed estimator of the c.d.f. since smoothed estimators perform better than nonsmoothed estimators. The kernel c.d.f. estimator is described as follows:
where $W(\cdot)$ is the c.d.f. of a positive kernel function $K(\cdot)$ as $W(x) =\int_{-\infty}^{x} K(t)\,{\rm d}t$ and h n is a bandwidth parameter.
The $ K (\cdot) $ kernel must be selected carefully; moreover, the bandwidth or smoothing parameter h n must be selected more carefully. To select the bandwidth parameter, we use two approaches. First, we find the optimal h n by minimizing the integrated mean squared error by considering the Gaussian kernel, which gives us
Usually, σ is estimated by $\min\{S, Q/1.34\}$, where S is the sample standard deviation and Q is the interquartile range. This choice of $h_n^{1}$ works well if the true density is very smooth. This approach was called the normal reference rule or rule-of-thumb in [Reference Silverman30].
In the second approach, the bandwidth is chosen by the cross-validation technique. Since we do not want to assume that the density of data is very smooth necessarily, it is usually better to estimate h n by using cross-validation. We use the leave-one-out cross-validation criterion defined by [Reference Sarda26] method, which considers the selection method as
where $ F_{h,-i}(X_i) $ is the leave-one-out version of the kernel-smoothed estimator of c.d.f., defined as
The support set of random variables that is considered in Sections 3 and 4 is bounded. Near the boundaries, the kernel density estimates have larger errors. Kernel density estimates tend to smooth the probability mass over the boundary points. To solve this boundary problem, we use a reflection boundary technique, which was illustrated in [Reference Scott28]. In this technique, to estimate the density $ f(\cdot)$ by using sample $x_{1},\dots,x_{n}$, first add the reflection of the entire sample; that is, append $-x_{1},\dots,-x_{n}$ to the data. Then estimate a density $ s(\cdot)$ using the 2 n points, but use n points to determine the bandwidth parameter. Thus, $\widehat{f}(x) = 2 \widehat{s}(x)$. This method is applied to the kernel density estimation, and the bandwidth is selected using the rule-of-thumb and cross-validation techniques.
Similarly, a nonparametric estimator of nth lower record value and parent random variable is defined by
where $ \widehat{G}(\cdot) $ and $ \widehat{g}(\cdot) $ denote the estimators of $ G(\cdot) $ and $ g(\cdot)$, respectively. These estimators can be computed by similar methods as discussed above for estimates $ \widehat{F}(\cdot) $ and $ \widehat{f}(\cdot) $.
6. Simulation
In this section, we utilize a Monte Carlo simulation study to assess our estimators in terms of Bias and root mean square error (RMSE). For the simulations, we consider the Weibull and GVE distributions with various parameters. The proposed estimators are computed for various values of sample sizes (20 and 50) and record values n (2, 3, and 5).
The bias and RMSE of four estimators of nth upper record value for the Weibull distribution and nth lower record value for the GVE distribution are computed and given in Tables 1 and 2, respectively. Simulation results show the density estimation with the reflection boundary technique, and as a result, the c.d.f. estimation with the kernel method by using the rule-of-thumb bandwidth selection, $ (\widehat{F}_{h_1},\widehat{f}_{h_1}) $, has the best performance between the proposed estimators under different parameters and record values. When the c.d.f. is estimated with the empirical method ( $ \widehat{F}_{n} $), estimators based on kernel density estimation with the rule-of-thumb technique ( $ \,\widehat{f}_{h_1} $) perform better than the cross-validation technique ( $ \,\widehat{f}_{h_2} $) for the bandwidth selection. The accuracy of the estimators using the cross-validation technique improves with increasing the sample size. The Bias and RMSE estimation of the proposed estimators are reduced by increasing the sample sizes or record values. These results can be seen in both Tables 1 and 2.
7. Applications
Here, we consider the real data set to show the behavior of the estimators in real cases. This data set includes an active repair time (in hours) for an airborne communication transceiver reported by [Reference Balakrishnan, Leiva, Sanhueza and Cabrera3], which was initially given by [Reference Chhikara and Folks5]. The actual observations are listed below:
0.2, 0.3, 0.5, 0.5, 0.5, 0.5, 0.6, 0.6, 0.7, 0.7, 0.7, 0.8, 0.8, 1.0, 1.0, 1.0, 1.0, 1.1, 1.3, 1.5, 1.5, 1.5, 1.5, 2.0, 2.0, 2.2, 2.5, 2.7, 3.0, 3.0, 3.3, 3.3, 4.0, 4.0, 4.5, 4.7, 5.0, 5.4, 5.4, 7.0, 7.5, 8.8, 9.0, 10.3, 22, 24.5.
We have fitted the data set using the weibull distribution by the maximum likelihood method for parameters estimation. The obtained parameters are $\widehat{\lambda}=3.391$ and $\widehat{\delta}=0.899$, and we have a Kolmogorov–Smirnov (K-S) statistic value 0.120 with p-value 0.517.
Figure 6 shows theoretical values and nonparametric estimations of $ J(\,f^u_n, f) $ for the Weibull distribution. We can see that nonparametric estimations of $ J(\,f^u_n, f) $ are based on the density estimation with the reflection boundary technique. Moreover, the c.d.f. estimation with the kernel method by using the rule-of-thumb bandwidth selection, $ (\widehat{F}_{h_1},\widehat{f}_{h_1}) $, has less error than other estimates and is closer to the theoretical value. It can also be seen in Figure 6 that as the record value increases, all estimates approach the theoretical value.
8. Conclusion and future remarks
In this paper, we have extended the Kerridge measure of inaccuracy for record values, and a characterization result has been studied. Some properties of the proposed measure have been discussed. It was also shown that the defined measure of inaccuracy is invariant under scale but not under location transformation. We have characterized certain specific lifetime distribution functions. The measure has been studied for some F β distributions also. Nonparametric estimators of inaccuracy based on extropy for nth upper and lower record values have been proposed. Simulation results showed that the density estimation with the reflection boundary technique and the c.d.f. estimation with the kernel method by using the rule-of-thumb bandwidth selection have the best performance between proposed estimators. The performance of estimators based on a real data set has been discussed as well.
Acknowledgments
The authors would like to thank referees and the editor for their useful comments and constructive criticisms on the original version of this manuscript, which led to this considerably improved version.