Smoothed Empirical Likelihood Inference for Nonlinear Quantile Regression Models with Missing Response ()
1. Introduction
Quantile regression (QR) proposed by Koenker and Bassett [1] has become a popular alternative to least squares method for providing comprehensive description of the response distribution and robustness against heavy-tailed error distributions. Because of these significant advantages, QR has become an effective method for statistical research. There are many literatures on the estimation of quantile regression models; among them, Koenker [2] is a monograph worth studying. As for some papers, see for example Kim [3] , Cai and Xu [4] , Wu et al. [5] , Cai and Xiao [6] and among others.
In recent years, quantile regression with missing data has attracted scholars’ considerable attention. There are several methods, such as complete-case (CC) analysis method, inverse probability weighted method (IPW) and imputation method to handle the missing data. For example, Wei et al. [7] proposed a multiple imputation estimator for parameter estimation in linear QR with missing covariates. Sherwood et al. [8] suggested the inverse probability weighted (IPW) method for linear QR when the covariates are missing at random. Chen et al. [9] also examined the estimation of linear QR model based on nonparametric inverse probability weighted, estimating equations projection, and a combination of both when observations are missing at random. Sherwood [10] investigated the variable selection for the additive partially linear quantile regression with missing covariates. Zhao et al. [11] studied several IPW estimators for parameters in QR when covariates or responses are subject to missing not at random.
It is well known that empirical likelihood (EL) method, introduced by Owen [12] [13] , has many advantages in constructing confidence intervals. For example, it does not need to construct a pivot quantity, and the confidence regions shape and orientation are determined entirely by the data itself. Some scholars have used this method to QR, and some good theoretical results have been obtained under this framework. See for example, Chen and Hall [14] , Wang and Zhu [15] , Tang and Leng [16] , Zhao et al. [17] , Zhao and Tang [18] , Luo and Pang [19] , Zhao and Zhou [20] , Zhou et al. [21] . However, estimation equations based on quantile regression models are not differentiable at parameter points, such that EL method fails to achieve the higher order accuracy. To achieve the higher-order asymptotic refinements, Whang [22] proposed to smooth the estimating equations for the empirical likelihood under the linear QR models. Later, Lv and Li [23] proposed the smoothed empirical likelihood (SEL) for partially linear quantile regression models with missing response, and the SEL statistics for the parameters and the nonparametric part were defined, and the asymptotic Chi-squared distributions were shown. Recently, for the linear QR models with missing response at random, Luo et al. [24] proposed three SEL ratios for the regression parameter, and the asymptotical distributions were shown to be standard
distribution under some conditions. Linear quantile regression models offer a flexible approach in many applications. It is also of considerable interest to investigate nonlinear quantile models. As far as we know, there is little work done for nonlinear quantile models with missing responses at random. Just like mentioned in Koenker [2] and other literature about nonlinear models, the computation of the entire process in the nonlinear case is considerably more challenging than the linear case where the computation task is quite easy. So it is not directly to extend the work of Whang [22] and Luo et al. [24] to nonlinear quantile models because of the complexity of nonlinear models with missing responses at random. Therefore, the main purpose of this paper is to develop the smoothed EL inferences on
with missing responses at random. The rest of this paper is organized as follows. In Section 2, the smoothed empirical likelihood ratios for the parametric vector are proposed, and the asymptotic properties of the proposed empirical log-likelihood ratios are investigated in Section 3. Section 4 is the proofs of the main results. Conclusions are given in Section 5.
2. Methodology
In this paper, we consider the nonlinear quantile regression model
(2.1)
where
is a d-dimensional covariate and
is a response variable,
is a p-dimensional parameter vector, and
is an independent and identically distributed random variable, which satisfies
, where the quantile level
. For simplicity, we write
as
throughout this paper. For model (2.1), we focus on the case where all values of
are completely observed, but some values of response
are missing. That is, we have the incomplete observations
from model (2.1), where
is an indicator variable, when
can be observed, then
, and when
is missing with
. Throughout this paper, we assume that
is missing at random (MAR), and the MAR assumption satisfies
(2.2)
The formula (2.2) implies that
is conditionally independent with
for given
, and
is called the propensity score or selection probability function.
2.1. Smoothed Quantile Empirical Likelihood with Complete-Case Data
In Koenker [2] , if
can be observed, the quantile estimator
of the parameter
in model (2.1) is obtained by minimizing the following objective function
(2.3)
where
is the quantile loss function and
is the indicator function.
can be obtained by solving the following equation, which is the optimal condition corresponding to (2.3), i.e.
(2.4)
where
,
is the quantile score function, and
if
is the true value.
Since some values of response
from model (2.1) are missing, with MAR assumption we can prove that
(2.5)
So based on complete-case data, the quantile estimator
of the parameter
is the solution of
(2.6)
As pointed out by Whang [22] , the function
in (2.6) is not differentiable at point
. This will cause some difficulties in higher-order asymptotic analysis, since most of the empirical likelihood-based research is based on a smooth function of sample moments. Then following Whang [22] , let
denote a bounded kernel function that is compactly supported on
and integrated to one, define
,
, where h is a positive bandwidth, then a smooth function of
is defined as
(2.7)
It can be proved that
is asymptotically unbiased.
Introducing the auxiliary random vector
(2.8)
According to the above discussion, we know that
, so the smoothed empirical log-likelihood ratio function of parameter
with complete-case data can be defined as
(2.9)
where log(.) is the logarithmic fuction based on e. If zero is inside the convex hull of
, then a unique value for
exists. Using the Lagrange multiplier method and some simple calculations,
can be written as
(2.10)
where
is a Lagrange multiplier which is determined by
(2.11)
2.2. Smoothed Weighted Quantile Empirical Likelihood
Similar to Section 2.1, we introduce the following auxiliary random vector
(2.12)
Using the MAR assumption, we can prove that
if
is the true value, thus the smoothed weighted quantile empirical log-likelihood ratio function for
can be defined accordingly.
Since (2.12) contains an unknown function
which needs to be estimated first. We can use a kernel smoothing method to estimate. Specifically,
can be defined by
(2.13)
where
is the d-dimensional kernel function, and a is the bandwidth.
When the dimension of the covariate X is very high, the nonparametric estimation will encounter with the curse of dimensionality. In this case, a parametric approach might be more feasible for the estimation of
. A commonly used model is the logistic regression given by
(2.14)
where
,
is
-dimensional unknown parameter vector. Here
can be estimated by maximizing the log-likelihood function
(2.15)
Let
be the maximum likelihood estimation of
, then the parameter estimator of
can be written as
. If the parametric model for
is correctly specified, the inverse probability weighted method is consistent and feasible.
For convenience, we use
to represent the estimation of
, it can be the estimator estimated by the parameter method or by nonparametric method. Denote
(2.16)
the smoothed weighted quantile empirical log-likelihood ratio function of parameter
is
(2.17)
2.3. Smoothed Imputed Quantile Empirical Likelihood
From above discussion, we can see neither approach makes full use of the information contained in the data. As pointed out in Xue [25] , discarding the missing data may lead to incorrect conclusion when there are a lot of missing values in the considered data set. To resolve the issue, we first use the nonlinear quantile imputation to impute
by
, with
obtained by (2.6), this kind of imputation is also used by Zhao and Tang [18] and Zhou et al. [21] . With the imputed value in hand, and then using the inverse probability weighted technique, we define the final imputed value by
(2.18)
where
is given in Section 2.2. Then the imputation based auxiliary random vector is
(2.19)
Accordingly, the smoothed imputed quantile empirical log-likelihood ratio function for
is defined as
(2.20)
The ratio is more appropriate, because it sufficiently uses the information contained in the data.
3. Asymptotic Properties
In this section, we will give the asymptotic distributions for the three smoothed quantile empirical log-likelihood ratios proposed in Section 2.1-2.3. Firstly, we give some symbols and assumptions that needed in proof.
Assuming the probability density function of X is
, let
and
be the density and distribution function of error
conditional on
, respectively.
C1.
, are independent and identically distributed random vectors.
C2. Both
and
have bounded derivatives up to order r almost surely, and
.
C3.
is a kernel function of order r, and there is a constant
, such that
.
C4. The kernel function
has bounded and compactly supported on
, and for a constant
, it satisfies
C5. For
, let
, where
. For any
satisfying
, there is a partition on
:
, such that
is either strictly positive or strictly negative on
.
C6. The bandwidth h satisfies
as
.
C7. The matrices
and
defined in Lemma 3 of Section 4 are non-singular.
C8.
, where
.
C9. The bandwidth a satisfies
.
C10. The maximum likelihood estimation
of
is
-consistent and satisfies the regularity condition of asymptotic normality.
The following Theorem states the asymptotic distribution of
.
Theorem 1. Suppose that conditions in C1-C10 hold, and
is the true value of the parameter, then
where
can be
,
is a chi-square distribution with d degrees of freedom, and
represents convergence in distribution.
According to the above theorem, the confidence region of the parameter
can be constructed. More specifically, for a given
with
, let
satisfies
, then the approximate
confidence region for
can be defined as
(3.1)
4. Proofs
Before giving the proof of the main theorem, some lemmas are useful for proving the main theorem.
Lemma 1 Suppose conditions C2, C3 and C8 hold, then
Lemma 1 is the Lemma 2 in Xue [25] , so the proof is omitted.
Lemma 2 Suppose that conditions C2, C10 hold and the
is correctly specified, then
where
is the true value.
Lemma 2 is the lemma A.2 of Tang and Zhao [26] , please see the proof details in Tang and Zhao [26] .
Lemma 3. Suppose conditions C1-C10 hold, then as
, we have
1)
2)
When
,
,
;
When
,
,
;
When
,
,
.
Proof a) We first prove the lemma when
.
For result 1), By a change of variable, we have
and then
(4.1)
Obviously, the first term on the right-hand of Equation (4.1) is
, and then applying Taylor expansion to the second term on the right-hand of Equation (4.1) can obtain result 1).
For result 2), noticing that
(4.2)
Obviously, the first term on the right-hand of Equation (4.2) is
, and again applying Taylor expansion to the second term on the right-hand of Equation (4.2) can obtain result 2).
b) When
, by direct calculation, we can derive
(4.3)
In addition, we can prove
(4.4)
According to Lemma 1 and Lemma 2, we can obtain that
for both nonparametric estimator and parameter estimator of
. Using this result and (4.4), we can derive
(4.5)
Further derivation leads to
(4.6)
Similar to the proof in the case of a), it can be seen that
(4.7)
Obviously, the first term on the right-hand of the Equation (4.7) is
, and using Taylor expansion to the second term on the right hand of (4.7), and result 1) is proved. Similarly, result 2) is also obtained.
c) When
, direct calculation yield
(4.8)
Using Taylor expansion to
at
, we have
Noticing that
and combined with
, we can derive
Using
and
, then we get
(4.9)
So we obtain
(4.10)
and then
(4.11)
Similar to the proof in case a), it is obvious that the first term on the right-hand of Equation (4.11) is
. Similarly, we can obtain result 2) with
.
Lemma 4 Suppose conditions C1-C10 hold, then as
,
1)
2)
uniformly in
, with
, where
when
;
when
;
when
, and
, and
are defined in Lemma 3.
Proof: By Taylor expansion, we derive
(4.12)
where
, and
lies between
and
.
a) This lemma will be proved first for the case of
.
Similar to the proof of Lemma 2 of Whang [22] , using Cauchy-Schwartz inequality, triangle inequality and arguments similar to the proof of Lemma 3, we have
(4.13)
with
. Therefore, according to (4.12), (4.13), law of iterated logarithm, Lemma 3 and condition
, it holds that
The proof for the first result is completed. The second result can be proved in a similar way, here we omit the details.
b) When
, according to the Equations (4.5) and (4.10) respectively, and by the similar arguments for
, we can derive the two results, here we omit the details.
Proof of Theorem 1 By the Lagrange multiplier method,
can be represented as
(4.14)
where
is the solution of the following equation
(4.15)
Similar to the proof in Owen [13] , we can prove that
(4.16)
Using Taylor expansion for (4.14), combined with Lemma 4 and (4.16), we obtain
(4.17)
When
is
and
, together with the results of Lemma 3 and Lemma 4 respectively, the asymptotic distribution of the smoothed empirical log-likelihood ratio can be proved to be a chi-square distribution with d degrees of freedom.
5. Conclusion
In this paper, we propose three smoothed empirical log-likelihood ratio functions for nonlinear model parameters with missing responses. We obtain the corresponding Wilks phenomenon under some regular conditions, and can easily construct the confidence interval of the parameters. For the type of data missing, we only consider the case where the covariate data is complete and the response variable data is missing. In addition, there are cases where the covariate data is missing and the response variable data is complete. Therefore, the smooth empirical likelihood inference for nonlinear quantile regression models with missing covariates is also worth studying.
Funding
Wang’s researches are supported by NSF project (ZR2021MA077) of Shandong Province of China.