Estimation of Hazard Function for Censoring Random Variable by Using Wavelet Decomposition and Evaluation of MISE, AMSE with Simulation ()
The failure time which is denoted by
,
, is the time that failure occurs for each individual. It’s not always possible to observe the failure time for each individual in such cases that censorship occurs.
Survival function, which is shown by
, indicates the ratio of people who survived since the base time which is the point they enter the experiment to the time unit t analysis. Hazard function for the failure continuous time is as follows:
![](https://www.scirp.org/html/1-2870020x\fbd4f774-e61b-4dfe-b4ec-2e033726e38b.jpg)
(1)
In this paper, we obtain estimator hazard function for censoring data by using wavelet method. We evaluate convergence ratio of given estimator by simulation.
2. Estimation of Hazard Function by Using Wavelet Method
Wavelets can be used for transient phenomena analysis or function analysis which sometimes changes rapidly. They are symmetrical and have limited period. A close relationship between wavelet coefficients and some spaces is wavelet bases orthogonally. Also useful properties of them in wavelet issues simplify the computational algorithms. As a result, numerous articles have been published about in statistical science.
The mathematical theorem of wavelets and their application in statistics have been studied as a technique for density function estimator, by Harr [1], Doukhan [2], Antoniadys [3], nonparametric curve estimators by Malat [4], Meyer [5], Daubechies [6], Donoho [7], Kyacharyan and Picard [8], Hall and Patil [9] have found a formula for the Mean Integrated Squared Error of Nonlinear Wavelet based on density estimators. Antoniadys et al. [10] achieved the density function estimator and the hazard function for right-censored data with the wavelets. Daubechies [11] studied and discussed the compactly supported wavelets which produce orthogonal bases. Afshari et al. [12-14] studied about density, derivative density function estimator, regression function for the mixing random variables.
Let the nested sequence of closed subspaces;
be a multiresolutuon approximation to
. Define
,
to be orthogonal complement of
in
. Wavelets basis for function
as scaling function
and mother wavelet
such that
forms an orthogonal basis for
and
forms an orthonormal basis for
. Other wavelets in the basis are then generated by translation of the scaling function and dilations of the mother wavelet by using the relationships:
(2)
Given above Wavelet basis, a function
can be written a formal expansion:
(3)
where
![](https://www.scirp.org/html/1-2870020x\52548b65-980f-45f1-9151-12e1560a9618.jpg)
As for general orthogonal series estimator, Daubechies [4], density estimator can be writhen as:
(4)
where the obvious coefficient estimator can be written:
(5)
In this article, we divide time axis into two parts, the intervals and the number of events in each interval. We determine number of events and hazard function according to the observations. Then we flatten them separately via linear wavelet density estimation on the whole time and then we calculate the function estimator and evaluate the asymptotic distribution.
Suppose
are failure time of n tests that are studied. They are non-negative, independent, identically distributed, with the density function f and distribution function F. Also suppose that
are corresponding to censored times, non-negative, independent, identically distributed, with the density function
and distribution function
.
Assuming independency of failure times and censored time of the observed random variable,
and the function
and hazard function are shown as below:
![](https://www.scirp.org/html/1-2870020x\6a55b7a5-65a7-4b0c-9a95-145f9027643f.jpg)
Such that
is indicator function of A. For data censoring, if
,
![](https://www.scirp.org/html/1-2870020x\55b4825b-8cd2-4cf6-bd6d-4f5fa380ec15.jpg)
We assume that,
![](https://www.scirp.org/html/1-2870020x\136450ca-d392-4b68-9e0a-844b5d8be2a8.jpg)
Such that
then we can write as follows:
(6)
To estimate
we need the estimator of
and
.
For estimating
, we divide the time axis into two parts of small intervals and the amounts of events (0 or 1) in each interval, and then we divide these values to the length of intervals.
Estimation procedures of
can be summarized as the following:
Select
and collect the observed failures in
intervals with the length
and using wavelet estimation on the collected data. We find an estimate of sub density. This means that we calculate the collected wavelet coefficients data on the scale of
by choosing the decomposition level
and then we estimate
. It is necessary to state the following symbols to show the details:
![](https://www.scirp.org/html/1-2870020x\4b8659ad-4adb-4716-a2d2-ce80444c9598.jpg)
We figure estimators on the finite interval
in which
. Note that if
is the ordinal order statistic
of the sequence
then,
. In fact we suppose
.
Suppose that N is an integer that could be dependent to n and the estimated points are as follows:
![](https://www.scirp.org/html/1-2870020x\e0e966ea-170c-430b-992d-a864cb33d874.jpg)
Suppose that
and we divide the interval
of time axis to
intervals with
long
![](https://www.scirp.org/html/1-2870020x\e5534591-039f-4b4b-ab93-d1b2dae2e1e1.jpg)
The k-th interval is marked by
so:
for
,
.
Now we define the following indicator function that indicates the number of uncensored failures in the time interval
: ![](https://www.scirp.org/html/1-2870020x\a5ffa43a-876a-4a3e-a04d-6b6b33568279.jpg)
![](https://www.scirp.org/html/1-2870020x\5d62ebde-ffab-444e-a51b-dbc352625760.jpg)
We assume that
is the observed failures ratio in the interval
, in other words: ![](https://www.scirp.org/html/1-2870020x\c9b67334-307d-495f-a3c1-24553d447e05.jpg)
We smooth the data
by an appropriate wavelet smoother to find the estimation of
.
We can write as the following:
(7)
where,
.
The complex structural polymorphism analysis causes an efficient tree construction algorithm for analysis of functions in VN with theoretic scale wavelet coefficients
. However, the integral scale
is not well available and we need an initial value for a fast wavelet transform. Antoniadis [4] suggested the following initial amount:
![](https://www.scirp.org/html/1-2870020x\acfbc0e9-bf77-4f7d-822a-58b8c6a06125.jpg)
As a result a reasonable estimate for image of ![](https://www.scirp.org/html/1-2870020x\2c6e407b-4ca9-4573-8886-e07a5d97bc50.jpg)
with clarity N is: ![](https://www.scirp.org/html/1-2870020x\d073d110-3b64-418c-a5d0-b082ec197569.jpg)
If we assume that the collected values
which are equal to the estimators of
, are in Sobolev space
and
is regular of degree
. We estimate the unknown function
as follows to level the data with a better rate for the sample size
and the sequence
:
(8)
That it is the orthogonal image of
on the leveler approximation space
.
Now we consider an appropriate consistent estimator of
, and finally we estimate the Hazard function.
We assume that
has distribution function
and density function
.
For estimating of
, we use an empirical distribution
as the following:
Such that
is Histogram estimator of
. Suppose that,
, we can write:
![](https://www.scirp.org/html/1-2870020x\67be70cf-30ec-4cf0-a4f1-8ad069cdef47.jpg)
Suppose that
as
, then we define:
so we can write as the following:
(9)
By substituting Equation (9) in Equation (8), we obtain the estimator ![](https://www.scirp.org/html/1-2870020x\3b0f2cdc-d4b5-45e3-b4af-454e7efb2833.jpg)
Theorem: Suppose that the sub density
is a continuous function on
and it’s
times differentiable, If
and ![](https://www.scirp.org/html/1-2870020x\0084aaa0-6501-43b4-9181-4dc5bb6048d1.jpg)
then,
,
![](https://www.scirp.org/html/1-2870020x\e39652c4-1866-45ca-856c-a697a2cb6099.jpg)
Proof:
![](https://www.scirp.org/html/1-2870020x\167b1ba3-fcc2-48eb-8f08-e120e9915453.jpg)
By using Chung-Smirnov property and Taylor’s theorem we can write as the following:
![](https://www.scirp.org/html/1-2870020x\e29c2999-0039-44b5-b029-05cf5354820d.jpg)
(10)
(11)
By using Equations (10) and (11), we can write:
![](https://www.scirp.org/html/1-2870020x\c16278d1-5d54-40d6-a275-f0bd6939c44d.jpg)
then the proof is completed.
3. Numerical Computation and Simulation
In this section, we simulate
and
on the data of size
by using Semlayt’s wavelet. We consider convergence ratio of given estimator by computing of average mean square error of given estimators. We use R software and wavelet package for simulations.
Example 1: We generate
and
from the samples of size
and
with K = 16,
,
and
for optimal surface ![](https://www.scirp.org/html/1-2870020x\223b81a1-21fb-489d-8403-b6a23c8bfa7c.jpg)
The solid line in the Figure 1 displays the wavelet estimate of hazard function with the denoted line representing the true hazard rate The results in Table 1 display the average mean square errors of hazard function estimator for sample sizes
and
.
Example 2: Suppose![](https://www.scirp.org/html/1-2870020x\a9c5bd9c-e298-4ee6-95d8-5d1b5291a830.jpg)
, where
and
. We generate
from sample size of n = 400 and n = 600 with K = 16, K = 32, K = 64 and
.
The solid line in the Figure 2 displays the wavelet estimate of hazard function with the denoted line representing the true hazard rate.
The results in Table 2 display the average mean square
![](https://www.scirp.org/html/1-2870020x\5230f1d6-0273-44e3-9849-70db3b73e810.jpg)
Figure 1. The panel in Figure 1 displays the wavelet estimator of hazard function
with the denoted line representing the true hazard rate.
![](https://www.scirp.org/html/1-2870020x\8bae4193-5311-4219-a817-b26295c30c27.jpg)
Table 1. Average mean square errors of hazard function estimator.
![](https://www.scirp.org/html/1-2870020x\69e1536b-d233-46d9-9b27-c0ccb57741de.jpg)
Figure 2. The solid line in the panel displays the wavelet estimate of hazard function with the denoted line representing the true hazard rate.
![](https://www.scirp.org/html/1-2870020x\853d5469-d48e-4992-913d-87859b3d9caa.jpg)
Table 2. Average mean square errors of hazard function estimator.
errors of hazard function estimator for sample sizes n = 400 and n = 600.
Acknowledgements
The support of Research Committee of Persian Gulf University is greatly acknowledged.
NOTES