|
83 | 83 | }
|
84 | 84 | }
|
85 | 85 | \details{
|
86 |
| - If \code{zerocor} is absent or given as \code{"none"}, |
| 86 | + If \code{zerocor} is missing, or given as \code{"none"}, |
87 | 87 | this function computes the fixed bandwidth kernel estimator of the
|
88 | 88 | probability density on the real line.
|
89 | 89 |
|
90 | 90 | If \code{zerocor} is given, it is assumed that the density
|
91 | 91 | is confined to the positive half-line, and a boundary correction is
|
92 |
| - applied: |
| 92 | + applied to compensate for bias arising due to truncation at the origin: |
93 | 93 | \describe{
|
94 |
| - \item{weighted}{The contribution from each point \eqn{x_i}{x[i]} |
| 94 | + \item{\code{zerocor="weighted"}:}{ |
| 95 | + The contribution from each data point \eqn{x_i}{x[i]} |
95 | 96 | is weighted by the factor \eqn{1/m(x_i)}{1/m(x[i])}
|
96 | 97 | where \eqn{m(x) = 1 - F(-x)} is the total mass of the kernel centred on
|
97 | 98 | \eqn{x} that lies in the positive half-line, and \eqn{F(x)} is the
|
98 |
| - cumulative distribution function of the kernel} |
99 |
| - \item{convolution}{The estimate of the density \eqn{f(r)} is |
| 99 | + cumulative distribution function of the kernel. |
| 100 | + This is the \dQuote{cut-and-normalization} method of |
| 101 | + Gasser and \ifelse{latex}{\out{M\"{u}ller}}{Mueller} (1979). |
| 102 | + Effectively the kernel is renormalized so that it integrates to 1, |
| 103 | + and the adjusted kernel conserves mass. |
| 104 | + } |
| 105 | + \item{\code{zerocor="convolution"}:}{ |
| 106 | + The estimate of the density \eqn{f(r)} is |
100 | 107 | weighted by the factor \eqn{1/m(r)} where \eqn{m(r) = 1 - F(-r)}
|
101 | 108 | is given above.
|
| 109 | + This is the \dQuote{convolution}, \dQuote{uniform} |
| 110 | + or \dQuote{zero-order} boundary correction method |
| 111 | + often attributed to Diggle (1985). |
| 112 | + This correction does not conserve mass. |
102 | 113 | }
|
103 |
| - \item{reflection}{ |
| 114 | + \item{\code{zerocor="reflection"}:}{ |
104 | 115 | if the kernel centred at data point \eqn{x_i}{x[i]}
|
105 | 116 | has a tail that lies on the negative half-line, this tail is
|
106 | 117 | reflected onto the positive half-line.
|
| 118 | + This is the reflection method first proposed by |
| 119 | + Boneva et al (1971). |
| 120 | + This correction conserves mass. |
107 | 121 | }
|
108 |
| - \item{bdrykern}{The density estimate is computed using the |
109 |
| - Boundary Kernel associated with the chosen kernel |
| 122 | + \item{\code{zerocor="bdrykern"}:}{ |
| 123 | + The density estimate is computed using the |
| 124 | + Linear Boundary Kernel associated with the chosen kernel |
110 | 125 | (Wand and Jones, 1995, page 47).
|
111 | 126 | That is, when estimating the density \eqn{f(r)} for values of
|
112 | 127 | \eqn{r} close to zero (defined as \eqn{r < h} for all kernels
|
113 | 128 | except the Gaussian), the kernel contribution
|
114 | 129 | \eqn{k_h(r - x_i)}{k[h](r - x[i])} is multiplied by a
|
115 |
| - term that is a linear function of \eqn{r - x_i}{r - x[i]}. |
| 130 | + term that is a linear function of \eqn{r - x_i}{r - x[i]}, |
| 131 | + with coefficients depending on \eqn{r}. |
| 132 | + This correction does not conserve mass and may result in |
| 133 | + negative values, but is asymptotically optimal. |
116 | 134 | }
|
117 |
| - \item{JonesFoster}{ |
| 135 | + \item{\code{zerocor="JonesFoster"}:}{ |
118 | 136 | The modification of the Boundary Kernel estimate
|
119 |
| - proposed by Jones and Foster (1996), equal to |
| 137 | + proposed by Jones and Foster (1996) is computed. This is equal to |
120 | 138 | \eqn{
|
121 | 139 | \overline f(r) \exp( \hat f(r)/\overline f(r) - 1)
|
122 | 140 | }{
|
123 | 141 | f#(r) exp(f*(r)/f#(r) - 1)
|
124 | 142 | }
|
125 | 143 | where \eqn{\overline f(r)}{f#(r)} is the convolution estimator
|
126 |
| - and \eqn{\hat f(r)}{f*(r)} is the boundary kernel estimator. |
| 144 | + and \eqn{\hat f(r)}{f*(r)} is the linear boundary kernel estimator. |
| 145 | + This ensures that the estimate is always nonnegative |
| 146 | + and retains the asymptotic optimality of the linear boundary kernel. |
127 | 147 | }
|
128 | 148 | }
|
129 | 149 | If \code{fast=TRUE}, the calculations are performed rapidly using
|
|
152 | 172 | \references{
|
153 | 173 | \ournewpaper
|
154 | 174 |
|
| 175 | + Boneva, L.I., Kendall, D.G. and Stefanov, I. (1971) |
| 176 | + Spline transformations: three new diagnostic aids for the |
| 177 | + statistical data-analyst (with discussion). |
| 178 | + \emph{Journal of the Royal Statistical Society, Series B}, |
| 179 | + \bold{33}, 1–-70. |
| 180 | +
|
| 181 | + Diggle, P.J. (1985) |
| 182 | + A kernel method for smoothing point process data. |
| 183 | + \emph{Journal of the Royal Statistical Society, Series C (Applied Statistics)}, |
| 184 | + \bold{34} 138–-147. |
| 185 | + |
| 186 | + Gasser, Th. and \ifelse{latex}{\out{M\"{u}ller}}{Mueller}, H.-G. (1979). |
| 187 | + Kernel estimation of regression functions. |
| 188 | + In Th. Gasser and M. Rosenblatt (editors) |
| 189 | + \emph{Smoothing Techniques for Curve Estimation}, pages |
| 190 | + 23-–68. Springer, Berlin. |
| 191 | +
|
155 | 192 | Jones, M.C. and Foster, P.J. (1996)
|
156 | 193 | A simple nonnegative boundary correction method for kernel density
|
157 | 194 | estimation.
|
|
0 commit comments