Summary
In this paper we combine the concepts of local smoothing and fitting with basis functions for multivariate predictor variables. We start with arbitrary basis functions and show that the asymptotic variance at interior points is independent of the choice of the basis. Moreover we calculate the asymptotic variance at boundary points. We are not able to compute the asymptotic bias since a Taylor theorem for arbitrary basis functions does not exist. For this reason we focus on basis functions without interactions and derive a Taylor theorem which covers this case. This theorem enables us to calculate the asymptotic bias for interior as well as for boundary points. We demonstrate how advantage can be taken of the idea of local fitting with general basis functions by means of a simulated data set, and also provide a data-driven tool to optimize the basis.
Similar content being viewed by others
References
Cleveland, W. S. (1979). Robust locally weighted regression and smoothing scatterplots. J. Amer. Statist. Assoc. 74, 829–836.
Cleveland, W. S. and Devlin, S. (1988). Locally weighted regression: An approach to regression analysis by local fitting. J. Amer. Statist. Assoc. 83, 596–610.
Einbeck, J. (2001). Local fitting with general basis functions, SFB 386, Discussion Paper No. 256. https://doi.org/www.stat.uni-muenchen.de/∼einbeck/powerpap06.ps.
Fahrmeir, L. and Hamerle, A. (1984). Multivariate statistische Verfahren. Berlin / New York: de Gruyter.
Fan, J. (1992). Design-adaptive nonparametric regression. J. Amer. Statist. Assoc. 87, 998–1004.
Nadaraya, E. A. (1964). On estimating regression. Theory Prob. Appl. 10, 186–190.
Ramsay, J. O. and Silverman, B. W. (1997). Functional Data Analysis. New York: Springer.
Ruppert, D. and Wand, M. P. (1994). Multivariate locally weighted least squares regression. Ann. Statist. 22, 1346–1370.
Staniswalis, J. G., Messer, K., and Finston, D. R. (1993). Kernel estimators for multivariate regression. Nonparametric Statistics 3, 103–121.
Stone, C. J. (1977). Consistent nonparametric regression. Ann. Statist. 5, 595–645.
Wand, M. P. (1992). Error analysis for general multivariate kernel estimators. Nonparametric Statistics 2, 1–15.
Wand, M. P. and Jones, M. C. (1993). Comparison of smoothing parametrizations in bivariate kernel density estimation. J. Amer. Statist. Assoc. 88, 520–528.
Watson, G. S. (1964). Smooth regression analysis. Sankhyā, Series A, 26, 359–372.
Yang, L. and Tschering, R. (1999). Multivariate bandwidth selection for local linear regression. J. R. Statist. Soc. B 61, 793–815.
Acknowledgements
The author is grateful to Gerhard Tutz (LMU, Dep. of Statistics) for helpful inspirations during this work, Daniel Rost (LMU, Math. Inst.) for contributions concerning the extendibility of Taylor’s theorem and to the referees for many suggestions which improved this paper. This work was finished while the author passed a term at the University São Paulo. Many thanks especially to Carmen D. S. de André and Júlio M. Singer (USP) for their support in various fields.
Parts of this work were presented and discussed at the Euroworkshop on Statististical Modelling (2001). There a variety of valuable comments and suggestions were given that enhanced the paper.
Author information
Authors and Affiliations
Appendices
Appendix
A Regularity conditions
-
(A1) The kernel K is bounded with compact support, ∫ uuTK(u)du = μ2Id, where μ2 is a scalar and Id the d × d identity matrix. In addition, all odd-order moments of K vanish, i.e. \(\int {u_1^{{l_1}}\; \cdots \;u_d^{{l_d}}K\left( u \right)du = 0} \) for all nonnegative integers l1, …, ld with an odd sum.
-
(A2) The point x is ∈ supp(f). At x, σ2 is continuous, f is continuously differentiable and all second-order derivatives of m are continuous. Further f(x) > 0, σ2(x) > 0.
-
(A3) The sequence of bandwidth matrices H1/2 is such that n−1∣H∣−1/2 and each entry of H tends to zero as n → ∞.
-
(A4) For a boundary point x, there exists a value on the boundary of supp(f) with x = xb + H1/2c, where c is a fixed element of supp(K), and a convex set \({\cal C}\) with nonnull interior containing xb, such that \(x \in {\cal C}\).
(A5) At x, all basis functions are continuously differentiable (for variance expressions in Theorem 1,3,4) resp. twice continuously differentiable (for bias expressions in Theorem 3 and 4). In either case, the point x is non-singular for all basis functions, i.e. ∇φj(x) ≠ 0 for j = 1, …, q.
For explanations and interpretations of conditions (Al) to (A4) see Ruppert & Wand (1994).
B Proofs
B.1 Proof of Theorem 1
Let 1 be a matrix of appropriate dimension having only entries equal to 1, further let
Note that for any u ∈ ℝd
holds. Let \({C_{x,H}} = \left\{ {t:{H^{ - 1/2}}\left( {t - x} \right) \in {{\cal D}_{x,H}}} \right\}\). For interior and boundary points we derive
and analogously
Substituting (19) and (20) into (6) leads to (8). In the special case of an interior point we have \({M_x} = \left( {\begin{array}{*{20}c} 1 \; 0 \\ 0 \; {{\mu _2}{{\bf{I}}_d}} \\ \end{array} } \right)\) and \({N_x} = \,\left( {\begin{array}{*{20}c} {{\nu _0}} \; 0 \\ 0 \; {\int u{u^T}\;{K^2}\left( u \right)du} \\ \end{array} } \right)\). Thus (8) reduces to
B.2 Proof of Theorem 2
We introduce the function M : [0, 1] → ℝ,
Then we have M(0) = m(x) and M(1) = m(z). We apply the univariate Taylor theorem on the function M ∈ Cp+1([0,1]) and obtain
where
Using the Inverse Function Theorem we obtain
Repeated application of the chain rule on M = m ○ yΦ leads to
Applying the latter formulas in (23) and substituting ζ = yΦ(τ) proves the allegation.
B.3 Proof of Theorem 3
The proof is kept shortly since it follows mainly the ideas of the corresponding proof for multivariate local linear fitting, see Ruppert & Wand (1994).
Asymptotic Bias
First note that, applying (10), we have
with
and Sm(x) = o(Qm(x)). Plugging (24) into (5) shows that
Let Wi = KH(Xi − x). Using matrix algebra (see e.g. Fahrmeir & Hamerle (1984)) we derive
and
so that substituting (26) and (27) into (25) proves (11).
Asymptotic variance
Similar like above we obtain
where
Plugging this result and (26) into (6) leads to (12).
B.4 Proof of Theorem 4
Let
Asymptotic bias
Note that
where Cx,h was defined in the proof of Theorem 1. Using the first step in (27),
holds. Assuming (A4), Mx is nonsingular and we have
where \(\mu_{x}^{11}=\left(\mu_{x, 11}-\mu_{x, 12} \mu_{x, 22}^{-1} \mu_{x, 21}\right)^{-1}, \mu_{x}^{12}=-\left(\mu_{x, 12} / \mu_{x, 11}\right) \mu_{x}^{22}\)and \(\mu_{x}^{22}=\left(\mu_{x, 22}-\mu_{x, 21} \mu_{x, 12} / \mu_{x, 11}\right)^{-1}\). Then substituting (28) and (29) into (25) and noticing that
yields formula (14).
Asymptotic variance
Similar considerations like in (28) lead to
With (6), (28) and (30) we get
what had to be proven.
Rights and permissions
About this article
Cite this article
Einbeck, J. Multivariate Local Fitting with General Basis Functions. Computational Statistics 18, 185–203 (2003). https://doi.org/10.1007/s001800300140
Published:
Issue Date:
DOI: https://doi.org/10.1007/s001800300140