Abstract
We construct a class of elementary nonparametric output predictors of an unknown discrete-time nonlinear fading memory system. Our algorithms predict asymptotically well for every bounded input sequence, every disturbance sequence in certain classes, and every linear or nonlinear system that is continuous and asymptotically time-invariant, causal, and with fading memory. The predictor is based on k n -nearest neighbor estimators from nonparametric statistics. It uses only previous input and noisy output data of the system without any knowledge of the structure of the unknown system, the bounds on the input, or the properties of noise. Under additional smoothness conditions we provide rates of convergence for the time-average errors of our scheme. Finally, we apply our results to the special case of stable LTI systems.
This work was partially supported by the National Science Foundation under NYI grant IRI-9457645.
This research was completed while Posner was at the Department of Electrical Engineering, Princeton University and the Department of Statistics, University of Toronto.
Preview
Unable to display preview. Download preview PDF.
References
T.M. Cover, “Estimation by the nearest neighbor rule,” IEEE Trans. Information Theory, vol. IT-14, pp. 50–55, Jan. 1968.
M.A. Dahleh, E.D. Sontag, D.N.C. Tse, and J.N. Tsitsiklis, “Worst-case identification of nonlinear fading memory systems,” Automatica, vol. 31, pp. 503–508, 1995.
M.A. Dahleh, T. Theodosopoulos, and J.N. Tsitsiklis, “The sample complexity of worst-case identification for f.i.r. linear systems,” Syst. Contr. Lett., vol. 20, pp. 157–166, March 1993.
L. Devroye, “Necessary and sufficient conditions for the pointwise convergence of nearest neighbor regression function estimates,” Z. Wahrscheinlichkeitstheorie verw. Gebiete vol. 61, pp. 467–481, 1982.
R.M. Dudley, Real Analysis and Probability, Chapman & Hall, 1989.
J.D. Farmer and J.J. Sidorowich, “Exploiting chaos to predict the future and reduce noise,” Evolution, Learning, and Cognition, pp. 265–289, World Scientific, Singapore, 1988.
M. Feder, N. Merhav, and M. Gutman, “Universal prediction of individual sequences,” IEEE Trans. Information Theory, vol. 38, pp. 1258–1270, July 1992.
W. Greblicki, “Nonparametric identification of Wiener systems by orthogonal series,” IEEE Trans. Automatic Control, vol. 30, pp. 2077–2086, 1994.
W. Greblicki and M. Pawlak, “Dynamic system identification with order statistics,” IEEE Trans. Information Theory, vol. 40, pp. 1474–1489, 1994.
L. Györfi, “The rate of convergence of k n -NN regression estimates and classification rules,” IEEE Trans. Information Theory, vol. IT-27, pp. 362–364, May 1981.
A.J. Helmicki, C.A. Jacobson, C.N. Nett, “Control Oriented System Identification: A Worst-Case/Deterministic Approach in H ∞,” IEEE Trans. Automatic Control, vol. 36, Oct. 1991.
A. Krzyżak, “On estimation of a class of nonlinear systems by the kernel regression estimate,” IEEE Trans. Information Theory, vol. 36, pp. 141–152, 1990.
A. Krzyżak, “Identification of nonlinear systems by recursive kernel regression estimates,” Int. J. Systems Sci., vol. 24, pp. 577–598, 1993.
S.R. Kulkarni, “Data-dependent nearest neighbor and kernel estimators consistent for arbitrary processes,” preprint.
S.R. Kulkarni and S.E. Posner, “Rates of convergence of nearest neighbor estimation under arbitrary sampling,” IEEE Trans. Information Theory, pp. 1028–1039, July 1995.
S.R. Kulkarni and S.E. Posner, “Nonparametric output prediction for nonlinear fading memory systems,” to appear IEEE Trans. Automatic Control, Nov., 1998.
L. Ljung, System Identification: Theory for the User, Prentice-Hall, 1987.
P.M. Mäkilä, “Robust identification and Galois sequences,” Int. J. Contr., vol. 54, pp. 1189–1200, 1991.
G. Morvai, S. Yakowitz, and L. Györfi, “Nonparametric inferences for ergodic stationary time series,” preprint.
K. Poolla and A. Tikku, “On the time complexity of worst-case system identification,” IEEE Trans. Automatic Control, vol. 39, pp. 944–950, May 1994.
S.E. Posner, “Nonparametric estimation, regression, and prediction under minimal regularity conditions,” Ph.D. Thesis, Department of Electrical Engineering, Princeton University, 1995.
C.J. Stone, “Consistent nonparametric regression,” Ann. Stat., vol. 8, pp. 1348–1360, 1977.
D.N.C. Tse, M.A. Dahleh, J.N. Tsitsiklis, “Optimal asymptotic identification under bounded disturbances,” IEEE Trans. Automatic Control, vol. 38, pp. 1176–1190, Aug. 1993.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this paper
Cite this paper
Kulkarni, S.R., Posner, S.E. (1999). Universal output prediction and nonparametric regression for arbitrary data. In: Yamamoto, Y., Hara, S. (eds) Learning, control and hybrid systems. Lecture Notes in Control and Information Sciences, vol 241. Springer, London. https://doi.org/10.1007/BFb0109733
Download citation
DOI: https://doi.org/10.1007/BFb0109733
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-85233-076-7
Online ISBN: 978-1-84628-533-2
eBook Packages: Springer Book Archive