Abstract
Selective sampling, a part of the active learning method, reduces the cost of labeling supplementary training data by asking only for the labels of the most informative, unlabeled examples. This additional information added to an initial, randomly chosen training set is expected to improve the generalization performance of a learning machine. We investigate some methods for a selection of the most informative examples in the context of one-class classification problems i.e. problems where only (or nearly only) the examples of the so-called target class are available. We applied selective sampling algorithms to a variety of domains, including real-world problems: mine detection and texture segmentation. The goal of this paper is to show why the best or most often used selective sampling methods for two- or multi-class problems are not necessarily the best ones for the one-class classification problem. By modifying the sampling methods, we present a way of selecting a small subset from the unlabeled data to be presented to an expert for labeling such that the performance of the retrained one-class classifier is significantly improved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
A. Blum, T. Mitchell ‘Combining labeled and unlabeled data with co-training’ Proceedings of the 1998 Conference on Computational Learning Theory
C. Cambell, N. Cristianini, A. Smola, ‘Query learning with large margin classifiers’
D. Cohn, L. Atlas, R. Ladner, ‘Improving generalization with active learning’, 1992
D. Cohn, Z. Ghahramani, M.I. Jordan ‘Active learning with statistical models’, Journal of artificial intelligence research 4, 1996 129–145
Y. Freund, H. S. Seung, E. Shamir, N. Tishby, ‘Selective sampling using the query by committee algorithm’, Machine Learning, 28, 133–168 (1997)
N. Japkowicz, ‘Concept-learning in the absence of counter-examples: an autoassociation-based approach to classification’, PhD thesis 1999
M.J. Kearns, U.V. Vazirani, ‘An introduction to computational learning theory’, The MIT Press 1994, ISBN 0-262-11193-4
Ion Muslea, Steve Minton, Craig Knoblock, ‘Selective sampling with redundant views’, Proceedings of the 15th National Conference on Artificial Intelligence, 621–626, AAAI-2000.
D.M.J. Tax and P. Juszczak, ‘Kernel whitening for data description’, International Workshop on Pattern Recognition with Support Vector Machines 2002
D.M.J. Tax, ‘One-class classification’, PhD thesis, Delft University of Technology, ISBN:90-75691-05-x, 2001
D.M.J. Tax, R.P.W. Duin, ‘Support Vector Data Description’, Pattern Recognition Letters, December 1999, vol. 20(11–13), pp. 1191–1199
M.K. Warmuth, G. Rätsch, M. Mathieson, J. Liao, C. Lemmen, ‘Active learning in the drug discovery process’
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Juszczak, P., Duin, R.P.W. (2003). Selective Sampling Methods in One-Class Classification Problems. In: Kaynak, O., Alpaydin, E., Oja, E., Xu, L. (eds) Artificial Neural Networks and Neural Information Processing — ICANN/ICONIP 2003. ICANN ICONIP 2003 2003. Lecture Notes in Computer Science, vol 2714. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44989-2_18
Download citation
DOI: https://doi.org/10.1007/3-540-44989-2_18
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40408-8
Online ISBN: 978-3-540-44989-8
eBook Packages: Springer Book Archive