Abstract
Machine learning is currently one of the hottest topics that enable machines to learn from data and build predictions without being explicitly programmed for that task, automatically without human involvement. Supervised learning is one of two broad branches of machine learning that makes the model enable to predict future outcomes after they are trained based on past data where we use input/output pairs or the labeled data to train the model with the goal to produce a function that is approximated enough to be able to predict outputs for new inputs when introduced to them. Supervised learning problems can be grouped into regression problems and classification problems. A regression problem is when outputs are continuous whereas a classification problem is when outputs are categorical. This paper tries to compare different types of classification algorithms precisely widely used ones on the basis of some basic conceptions though it is obvious that a complete and comprehensive review and survey of all the supervised learning classification algorithms possibly cannot be accomplished by a single paper, but the references cited in this paper hopefully cover the significant theoretical issues and our survey has been kept limited to the widely used algorithms because the field is highly growing and not possible to cover all the algorithms in a single paper. One more point to be mentioned here that any study of complex procedure like neural networks has not been included as it has been tried to keep the content as much simple as possible.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
References
J.H. Friedman, Regularized discriminant analysis. J. Am. Stat. Assoc. 84(405), 165–175 (1989)
N. Friedman, D. Geiger, M. Goldszmidt, Bayesian network classifiers. Mach. Learn. 29, 131–163 (1997)
N. Friedman, D. Koller, Being Bayesian about network structure: A Bayesian approach to structure discovery in Bayesian networks. Mach. Learn. 50(1), 95–125 (2003)
R.G. Cowell, Conditions under which conditional independence and scoring methods lead to identical selection of Bayesian network models, in Proceedings of 17th International Conference on Uncertainty in Artificial Intelligence
R.L. De Mantaras, E. Armengol, Machine learning from examples: inductive and lazy methods. Data Knowl. Eng. 25(1–2), 99–123 (1998)
D. Heckerman, C. Meek, G. Cooper, A Bayesian approach to causal discovery, in Computation, Causation, and Discovery, ed. by C. Glymour, G. Cooper (MIT Press, Cambridge, 1999), pp. 141–165
N. Japkowicz, S. Stephen, The class imbalance problem: a systematic study. Intell. Data Anal. 6(5), 429–449 (2002)
D.E. Rumelhart, G.E. Hinton, R.J. Williams, Learning internal representations by error propagation, in Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol. 1, ed. by D.E. Rumelhart, J.L. McClelland et al. (MIT Press, Cambridge, MA), pp. 318–362.
A. Roy, On connectionism, rule extraction, and brain-like learning. IEEE Trans. Fuzzy Syst. 8(2), 222–227; L. Breiman, Bagging predictors. Mach. Learn. 24, 123–140
I.J. Good, Probability and the Weighing of Evidence (London, Charles Grin)
N.J. Nilsson, Learning Machines (McGraw-Hill, New York)
B. Cestnik, I. Kononenko, I. Bratko, Assistant 86: a knowledge elicitation tool for sophisticated users, in Proceedings of the Second European Working Session on Learning, pp. 31–45
B. Cestnik, Estimating probabilities: a crucial task in machine learning, in Proceedings of the European Conference on Artificial Intelligence, pp. 147–149
T. Cover, P. Hart, Nearest neighbor pattern classification. IEEE Trans. Inf. Theor. 13(1), 21–27 (1967)
W. Cohen, Fast effective rule induction, in Proceedings of ICML-95, pp. 115–123
J.M. Kalyan Roy, Image similarity measure using color histogram, color coherence vector, and sobel method. Int. J. Sci. Res. (IJSR) 2(1), 538–543 (2013), India Online ISSN: 23197064
A. Smola, S. Vishwanathan, Introduction to Machine Learning (United Kingdom at the University Press, Cambridge, 2010)
R.G. Cowell, Conditions under which conditional independence and scoring methods lead to identical selection of Bayesian network models, in Proceedings of 17th International Conference on Uncertainty in Artificial Intelligence (2001)
W. Gerstner, Supervised learning for neural networks: a tutorial with JAVA exercises
R. Olshen L. Breiman, J.H. Friedman, “Classification and regression trees.” Belmont CA Wadsworth International group, 1984. B. C. U. P.E.tgoff, “Multivariate decision trees: machine learning,” no. 19, 1995, pp. 45–47.
T. Dietterich, M. Kearns, Y. Mansour, Applying the weak learning framework to understand and improve C4. 5 (Sanfrancisco, Morgan), pp. 96–104
Kufmann, in Proceeding of the 13th International Conference on Machine Learning (1996)
K.M.A. Chai, H.L. Chieu, H.T. Ng, Bayesian online classifiers for text classification and filtering, in Proceedings of the 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (August 2002), pp. 97–104
T. Elomaa, The biases of decision treepruning strategies (Springer, 1999), Lecture Notes in Computer Science, vol. 1642, pp. 63–74
A. Kalousis, G. Gama, On data and algorithms: understanding inductive performance. Mach. Learn. 54, 275–312 (2004)
P. Brazdil, C. Soares, J. Da Costa, ranking learning algorithms: using IBL and meta-learning on accuracy and time results. Mach. Learn. 50, 251–277 (2003)
G. Batista, M.C. Monard, An analysis of four missing data treatment methods for supervised learning. Appl. Artif. Intell. 17, 519–533 (2003)
J. Basak, R. Kothari, A classification paradigm for distributed vertically partitioned data. Neural Comput. 16(7), 1525–1544 (2004)
A. Blum, Empirical support for winnow and weighted-majority algorithms: results on a calendar scheduling domain. Mach. Learn. 26(1), 5–23 (1997)
A. Bonarini, An Introduction to Learning Fuzzy Classifier Systems (2000), Lecture Notes in Computer Science, vol. 1813, pp. 83–92
R. Bouckaert, Choosing between two learning algorithms based on calibrated tests, in Proceedings of 20th International Conference on Machine Learning (Morgan Kaufmann, 2003), pp. 51–58
R. Bouckaert, Naive Bayes Classifiers that Perform Well with Continuous Variables (2004), Lecture Notes in Computer Science, vol. 3339, pp. 1089–1094
P. Brazdil, C. Soares, J. Da Costa, ranking learning algorithms: using IBL and meta-learning on accuracy and time results. Mach. Learn. 50, 251–277 (2003)
L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone, Classification and Regression Trees (Wadsforth International Group, 1984)
L. Breiman, Bagging predictors. Mach. Learn. 24, 123–140 (1996)
L.A. Breslow, D.W. Aha, Simplifying decision trees: a survey. Knowl. Eng. Rev. 12, 1–40 (1997)
H. Brighton, C. Mellish, Advances in instance selection for instance-based learning algorithms. Data Min. Knowl. Disc. 6, 153–172 (2002)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Sen, P.C., Hajra, M., Ghosh, M. (2020). Supervised Classification Algorithms in Machine Learning: A Survey and Review. In: Mandal, J., Bhattacharya, D. (eds) Emerging Technology in Modelling and Graphics. Advances in Intelligent Systems and Computing, vol 937. Springer, Singapore. https://doi.org/10.1007/978-981-13-7403-6_11
Download citation
DOI: https://doi.org/10.1007/978-981-13-7403-6_11
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-7402-9
Online ISBN: 978-981-13-7403-6
eBook Packages: EngineeringEngineering (R0)