A Novel Approach for Feature Selection

  • Conference paper
  • First Online:
Information Systems Design and Intelligent Applications

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 339))

  • 1703 Accesses

Abstract

Clustering problem suffers from the curse of dimensionality. Dimensionality reduction of a feature set refers to the problem of selecting relevant features which produce the most predictive outcome and similarity functions that use all input features with equal relevance may not be effective. We introduce an algorithm that discovers clusters by different combinations of dimensions via local weightings of features. This approach avoids the risk of loss of information encountered in global dimensionality reduction techniques, and does not assume any data distribution model. Our method associates to each cluster a weight vector, whose values capture the relevance of features within the corresponding cluster. To judge the efficiency of the proposed method the results are experimentally compared with other optimization methods such as Genetic Algorithm (GA), Particle Swarm Optimization (PSO) and Differential Evolution (DE) for feature selection.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (Canada)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 299.00
Price excludes VAT (Canada)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 379.99
Price excludes VAT (Canada)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97, 245–271 (1997)

    Google Scholar 

  2. Hall, M.A.: Correlation-based feature selection for machine learning. Ph.D. thesis, The University of Waikato (1999)

    Google Scholar 

  3. Kohavi, R.: Wrappers for performance enhancement and oblivious decision graphs. Ph.D. thesis, Stanford University (1995)

    Google Scholar 

  4. Bottou, L., Vapnik, V.: Local learning algorithms. Neural Comput. 4(6), 888–900 (1992)

    Article  Google Scholar 

  5. Satapathy, S.C., Naik, A.: Hybridization of Rough Set and Differential Evolution Technique for Optimal Features Selection, vol. 132, pp. 453–460. Springer-AISC, Heidelberg (2012)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ch. Swetha Swapna .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer India

About this paper

Cite this paper

Swetha Swapna, C., Vijaya Kumar, V., Murthy, J.V.R. (2015). A Novel Approach for Feature Selection. In: Mandal, J., Satapathy, S., Kumar Sanyal, M., Sarkar, P., Mukhopadhyay, A. (eds) Information Systems Design and Intelligent Applications. Advances in Intelligent Systems and Computing, vol 339. Springer, New Delhi. https://doi.org/10.1007/978-81-322-2250-7_87

Download citation

  • DOI: https://doi.org/10.1007/978-81-322-2250-7_87

  • Published:

  • Publisher Name: Springer, New Delhi

  • Print ISBN: 978-81-322-2249-1

  • Online ISBN: 978-81-322-2250-7

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics

Navigation