Skip to main content

Abstract

The rapidly growing big data generated by connected sensors, devices, the web and social network platforms, etc., have stimulated the advancement of data science, which holds tremendous potential for problem solving in various domains. How to properly utilize the data in model building to obtain accurate analytics and knowledge discovery is a topic of great importance in data mining, and wherefore two issues arise: how to select a critical subset of features and how to select a critical subset of data points for sampling. This paper presents ongoing research that suggests: 1. the critical feature dimension problem is theoretically intractable, but simple heuristic methods may well be sufficient for practical purposes; 2. there are big data analytic problems where evidence suggest that the success of data mining depends more on the critical feature dimension than the specific features selected, thus a random selection of the features based on the dataset’s critical feature dimension will prove sufficient; and 3. The problem of critical sampling has the same intractable complexity as critical feature dimension, but again simple heuristic methods may well be practicable in most applications; experimental results with several versions of the heuristic method are presented and discussed. Finally, a set of metrics for data quality is proposed based on the concepts of critical features and critical sampling.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(1), 245–271 (1997)

    Article  MathSciNet  Google Scholar 

  2. Domingo, C., Gavaldà, R., Watanabe, O.: Adaptive sampling methods for scaling up knowledge discovery algorithms. Data Min. Knowl. Discov. 6(2), 131–152 (2002)

    Article  MathSciNet  Google Scholar 

  3. UCI machine learning repository. http://archive.ics.uci.edu/ml. Accessed 20 Oct 2017

  4. Garey, M.R., Johnson, D.S.: A Guide to the Theory of NP-Completeness, p. 70. WH Freemann, New York (1979)

    Google Scholar 

  5. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)

    MATH  Google Scholar 

  6. National Research Council: Frontiers in massive data analysis. The National Academies Press, Washington, DC (2013)

    Google Scholar 

  7. Papadimitriou, C.H., Yannakakis, M.: The complexity of facets (and some facets of complexity). J. Comput. Syst. Sci. 28, 244–259 (1984)

    Article  MathSciNet  Google Scholar 

  8. Provost, F., Jensen, D., Oates, T.: Progressive sampling. In: Liu, H., Motoda, H. (eds.) Instance Selection and Construction for Data Mining, pp. 151–170. Springer, Boston (2001)

    Chapter  Google Scholar 

  9. Suryakumar, D.: The critical dimension problem; no compromise feature selection. Ph.D. dissertation. New Mexico Institute of Mining and Technology (2013)

    Google Scholar 

  10. Suryakumar, Divya, Sung, A. H., and Liu, Q.: The critical dimension problem: No compromise feature selection. In: Proceedings of eKNOW 2014, the Sixth International Conference on Information, Process, and Knowledge Management, pp. 145–151. IARIA, Barcelona, Spain (2014)

    Google Scholar 

Download references

Acknowledgements

The authors gratefully acknowledge the reviewer of earlier versions of our papers whose insightful comments point them to fruitful directions of study, and their other colleagues and students who contributed to many helpful discussions and experiments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Andrew H. Sung .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ribeiro, B., Silva, J., Sung, A.H., Suryakumar, D. (2018). Critical Feature Selection and Critical Sampling for Data Mining. In: Ganapathi, G., Subramaniam, A., Graña, M., Balusamy, S., Natarajan, R., Ramanathan, P. (eds) Computational Intelligence, Cyber Security and Computational Models. Models and Techniques for Intelligent Systems and Automation. ICC3 2017. Communications in Computer and Information Science, vol 844. Springer, Singapore. https://doi.org/10.1007/978-981-13-0716-4_2

Download citation

  • DOI: https://doi.org/10.1007/978-981-13-0716-4_2

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-13-0715-7

  • Online ISBN: 978-981-13-0716-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics