skip to main content
10.1145/3478905.3479002acmotherconferencesArticle/Chapter ViewAbstractPublication PagesdsitConference Proceedingsconference-collections
research-article

Absolute Redundancy Analysis Based on Features Selection

Published:28 September 2021Publication History

ABSTRACT

The goal of feature selection (FS) in machine learning is to find the best subset of features to create efficient models for a learning task. Different FS methods are then used to assess features relevancy. An efficient feature selection method should be able to select relevant and non-redundant features in order to improve learning performance and training efficiency on large data. However in the case of non-independents features, we saw existing features selection methods inappropriately remove redundancy which leads to performance loss. We propose in this article a new criteria for feature redundancy analysis. Using our proposed criteria, we then design an efficient features redundancy analysis method to eliminate redundant features and optimize the performance of a classifier. We experimentally compare the efficiency and performance of our method against other existing methods which may remove redundant features. The results obtained show that our method is effective in maximizing performance while reducing redundancy.

References

  1. Thushara Amarasinghe, Achala Aponso, and Naomi Krishnarajah. 2018. Critical Analysis of Machine Learning Based Approaches for Fraud Detection in Financial Transactions. In Proceedings of the 2018 International Conference on Machine Learning Technologies (Jinan, China) (ICMLT ’18). Association for Computing Machinery, New York, NY, USA, 12–17. https://doi.org/10.1145/3231884.3231894Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Girish Chandrashekar and Ferat Sahin. 2014. A Survey on Feature Selection Methods. Comput. Electr. Eng. 40, 1 (Jan. 2014), 16–28. https://doi.org/10.1016/j.compeleceng.2013.11.024Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Dheeru Dua and Casey Graff. 2017. UCI Machine Learning Repository. http://archive.ics.uci.edu/mlGoogle ScholarGoogle Scholar
  4. Boli Fang, Miao Jiang, P. Cheng, J. Shen, and Yi Fang. 2020. Achieving Outcome Fairness in Machine Learning Models for Social Decision Problems. In IJCAI.Google ScholarGoogle Scholar
  5. Bahar Farahani, Mojtaba Barzegari, and Fereidoon Shams Aliee. 2019. Towards Collaborative Machine Learning Driven Healthcare Internet of Things. In Proceedings of the International Conference on Omni-Layer Intelligent Systems (Crete, Greece) (COINS ’19). Association for Computing Machinery, New York, NY, USA, 134–140. https://doi.org/10.1145/3312614.3312644Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. I. Gheyas and L. Smith. 2010. Feature subset selection in large dimensionality domains. Pattern Recognit. 43(2010), 5–13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Mark Andrew Hall. 1999. Correlation-based feature selection for machine learning. (1999).Google ScholarGoogle Scholar
  8. George H. John, Ron Kohavi, and Karl Pfleger. 1994. Irrelevant Features and the Subset Selection Problem. In Proceedings of the Eleventh International Conference on International Conference on Machine Learning(New Brunswick, NJ, USA) (ICML’94). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 121–129.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Ioannis K. Paparrizos, B. B. Cambazoglu, and A. Gionis. 2011. Machine learned job recommendation. In RecSys ’11.Google ScholarGoogle Scholar
  10. Hanchuan Peng, Fuhui Long, and Chris Ding. 2005. Feature Selection Based on Mutual Information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 27, 8 (Aug. 2005), 1226–1238. https://doi.org/10.1109/TPAMI.2005.159Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Zhiwei (Tony) Qin, Jian Tang, and Jieping Ye. 2019. Deep Reinforcement Learning with Applications in Transportation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discover & Data Mining(Anchorage, AK, USA) (KDD ’19). Association for Computing Machinery, New York, NY, USA, 3201–3202. https://doi.org/10.1145/3292500.3332299Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Laura Elena Raileanu and Kilian Stoffel. 2004. Theoretical comparison between the gini index and information gain criteria. Annals of Mathematics and Artificial Intelligence 41, 1(2004), 77–93.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. G. Thippa Reddy, M. Praveen Kumar Reddy, Kuruva Lakshmanna, Rajesh Kaluri, Dharmendra Singh Rajput, Gautam Srivastava, and Thar Baker. 2020. Analysis of Dimensionality Reduction Techniques on Big Data. IEEE Access 8(2020), 54776–54788. https://doi.org/10.1109/ACCESS.2020.2980942Google ScholarGoogle ScholarCross RefCross Ref
  14. Matthias Ring and Bjoern M. Eskofier. 2016. An Approximation of the Gaussian RBF Kernel for Efficient Classification with SVMs. Pattern Recogn. Lett. 84, C (Dec. 2016), 107–113. https://doi.org/10.1016/j.patrec.2016.08.013Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. S. Salzberg. 1994. C4.5: Programs for Machine Learning by J. Ross Quinlan. Morgan Kaufmann Publishers, Inc., 1993. Machine Learning 16(1994), 235–240.Google ScholarGoogle ScholarCross RefCross Ref
  16. Nirmalya Thakur and Chia Y. Han. 2020. An Approach for Detection of Walking Related Falls During Activities of Daily Living. In 2020 International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE). 280–283. https://doi.org/10.1109/ICBAIE49996.2020.00066Google ScholarGoogle Scholar
  17. Adnan Ullah, Usman Qamar, Farhan Hassan Khan, and Saba Bashir. 2017. Dimensionality Reduction Approaches and Evolving Challenges in High Dimensional Data. In Proceedings of the 1st International Conference on Internet of Things and Machine Learning (Liverpool, United Kingdom) (IML ’17). Association for Computing Machinery, New York, NY, USA, Article 67, 8 pages. https://doi.org/10.1145/3109761.3158407Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. B. Venkatesh and J. Anuradha. 2019. A Review of Feature Selection and Its Methods. Cybernetics and Information Technologies 19 (2019), 26 – 3.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Mei Wang, Xinrong Tao, and Fei Han. 2020. A New Method for Redundancy Analysis in Feature Selection. In 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence (Sanya, China) (ACAI 2020). Association for Computing Machinery, New York, NY, USA, Article 21, 5 pages. https://doi.org/10.1145/3446132.3446153Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. L. Yu and H. Liu. 2004. Efficient Feature Selection via Analysis of Relevance and Redundancy. J. Mach. Learn. Res. 5(2004), 1205–1224.Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    DSIT 2021: 2021 4th International Conference on Data Science and Information Technology
    July 2021
    481 pages
    ISBN:9781450390248
    DOI:10.1145/3478905

    Copyright © 2021 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 28 September 2021

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate114of277submissions,41%
  • Article Metrics

    • Downloads (Last 12 months)14
    • Downloads (Last 6 weeks)1

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format