
Feature Selection Method Based on Consecutive Forward Selection and Backward Elimination Concepts Using a Weighted Vector
Feature selection is an essential preprocessing task in many disciplines, including Machine Learning (ML) and the Internet of Things (IoT), and it is the most demanding process for data analysis. This process attempts to identify and remove as much irrelevant and redundant information as possible in a controlled manner. Existing algorithms still have limitations in selecting the most informative features maintaining high classification accuracy results. This study proposed a consecutive Forward selection and Backward Elimination algorithm (FBWV) that enhances feature selection by applying the forward selection concept, backward elimination concept, weighted chi-square vector, and custom decision threshold value. The FBWV model framework was optimized through data preprocessing and parameter tuning. The effectiveness of the proposed method was evaluated by comparing it with other state-of-the-art Feature Selection Algorithms (FSA), namely, Rough Set (RS), Weight- Guided (WG) feature selection, and Stability-correlation and Correlation (ScC). The reduced subsets were trained by several classifiers using different measures, including accuracy, F-measure, reduction rate, and AUC. The results revealed that the FBWV effectively reduced the size of the given datasets. It achieved the highest accuracies of 85.28%, 88.33%, 96.26%, 81.36%, 96%, 74.39%, 81.89%, 65.26%, and 98.69% for Austra, Heart Disease, Phishing, Sonar, Iono, SGC, and SpamBase, respectively. The Messidor and Pop-Failure datasets outperform the other FSAs. Moreover, it achieved the highest F-measure and AUC rates of 97.94% each for the Pop-Failure dataset. The FBWV proved the capability of handling different types of datasets and reduced computational complexity, storage, and cost.
[1] Aggarwal N., Shukla U., Saxena G., Rawat M., Bafila A., Singh S., and Pundir A., “Mean Based Relief: An Improved Feature Selection Method Based on ReliefF,” Applied Intelligence, vol. 53, pp. 23004-23028, 2023. DOI:10.1007/s10489- 023-04662-w
[2] Ali W. and Saeed F., “Hybrid Filter and Genetic Algorithm-Based Feature Selection for Improving Cancer Classification in High-Dimensional Microarray Data,” Processes, vol. 11, no. 2, pp. 1- 22, 2023. https://doi.org/10.3390/pr11020562
[3] Al-Shalabi L., “Hybrid Feature Selection based ScC and Forward Selection Methods,” International Journal of Data and Network Science, vol. 8, no. 2, pp. 1117-1128, 2024. DOI:10.5267/j.ijdns.2023.11.022
[4] Al-Shalabi L., “New Feature Selection Algorithm Based on Feature Stability and Correlation,” IEEE Access, vol. 10, pp. 4699-4713, 2022. DOI:10.1109/ACCESS.2022.3140209
[5] Al-Shalabi L., “Rough Set-Based Reduction of Incomplete Medical Datasets by Reducing the Number of Missing Values,” The International Arab Journal of Information Technology, vol. 16, no. 2, pp. 203-210, 2019. https://iajit.org/portal/PDF/March%202019,%20 No.%202/11460.pdf
[6] Al-Shalabi L., Shaaban Z., and Kasasbeh B., “Data mining: A Preprocessing Engine,” Journal of Computer Science, vol. 2, no. 9, pp. 735-739, 2006. DOI:10.3844/jcssp.2006.735.739
[7] Alsini R., Naz A., Khan H., Bokhari A., Daud A., and Ramzan M., “Using Deep Learning and Word Embeddings for Predicting Human Agreeableness Behavior,” Scientific Reports, vol. 14, no. 29875, pp. 1-19, 2024. https://doi.org/10.1038/s41598- 024-81506-8
[8] Aslam M. and Smarandache F., “Chi-Square Test for Imprecise Data in Consistency Table,” Frontiers in Applied Mathematics and Statistics, vol. 9, pp. 1-5, 2023. DOI:10.3389/fams.2023.1279638
[9] Bermejo P., Gamez J., and Puerta J., “Adapting the CMIM Algorithm for Multilabel Feature Selection. A Comparison with Existing Methods,” Expert Systems, vol. 35, no. 1, pp. e12230, 2017. https://doi.org/10.1111/exsy.12230
[10] Bolon-Canedo V., Sanchez-Marono N., and Alonso-Betanzos A., “A Review of Feature Selection Methods on Synthetic Data,” Knowledge Information Systems, vol. 34, no. 3, pp. 483-519, 2013. https://doi.org/10.1007/s10115-012-0487-8
[11] Bommert A., Sun X., Bischl B., Rahnenfuhrer J., and Lang M., “Benchmark for Filter Methods for Feature Selection in High-Dimensional Classification Data,” Computational Statistics and Data Analysis, vol. 143, pp. 106839, 2020. https://doi.org/10.1016/j.csda.2019.106839
[12] Borboudakis G. and Tsamardinos I., “Forward- Backward Selection with Early Dropping,” The Journal of Machine Learning Research, vol. 20, no. 1, pp. 1-39, 2019. https://jmlr.org/papers/volume20/17-334/17- 334.pdf
[13] Breiman L., “Random Forests,” Machine Learning, vol. 45, pp. 5-32, 2001. https://doi.org/10.1023/A:1010933404324
[14] Chaudhary A., Kolhe S., and Rajkamal, “Performance Evaluation of Feature Selection Methods for Mobile Devices,” International Journal of Engineering Research and Applications, vol. 3, no. 6, pp. 587-594, 2013. https://www.ijera.com/papers/Vol3_issue6/CW36 587594.pdf
[15] Cherrington M., Thabtah F., Lu J., and Xu Q., “Feature Selection: Filter Methods Performance Challenges,” in Proceedings of the International Conference on Computer and Information Sciences, Sakaka, pp. 1-4, 2019. DOI:10.1109/ICCISci.2019.8716478
[16] Chowdhury M. and Turin T., “Variable Selection Strategies and its Importance in Clinical Prediction Modelling,” Fam Med Com Health, vol. 8, no.1, pp. 1-7, 2020. DOI:10.1136/fmch- 2019-000262
[17] Christen P., Hand D., and Kirielle N., “A Review of the F-Measure: Its History, Properties, Criticism, and Alternatives,” ACM Computing Surveys, vol. 56, no. 3, pp. 1-24, 2023. https://doi.org/10.1145/3606367
[18] Cilia N., D’Alessandro T., Stefano C., Fontanella F., and Di Freca A., “Comparing Filter and Wrapper Approaches for Feature Selection in Handwritten Character Recognition,” Pattern Recognition Letters, vol. 168, pp. 39-46, 2023. https://doi.org/10.1016/j.patrec.2023.02.028
[19] Cormen T., Leiserson C., Rivest R., and Stein C., Introduction to Algorithms, MIT Press, 2009. https://enos.itcollege.ee/~japoia/algorithms/GT/I ntroduction_to_algorithms-3rd%20Edition.pdf
[20] Das A., Sengupta S., and Bhattacharyya S., “A Group Incremental Feature Selection for Classification Using Rough Set Theory based Genetic Algorithm,” Applied Soft Computing, vol. 65, pp. 400-411, 2018.
[21] Frederick O., Maxwell O., Ifunanya O., Udochukwu E., Kelechi O., Ngonadi L., and Idris H., “Comparison of Some Variable Selection Techniques in Regression Analysis,” American Journal of Biomedical Science and Research, vol. 812 The International Arab Journal of Information Technology, Vol. 22, No. 4, July 2025 6, no. 4, pp. 281-293, 2019. https://doi.org/10.3389/fams.2023.1279638
[22] Guyon I. and Elisseeff A., “An Introduction to Variable and Feature Selection,” Journal of Machine Learning Research, vol. 3, pp. 1157- 1182, 2003.
[23] Hall M., Feature Selection for Discrete and Numeric Class Machine Learning, Technical Report, 1999. https://ml.cms.waikato.ac.nz/publications/1999/9 9MH-Feature-Select.pdf
[24] Haq A., Zeb A., Lei Z., and Zhang D., “Forecasting Daily Stock Trend Using Multi- Filter Feature Selection and Deep Learning,” Expert Systems with Applications, vol. 168, pp. 114444, 2021. https://doi.org/10.1016/j.eswa.2020.114444
[25] Hoque N., Singh M., and Bhattacharyya D., “EFS- MI: An Ensemble Feature Selection Method for Classification,” Complex Intelligent System, vol. 4, no. 2, pp. 105-118, 2017. https://doi.org/10.1007/s40747-017-0060-x
[26] Hu M., Tsang E., Guo Y., and Xu W., “Fast and Robust Attribute Reduction Based on the Separability in Fuzzy Decision Systems,” IEEE Transactions on Cybernetics, vol. 52, no. 6, pp. 5559-5572, 2022. DOI:10.1109/TCYB.2020.3040803
[27] Huanhuan G., Yanying L., Jiaoni Z., Baoshuang Z., and Xialin W., “A New Filter Feature Selection Algorithm for Classification Task by Ensembling Pearson Correlation Coefficient and Mutual Information,” Engineering Applications of Artificial Intelligence, vol. 131, pp. 107865, 2024. https://doi.org/10.1016/j.engappai.2024.107865
[28] Jenifa G., Padmapriya K., Sevanthi P., Karthika K., Pandi V., and Arumugam D., “An Effective Personality Recognition Model Design using Generative Artificial Intelligence based Learning Principles,” in Proceedings of the International Conference on Computing and Data Science, Chennai, pp. 1-6, 2024. DOI:10.1109/ICCDS60734.2024.10560368
[29] Ke W., Wu C., Wu Y., and Xiong N., “A New Filter Feature Selection Based on Criteria Fusion for Gene Microarray Data,” IEEE Access, vol. 6, pp. 61065-61076, 2018. DOI:10.1109/ACCESS.2018.2873634
[30] Lazar C., Taminau J., Meganck S., Steenhoff D., Coletta A., Molter C., De Schaetzen V., Duque R., Bersini H., and Nowe A., “A Survey on Filter Techniques for Feature Selection in Gene Eexpression Microarray Analysis,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 9, no. 4, pp. 1106-1119, 2012. DOI:10.1109/TCBB.2012.33
[31] Lu J., Zhou S., Feng T., and Tong Z., “A Chi- Squared Approximation-based Weighting Method for Distributed Detection,” in Proceedings of the IEEE Radar Conference, New York, pp. 1-6, 2022. DOI:10.1109/RadarConf2248738.2022.9764160
[32] Maseno E. and Wang Z., “Hybrid Wrapper Feature Selection Method based on Genetic Algorithm and Extreme Learning Machine for Intrusion Detection,” Journal of Big Data, vol. 11, pp. 1-25, 2024. https://doi.org/10.1186/s40537-024-00887- 9
[33] Mohtasham F., Pourhoseingholi M., Hashemi Nazari S., Kavousi K., and Reza Zali M., “Comparative Analysis of Feature Selection Techniques for COVID-19 Dataset,” Scientific Reports, vol. 14, pp. 1-20, 2024. DOI:10.1038/s41598-024-69209-6
[34] Mohtashami M. and Eftekhari M., “A Hybrid Filter-Based Feature Selection Method via Hesitant Fuzzy and Rough Sets Concepts,” Iranian Journal of Fuzzy Systems, vol. 16, no. 2, pp. 165-182, 2019. DOI: 10.22111/ijfs.2019.4550
[35] Natekin A. and Knoll A., “Gradient boosting Machines, A Tutorial,” Frontiers in Neurorobotics, vol. 7, pp. 1-21, 2013. DOI:10.3389/fnbot.2013.00021
[36] Ngo N., Michel P., and Giorgi R., “Multivariate Filter Methods for Feature Selection with the γ- Metric,” BMC Medical Research Methodology, vol. 24, no. 307, pp. 1-22, 2024. DOI:10.1186/s12874-024-02426-9
[37] Parmar M., Sonker A., and Sejwar V., “A Comparative Analysis for Filter-Based Feature Selection Techniques with Tree-based Classification,” International Journal on Recent and Innovation Trends in Computing and Communication, vol. 11, no. 10, pp. 360-369, 2023. https://doi.org/10.17762/ijritcc.v11i10s.7643
[38] Parodi S., Verda D., Bagnasco F., and Muselli M., “The Clinical Meaning of the Area Under a Receiver Operating Characteristic Curve for the Evaluation of the Performance of Disease Markers,” Epidemiology Health, vol. 44, pp. 1-10, 2022. https://doi.org/10.4178/epih.e2022088
[39] Patel D., Saxena A., and Wang J., “A Machine Learning-Based Wrapper Method for Feature Selection,” International Journal of Data Warehousing and Mining, vol. 20, no. 1, pp. 1-33, 2024. https://doi.org/10.4018/IJDWM.352041
[40] Peng C., Lee K., and Ingersoll G., “An Introduction to Logistic Regression Analysis and Reporting,” The Journal of Educational Research, vol. 96, no. 1, pp. 3-14, 2002. https://doi.org/10.1080/00220670209598786
[41] Pirgazi J., Kallehbasti M., Sorkhi A., and Kermani A., “An Efficient Hybrid Filter-Wrapper Method based on Improved Harris Hawks Optimization for Feature Selection,” BioImpacts, vol. 15, pp. 30340, 2025. DOI:10.34172/bi.30340 Feature Selection Method Based on Consecutive Forward Selection and Backward Elimination ... 813
[42] Pourpanah F., Shi Y., Lim C., Hao Q., and Tan C., “Feature Selection Based on Brain Storm Optimization for Data Classification,” Applied Soft Computing Journal, vol. 80, pp. 761-775, 2019. https://doi.org/10.1016/j.asoc.2019.04.037
[43] Qian Y., Liang X., Wang Q., Liang J., Liu B., Skowron A., Yao Y., Ma J., and Dang C., “Local Rough Set: A Solution to Rough Data Analysis in Big Data,” International Journal of Approximate Reasoning, vol. 97, pp. 38-63, 2018. https://doi.org/10.1016/j.ijar.2018.01.008
[44] Quinlan J., “Unknown Attribute Values in Induction,” in Proceedings of the 6th International Workshop on Machine Learning, New York, pp. 164-168, 1989. https://doi.org/10.1016/B978-1- 55860-036-2.50048-5
[45] Rebbah F., Chamlal H., and Ouaderhman T., “Accurate Analysis for Univariate-Based Filter Methods for Microarray Data Classification,” Journal of Algorithms and Computational Technology, vol. 18, pp. 1-11, 2024. DOI:10.1177/17483026241232295
[46] Ringsquandl M., Lamparter S., Brandt S., Hubauer T., and Lepratti R., “Semantic-Guided Feature Selection for Industrial Automation Systems,” in Proceedings of the the Semantic Web-ISWC, Lecture Notes in Computer Science, Bethlehem, pp. 225-240, 2015. https://doi.org/10.1007/978-3-319-25010-6_13
[47] Saeidi S., “Identifying Personality Traits of WhatsApp Users Based on Frequently Used Emojis Using Deep Learning,” Multimedia Tools and Applications, vol. 83, pp. 13873-13886, 2024. DOI:10.1007/s11042-023-15209-z
[48] Sinayobye J., Kaawaase K., Kiwanuka F., and Musabe R., “Hybrid Model of Correlation Based Filter Feature Selection and Machine Learning Classifiers Applied on Smart Meter Data Set,” in Proceedings of the IEEE/ACM Symposium on Software Engineering in Africa, Montreal, pp. 1- 10, 2019. DOI:10.1109/SEiA.2019.00009
[49] Soheili M. and Moghadam A., “DQPFS: Distributed Quadratic Programming-Based Feature Selection for Big Data,” Journal of Parallel and Distributed Computing, vol. 138, pp. 1-14, 2020. https://doi.org/10.1016/j.jpdc.2019.12.001
[50] Velayutham C. and Thangavel K., “Unsupervised Quick Reduct Algorithm Rough Set Theory,” Journal of Electronic Science and Technology, vol. 9, no. 3, pp. 193-201, 2011.
[51] Velusamy K. and Manavalan R., “Performance Analysis of Unsupervised Classification based on Optimization,” International Journal of Computer Applications, vol. 42, no. 19, pp. 22-27, 2012. DOI:10.5120/5801-8090
[52] Wah Y., Ibrahim N., Hamid H., Abdul-Rahman S., and Fong S., “Feature Selection Methods: Case of Filter and Wrapper Approaches for Maximizing Classification Accuracy,” Pertanika Journal of Science and Technology, vol. 26, no. 1, pp. 329- 340, 2018. https://www.researchgate.net/publication/322920 304_Feature_selection_methods_Case_of_filter_ and_wrapper_approaches_for_maximising_classi fication_accuracy
[53] Wisniewski G. and Yvon F., “Fast Large-Margin Learning for Statistical Machine Translation,” International Journal of Computational Linguistics and Applications, vol. 4, no. 2, pp. 45- 62, 2013.
[54] Xue B., Zhang M., and Browne W., “A Comprehensive Comparison on Evolutionary Feature Selection Approaches to Classification,” International Journal of Computational Intelligence and Applications, vol. 14, no. 2, pp. 1-22, 2015. https://doi.org/10.1142/S146902681550008X
[55] Zhang Y., Gong D., Hu Y., and Zhang W., “Feature Selection Algorithm Based on Bare Bones Particle Swarm Optimization,” Neurocomputing, vol.148, pp. 150-157, 2015. https://doi.org/10.1016/j.neucom.2012.09.049
[56] Zhu Z., Ong Y., and Dash M., “Wrapper-Filter Feature Selection Algorithm Using a Memetic Framework,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 37, no. 1, pp. 70-76, 2007. DOI:10.1109/TSMCB.2006.883267