PERTANIKA JOURNAL OF SCIENCE AND TECHNOLOGY

 

e-ISSN 2231-8526
ISSN 0128-7680

Home / Regular Issue / JST Vol. 32 (5) Aug. 2024 / JST-4731-2023

 

Learning Discriminative Features Using ANN-based Progressive Learning Model for Efficient Big Data Classification

Nandita Bangera and Kayarvizhy

Pertanika Journal of Science & Technology, Volume 32, Issue 5, August 2024

DOI: https://doi.org/10.47836/pjst.32.5.06

Keywords: Artificial neural network, big data classification, data imbalance, Pearson correlation coefficient-based feature selection, progressive learning model, weight updating

Published on: 26 August 2024

Progressive techniques encompass iterative and adaptive approaches that incrementally enhance and optimize data processing by iteratively modifying the analysis process, resulting in improved efficiency and precision of outcomes. These techniques contain a range of components, such as data sampling, feature selection, and learning algorithms. This study proposes the integration of an Artificial Neural Network (ANN) with a Progressive Learning Model (PLM) to enhance the efficacy of learning from large-scale datasets. The SMOTE and Pearson Correlation Coefficient (PCC) methods are commonly employed in imbalanced dataset handling and feature selection. The utilization of progressive weight updating is a notable strategy for improving performance optimization in neural network models. This approach involves the incremental modification of the network’s progressive weights during the training phase rather than relying on gradient values. The proposed method gradually obtains the localization of discriminative data by incorporating information from local details into the overall global structure, effectively reducing the training time by iteratively updating the weights. The model has been examined using two distinct datasets: the Poker hand and the Higgs. The performance of the suggested method is compared with that of classification algorithms: Population and Global Search Improved Squirrel Search Algorithm (PGS-ISSA) and Adaptive E-Bat (AEB). The convergence of Poker’s is achieved after 50 epochs with ANN-PLM; however, without PLM, it takes 65 epochs. Similarly, with the Higgs, convergence is achieved after 25 epochs with PLM and 40 without PLM.

  • Abhilasha, A., & Naidul, P. A. (2022). Self-boosted with dynamic semi-supervised clustering method for imbalanced big data classification. International Journal of Software Innovation, 10(1), 1-24. https://doi.org/10.1007/s11042-022-12038-4

  • Ali, I. M. S., & Balakrishnan, M. (2021). Population and global search improved squirrel search algorithm for feature selection in big data classification. International Journal of Intelligent Engineering & Systems, 14(4), 177-189. https://doi.org/10.22266/ijies2021.0831.17

  • Al-Thanoon, N. A., Algamal, Z. Y., & Qasim, O. S. (2021). Feature selection based on a crow search algorithm for big data classification. Chemometrics and Intelligent Laboratory Systems, 212, Article 104288. https://doi.org/10.1016/j.chemolab.2021.104288

  • Banchhor, C., & Srinivasu, N. (2021). Analysis of Bayesian optimization algorithms for big data classification based on map reduce framework. Journal of Big Data, 8(1), Article 81. https://doi.org/10.1186/s40537-021-00464-4

  • Basgall, M. J., Naiouf, M., & Fernández, A. (2021). FDR2-BD: A fast data reduction recommendation tool for tabular big data classification problems. Electronics, 10(15), Article 1757.https://doi.org/10.3390/electronics10151757

  • BenSaid, F., & Alimi, A. M. (2021). Online feature selection system for big data classification based on multi-objective automated negotiation. Pattern Recognition, 110, Article 107629. https://doi.org/10.1016/j.patcog.2020.107629

  • Brahmane, A. V., & Krishna, B. C. (2021). Big data classification using deep learning and apache spark architecture. Neural Computing and Applications, 33(2),15253-15266. https://doi.org/ 10.1007/s00521-021-06145-w

  • Castro, F. M., Marín-Jiménez, M. J., Guil, N., Schmid, C., & Alahari, K. (2018, September 8-14). End-to-end incremental learning. [Paper presentation]. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany. https://doi.org/10.48550/arXiv.1807.09536

  • Chatterjee, S., Javid, A. M., Sadeghi, M., Mitra, P. P., & Skoglund, M. (2017). Progressive learning for systematic design of large neural networks. arXiv, Article 1710.08177. https://doi.org/10.48550/arXiv.1710.08177

  • Dubey, A. K., Kumar, A., & Agrawal, R. (2021). An efficient ACO-PSO-based framework for data classification and preprocessing in big data. Evolutionary Intelligence, 14, 909-922. https://doi.org/10.1007/s12065-020-00477-7

  • Du, R., Xie, J., Ma, Z., Chang, D., Song, Y. Z., & Guo, J. (2021). Progressive learning of category-consistent multi-granularity features for fine-grained visual classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(12), 9521-9535. https://doi.org/10.1109/TPAMI.2021.3126668

  • Hassanat, A. B., Ali, H. N., Tarawneh, A. S., Alrashidi, M., Alghamdi, M., Altarawneh, G. A., & Abbadi, M. A. (2022). Magnetic force classifier: A novel method for big data classification. IEEE Access, 10,12592-12606. htttps://doi.org/10.1109/ACCESS.2022.3142888

  • Hassib, E. M., El-Desouky, A. I., Labib, L. M., & El-Kenawy, E. S. M. (2020). WOA+ BRNN: An imbalanced big data classification framework using whale optimization and deep neural network. Soft Computing, 24(8), 5573-5592. https://doi.org/10.1007/s00500-019-03901-y

  • Jain, D. K., Boyapati, P., Venkatesh, J., & Prakash, M. (2022). An intelligent cognitive-inspired computing with big data analytics framework for sentiment analysis and classification. Information Processing Management, 59(1), Article 102758. https://doi.org/10. 1016/j.ipm.2021.102758

  • Juez-Gil, M., Arnaiz-Gonzalez, A., Rodriguez, J. J., Lopez-Nozal, C., & Garcia-Osorio, C. (2021). Approx-SMOTE: Fast SMOTE for big data on Apache spark. Neurocomputing, 464, 432-437. https://doi.org/10.1016/j.neucom.2021.08.086

  • Kantapalli, B., & Markapudi, B. R. (2023). SSPO-DQN spark: Shuffled student psychology optimization based deep Q network with spark architecture for big data classification. Wireless Networks, 29(1),369-385.https://doi.org/10.1007/s11276-022-03103-9

  • Li, Z., Liu, C., Yuille, A., Ni, B., Zhang, W., & Gao, W. (2021, June 19-25). Progressive stage-wise learning for unsupervised feature representation enhancement. [Paper presentation]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA. https://doi.org/10.48550/arXiv.2106.05554

  • Movassagh, A. A., Alzubi, J. A., Gheisari, M., Rahimi, M., Mohan, S., Abbasi, A. A., & Nabipour, N. (2021). Artificial neural networks training algorithm integrating invasive weed optimization with differential evolutionary model. Journal of Ambient Intelligence and Humanized Computing, 14, 6017–6025. https://doi.org/10.1007/s12652-020-02623-6

  • Mujeeb, S. M., Sam, R. P., & Madhavi, K. (2021). Adaptive exponential bat algorithm and deep learning for big data classification. Sādhanā, 46(1), Article 15. https://doi.org/10.1007/ s12046-020-01521-z

  • Park, S. T., Kim, D. Y., & Li, G. (2021). An analysis of environmental big data through the establishment of emotional classification system model based on machine learning: Focus on multimedia contents for portal applications. Multimedia Tools and Applications, 80, 34459-34477. https://doi.org/ 10.1007/s11042-020-08818-5

  • Rebuffi, S. A., Kolesnikov, A., Sperl, G., & Lampert, C. H. (2017). icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 2001-2010). IEEE Publishing. https://doi.org/10.48550/arXiv.1611.07725

  • Rusu, A. A., Rabinowitz, N. C., Desjardins, G., Soyer, H., Kirkpatrick, J., Kavukcuoglu, K., Pascanu R. & Hadsell, R. (2016). Progressive neural networks. arXiv, Article 1606.04671. https://doi.org/10.48550/arXiv.1606.04671

  • Siddiqui, Z. A., & Park, U. (2021). Progressive convolutional neural network for incremental learning. Electronics, 10(16), Article 1879. https://doi.org/10.3390/electronics10161879

  • Sleeman IV, W. C., & Krawczyk B. (2021). Multi-class imbalanced big data classification on spark. Knowledge-Based Systems, 212, Article 106598. https://doi.org/10.1016/j.knosys. 2020. 106598

  • Venkatesan, R., & Er, M. J. (2016). A novel progressive learning technique for multi-class classification. Neurocomputing, 207, 310-321. https://doi.org/10.1016/j.neucom.2016.05.006

  • Wang, H., Xiao, M., Wu, C., & Zhang, J. (2021). Distributed classification for imbalanced big data in distributed environments. Wireless Networks, 2021, 1-12. https://doi.org/10.1007/s11276-021-02552-y

  • Xing, W., & Bei, Y. (2019). Medical health big data classification based on KNN classification algorithm. IEEE Access, 8, 28808-28819. https://doi.org/10.1109/ACCESS.2019.2955754

  • Zhou, J., Li, J., Wang, C., Wu, H., Zhao, C., & Wang, Q. (2021). A vegetable disease recognition model for complex background based on region proposal and progressive learning. Computers and Electronics in Agriculture, 184, Article 106101. https://doi.org/10.1016/j.compag.2021.106101

ISSN 0128-7680

e-ISSN 2231-8526

Article ID

JST-4731-2023

Download Full Article PDF

Share this article

Related Articles