Accelerated fuzzy min–max neural network and arithmetic optimization algorithm for optimizing hyper-boxes and feature selection

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

The fuzzy min–max (FMM) neural network effectively solves classification problems. Despite its success, it has been observed recently that FMM has overlapping between hyper-boxes in some datasets which certainly the overall classification performance, as well as FMM has a high compactional complexity, especially when dealing with high-dimensional datasets. a hybrid model combining Arithmetic Optimization Algorithm (AOA) and Accelerated fuzzy min–max (AFMM) neural network is proposed to produce an AFMM-AOA model, where AFMM is used to speed up the hyper-boxes contraction process and to reduce the number of hyper-boxes, then AOA is employed for selecting the optimal feature set in each hyper-box, which results in lowering the compactional complexity and overcoming the overlapping problem. Furthermore, the AOA algorithm has been modified (MAOA) to enhance the exploiting ability of the original AOA algorithm for handling the high dimensionality in hyper-box representation by introducing both random and neighbor search methods. The performance of the proposed methods is evaluated using twelve datasets, as a result, the neighbor search method shows better performance than the random search. In addition, both methods showed superior performance compared with the original AOA and some state-of-the-art algorithms.

Original languageEnglish
Pages (from-to)1553-1568
Number of pages16
JournalNeural Computing and Applications
Volume36
Issue number4
DOIs
StatePublished - Feb 2024

Keywords

  • Arithmetic optimization algorithm
  • Feature selection
  • Fuzzy minimum maximum
  • Hyperbox optimization

Fingerprint

Dive into the research topics of 'Accelerated fuzzy min–max neural network and arithmetic optimization algorithm for optimizing hyper-boxes and feature selection'. Together they form a unique fingerprint.

Cite this