An improve unsupervised discretization using optimization algorithms for classification problems

This paper addresses the classification problem in machine learning focusing on predicting class labels for datasets with continuous features. Recognizing the critical role of discretization in enhancing classification performance, the study integrates equal width binning (EWB) with two optimization...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohamed, Rozlini, Samsudin, Noor Azah
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:http://eprints.uthm.edu.my/11092/1/J17588_57f47374243fbf4ecaff387b409cf14f.pdf
http://eprints.uthm.edu.my/11092/
https://doi.org/10.11591/ijeecs.v34.i2
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This paper addresses the classification problem in machine learning focusing on predicting class labels for datasets with continuous features. Recognizing the critical role of discretization in enhancing classification performance, the study integrates equal width binning (EWB) with two optimization algorithms: the bat algorithm (BA), referred to as EB, and the whale optimization algorithm (WOA), denoted as EW. The primary objective is to determine the optimal technique for predicting relevant class labels. The paper emphasizes the significance of discretization in data preprocessing, offering a comprehensive approach that combines discretization techniques with optimization algorithms. An investigative study was undertaken to assess the efficiency of EB and EW by evaluating their classification performance using Naive Bayes and K-nearest neighbor algorithms on four continuous datasets sourced from the UCI datasets. According to the experimental findings, the suggested EB has a major effect on the accuracy, recall, and F-measure of data classification. The classification performance using EB outperforms other existing approaches for all datasets.