Comparative performance of deep learning and machine learning algorithms on imbalanced handwritten data

Imbalanced data is one of the challenges in a classification task in machine learning. Data disparity produces a biased output of a model regardless how recent the technology is. However, deep learning algorithms, such as deep belief networks showed promising results in many domains, especially in...

Full description

Bibliographic Details
Main Authors: Amri, A’inur A’fifah, Ismail, Amelia Ritahani, Zarir, Abdullah Ahmad
Format: Article
Language:English
English
English
Published: The Science and Information (SAI) Organization 2018
Subjects:
Online Access:http://irep.iium.edu.my/62456/
http://irep.iium.edu.my/62456/
http://irep.iium.edu.my/62456/7/62456-Comparative%20performance%20of%20deep%20learning%20_scopus.pdf
http://irep.iium.edu.my/62456/8/62456-Comparative%20Performance%20of%20Deep%20Learning_%20article.pdf
http://irep.iium.edu.my/62456/19/62456_Comparative%20performance%20of%20deep%20learning_WOS.pdf
Description
Summary:Imbalanced data is one of the challenges in a classification task in machine learning. Data disparity produces a biased output of a model regardless how recent the technology is. However, deep learning algorithms, such as deep belief networks showed promising results in many domains, especially in image processing. Therefore, in this paper, we will review the effect of imbalanced data disparity in classes using deep belief networks as the benchmark model and compare it with conventional machine learning algorithms, such as backpropagation neural networks, decision trees, naïve Bayes and support vector machine with MNIST handwritten dataset. The experiment shows that although the algorithm is stable and suitable for multiple domains, the imbalanced data distribution still manages to affect the outcome of the conventional machine learning algorithms.