Comparative performance of deep learning and machine learning algorithms on imbalanced handwritten data
Imbalanced data is one of the challenges in a classification task in machine learning. Data disparity produces a biased output of a model regardless how recent the technology is. However, deep learning algorithms, such as deep belief networks showed promising results in many domains, especially in...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English English English |
Published: |
The Science and Information (SAI) Organization
2018
|
Subjects: | |
Online Access: | http://irep.iium.edu.my/62456/ http://irep.iium.edu.my/62456/ http://irep.iium.edu.my/62456/7/62456-Comparative%20performance%20of%20deep%20learning%20_scopus.pdf http://irep.iium.edu.my/62456/8/62456-Comparative%20Performance%20of%20Deep%20Learning_%20article.pdf http://irep.iium.edu.my/62456/19/62456_Comparative%20performance%20of%20deep%20learning_WOS.pdf |
Summary: | Imbalanced data is one of the challenges in a classification task in machine learning. Data disparity produces a biased output of a model regardless how recent the technology is.
However, deep learning algorithms, such as deep belief networks showed promising results in many domains, especially in image processing. Therefore, in this paper, we will review the effect of imbalanced data disparity in classes using deep belief networks as the benchmark model and compare it with conventional machine learning algorithms, such as backpropagation neural networks, decision trees, naïve Bayes and support vector machine
with MNIST handwritten dataset. The experiment shows that
although the algorithm is stable and suitable for multiple domains, the imbalanced data distribution still manages to affect the outcome of the conventional machine learning algorithms. |
---|