PARALLEL IMPLEMENTATION OF EXPECTATION-MAXIMISATION ALGORITHM FOR THE TRAINING OF GAUSSIAN MIXTURE MODELS
- 1 , Brazil
Abstract
Most machine learning algorithms need to handle large data sets. This feature often leads to limitations on processing time and memory. The Expectation-Maximization (EM) is one of such algorithms, which is used to train one of the most commonly used parametric statistical models, the Gaussian Mixture Models (GMM). All steps of the algorithm are potentially parallelizable once they iterate over the entire data set. In this study, we propose a parallel implementation of EM for training GMM using CUDA. Experiments are performed with a UCI dataset and results show a speedup of 7 if compared to the sequential version. We have also carried out modifications to the code in order to provide better access to global memory and shared memory usage. We have achieved up to 56.4% of achieved occupancy, regardless the number of Gaussians considered in the set of experiments.
DOI: https://doi.org/10.3844/jcssp.2014.2124.2134
Copyright: © 2014 G. F. Araújo, H. T. Macedo, M. T. Chella, C. A.E. Montesco and M. V.O. Medeiros. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
- 3,192 Views
- 2,793 Downloads
- 5 Citations
Download
Keywords
- Expectation-Maximization (EM)
- Gaussian Mixture Models (GMM)
- CUDA