@article {10.3844/jcssp.2014.2124.2134, article_type = {journal}, title = {PARALLEL IMPLEMENTATION OF EXPECTATION-MAXIMISATION ALGORITHM FOR THE TRAINING OF GAUSSIAN MIXTURE MODELS}, author = {Araújo, G. F. and Macedo, H. T. and Chella, M. T. and Montesco, C. A.E. and Medeiros, M. V.O.}, volume = {10}, number = {10}, year = {2014}, month = {Jul}, pages = {2124-2134}, doi = {10.3844/jcssp.2014.2124.2134}, url = {https://thescipub.com/abstract/jcssp.2014.2124.2134}, abstract = {Most machine learning algorithms need to handle large data sets. This feature often leads to limitations on processing time and memory. The Expectation-Maximization (EM) is one of such algorithms, which is used to train one of the most commonly used parametric statistical models, the Gaussian Mixture Models (GMM). All steps of the algorithm are potentially parallelizable once they iterate over the entire data set. In this study, we propose a parallel implementation of EM for training GMM using CUDA. Experiments are performed with a UCI dataset and results show a speedup of 7 if compared to the sequential version. We have also carried out modifications to the code in order to provide better access to global memory and shared memory usage. We have achieved up to 56.4% of achieved occupancy, regardless the number of Gaussians considered in the set of experiments.}, journal = {Journal of Computer Science}, publisher = {Science Publications} }