@article {10.3844/jcssp.2019.785.799, article_type = {journal}, title = {Performance Analysis of Deep Learning Libraries: TensorFlow and PyTorch}, author = {Florencio, Felipe and Valenç, Thiago and Moreno, Edward David and Junior, Methanias Colaço}, volume = {15}, number = {6}, year = {2019}, month = {Apr}, pages = {785-799}, doi = {10.3844/jcssp.2019.785.799}, url = {https://thescipub.com/abstract/jcssp.2019.785.799}, abstract = {Through the increase in deep learning study and use, in the last years there was a development of specific libraries for Deep Neural Network (DNN). Each one of these libraries has different performance results and applies different techniques to optimize the implementation of algorithms. Therefore, even though implementing the same algorithm and using different libraries, the performance of their executions may have a considerable variation. For this reason, developers and scientists that work with deep learning need scientific experimental studies that examine the performance of those libraries. Therefore, this paper has the aim of evaluating and comparing these two libraries: TensorFlow and PyTorch. We have used three parameters: Hardware utilization, hardware temperature and execution time in a context of heterogeneous platforms with CPU and GPU. We used the MNIST database for training and testing the LeNet Convolutional Neural Network (CNN). We performed a scientific experiment following the Goal Question Metrics (GQM) methodology, data were validated through statistical tests. After data analysis, we show that PyTorch library presented a better performance, even though the TensorFlow library presented a greater GPU utilization rate.}, journal = {Journal of Computer Science}, publisher = {Science Publications} }