Recent Advances in Electrical & Electronic Engineering

Recent Advances in Electrical & Electronic Engineering

Editor-in-Chief

ISSN (Print): 2352-0965
ISSN (Online): 2352-0973

Back Subscribe
Research Article

Comparative Analysis of CNN Performances Using CIFAR-100 and MNIST Databases: GPU vs. CPU Efficiency

Author(s): Rania Boukhenounorcid of author, Hakim Doghmaneorcid of author, Kamel Messaoudi*orcid of author and El-Bay Bourennaneauthors OrcID

Volume 18, Issue 10, 2025

Published on: 16 June, 2025

Article ID: e23520965348453 Pages: 13

DOI: 10.2174/0123520965348453250226080233

Price: $65

Become a Editorial Board Member
Become a Reviewer
Become a Editor
Become a Section Editor

Abstract

Introduction: Convolutional Neural Networks (CNNs) and deep learning algorithms have significantly advanced image processing and classification. This study compares three CNN architectures VGG-16, ResNet-50, and ResNet-18, and evaluates their performance on the CIFAR- 100 and MNIST datasets. Training time is prioritized as a critical metric, along with test accuracy and training loss, under varying hardware configurations (NVIDIA-GPU and Intel-CPU).

Methods: Experiments were conducted on an Ubuntu 22.04 system using the PyTorch framework. The hardware configurations included an NVIDIA GeForce GTX 1660 Super GPU and an Intel Core i5-10400 CPU. The CIFAR-100 dataset, containing 60,000 color images across 100 classes, and the MNIST dataset, comprising 70,000 grayscale images, were used for benchmarking.

Results: The results highlight the superior efficiency of GPUs, with training times reduced by up to 10x compared to CPUs. For CIFAR-100, VGG-16 required 13,000 seconds on the GPU versus 130,000 seconds on the CPU, while ResNet-18, the most time-efficient model, completed training in 150 seconds on the GPU and 1,740 seconds on the CPU. ResNet-50 achieved the highest test accuracy (~80%) on CIFAR-100. On MNIST, ResNet-18 was the most efficient, with training times of 185 seconds on the GPU and 22,000 seconds on the CPU.

Discussion: This study highlights the clear advantage of GPUs over CPUs in reducing training times, particularly for complex models such as VGG-16 and ResNet-50. ResNet-50 achieved the highest accuracy, while ResNet-18 was the most time efficient. However, the use of simpler datasets (MNIST and CIFAR-100) may not fully capture the complexity of the real world.

Conclusion: This study emphasizes the importance of hardware selection for deep learning workflows. Using the CIFAR-100 and MNIST datasets, we demonstrated that GPUs significantly outperform CPUs, achieving up to a 10x reduction in training time while maintaining competitive accuracy. Among the architectures tested, ResNet-50 delivered the highest test accuracy (~80%) on CIFAR-100, demonstrating superior feature extraction capabilities compared to VGG-16 and ResNet-18. Meanwhile, ResNet-18 proved to be the most time-efficient architecture, completing training in 150 seconds on the GPU, a significant improvement over VGG-16's 13,000 seconds. These results highlight the advantage of residual connections in reducing training complexity and achieving higher performance. The results underscore the critical role of both architecture selection and hardware optimization in advancing deep learning workflows.

Keywords: CNN architecture, deep learning, image classification, training time, gpu-nvidia, CPU, GPU.

Graphical Abstract

[1]
K. Simonyan, and A. Zisserman, "Very deep convolutional networks for large-scale image recognition", arXiv preprint arXiv:1409.1556., pp. 1-6, 2014.
[2]
K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition", In Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR) 2016+, pp. 770-778
[3]
R. Elshawi, A. Wahab, A. Barnawi, and S. Sakr, "DLBench: A comprehensive experimental evaluation of deep learning frameworks", Cluster Comput., vol. 24, no. 3, pp. 2017-2038, 2021.
[http://dx.doi.org/10.1007/s10586-021-03240-4]
[4]
A. Riahi, A. Savadi, and M. Naghibzadeh, "Comparison of analytical and ML-based models for predicting CPU–GPU data transfer time", Computing, vol. 102, no. 9, pp. 2099-2116, 2020.
[http://dx.doi.org/10.1007/s00607-019-00780-x]
[5]
A. Jayasimhan, and P. Pabitha, "A comparison between CPU and GPU for image classification using convolutional neural networks", In International Conference on Communication, Computing and Internet of Things (IC3IoT) IEEE, 2022. pp. 1-4.
[http://dx.doi.org/10.1109/IC3IOT53935.2022.9767990]
[6]
A. Todi, N. Narula, M. Sharma, and U. Gupta, "ConvNext: A contemporary architecture for convolutional neural networks for image classification", In Proc. 2023 Int. Conf. Innovative Sustainable Comput. Technol. (CISCT) 2023, PP. 1-6.
[http://dx.doi.org/10.1109/CISCT57197.2023.10351320]
[7]
R. Azad, "Advances in medical image analysis with vision transformers: A comprehensive review", Med. Image Ana., vol. 91, p. 103000, 2023.
[8]
A. Mondal, and P. Saha, "Classification of brain tumor from magnetic resonance imaging using vision transformers ensembling,” Biomed. Signal Process. Control", Online, vol. 74, p. 103486, 2022.
[http://dx.doi.org/10.1016/j.bspc.2021.103486]
[9]
S.K. Roy, "Spectral–spatial morphological attention transformer for hyperspectral image classification", ., vol. 61, IEEE Trans. Geosci. Remote Sens., pp. 1-15, 2023.
[http://dx.doi.org/10.1109/TGRS.2023.3242346]
[10]
H. Rasheed, "Efficient CNN architectures for edge computing environments", Open Comp. Sci. J., vol. 15, no. 2, pp. 201-215, 2023.
[11]
T.A. Bukhari, "Review of lightweight deep learning architectures for real-time object detection", Cur. Arti. Inte. Res., vol. 18, no. 3, pp. 219-234, 2022.
[12]
A. Krizhevsky, and G. Hinton, "Learning multiple layers of features from tiny images",
[13]
Z. Liu, "ConvNext: Revisiting ResNets for scalability", In Proc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 2022, pp. 1-9.
[14]
T. Touvron, "Going deeper with image transformers", In Proc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 2023, pp. 1-10.
[15]
Y. LeCun, The MNIST database of handwritten digits., 1998. Available from: http://yann.lecun.com/exdb/mnist/
[16]
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, "Gradient-based learning applied to document recognition", Proc. IEEE, vol. 86, no. 11, pp. 2278-2324, 1998.
[http://dx.doi.org/10.1109/5.726791]
[17]
D. Ciregan, U. Meier, and J. Schmidhuber, "Multi-column deep neural networks for image classification", In Proc. 2012 IEEE Conf. Computer Vision and Pattern Recognition. (CVPR), IEEE, 2012.
[http://dx.doi.org/10.1109/CVPR.2012.6248110]
[18]
Y. Lecun, Handwritten digit recognition with a back-propagation network., Advances in Neural Information Processing Systems (NeurIPS): New Orleans, 1990, pp. 396-404.
[19]
A. Krizhevsky, I. Sutskever, and G.E. Hinton, "Imagenet classification with deep convolutional neural networks", In: Adv. Neural Inf. Process. Syst., vol. 25. NIPS, 2012, no. 6, pp. 84-90.
[20]
A. Alex, "Exploring the limitations of deep learning for robust image recognition", In Proc. IEEE Int. Conf. Comput. Vision (ICCV). 2012, pp. 99–106.
[21]
C. Szegedy, "Going deeper with convolutions", In Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR) 2015, pp. 1–9.
[22]
F. Chollet, "Xception: Deep learning with depthwise separable convolutions", In Proc. IEEE Conf. Comput. Vision and Pattern Recognit. 2017, pp. 1251-1258.
[http://dx.doi.org/10.1109/CVPR.2017.195]
[23]
G. Huang, Z. Liu, L. van der Maaten, and K.Q. Weinberger, "Densely connected convolutional networks", In Proc. IEEE Conf. Comput. Vision and Pattern Recognit. 2017, pp. 4700–4708.
[http://dx.doi.org/10.1109/CVPR.2017.243]
[24]
A.G. Howard, "MobileNets: Efficient convolutional neural networks for mobile vision applications", arXiv:1704.04861, pp. 1-6, 2017.
[25]
M. Tan, and Q. Le, "EfficientNet: Rethinking model scaling for convolutional neural networks", In Proc. Int. Conf. Mach. Learn. 2019, pp. 6105-6114.
[26]
H. Touvron, "Training data-efficient image transformers and distillation through attention", In Proc. Int. Conf. Mach. Learn. 2021, pp. 10347-10357.
[27]
NVIDIA, GeForce GTX 1660 Super, 2024. Available from: http://www.nvidia.com/en-us/geforce/graphics-cards/gtx-1660-super/
[28]
"NVIDIA GeForce GTX 1660 Super", TechPowerUp, 2019. Available from: www.techpowerup.com/gpu-specs/geforce-gtx-1660-super.c3450
[29]
A. Krizhevsky, "GPU acceleration for deep learning: Analysis and insights", Adv. Mach. Lear., vol. 1, pp. 101-118, 2022.
[30]
"NVIDIA GeForce GTX 1660 Super Review", Tom's Hardware, 2019. Available from: www.tomshardware.com/reviews/nvidia-geforce-gtx-1660-super-review/
[31]
G. Sharma, "Comparative evaluation of GPUs for machine learning workloads", In Proc. Int. Conf. Comput. Arch. 2023, pp. 123-136.
[32]
Intel, "Intel Core i5-10400 Processor", Intel Official Website., 2024. Available from: www.intel.com/content/www/us/en/products/processors/core/i5-processors/i5-10400.html
[33]
"Intel Core i5-10400 Review", TechSpot, 2020. Available from: www.techspot.com/review/2011-intel-core-i5-10400/
[34]
"Intel Core i5-10400: Everything You Need to Know", Tom's Hardware, 2020. Available from: www.tomshardware.com/review/intel-core-i5-10400-review/
[35]
C.M. Bishop, and N.M. Nasrabadi, Pattern Recognition and Machine Learning., Springer: New York, 2006, no. 4, p. 738.4.
[36]
J. Brownlee, "Deep Learning for Computer Vision: Image Classification, Object Detection, and Face Recognition in Python", Mach. Lear. Mast.y, vol. 1, pp. 1-7, 2019.
[37]
A. Oliver, "Realistic evaluation of deep semi-supervised learning algorithms", Proc. NeurIPS, vol. 31, pp. 3235-3246, 2018.
[38]
R. Kohavi, "A study of cross-validation and bootstrap for accuracy estimation and model selection", In Proc. Int. Joint Conf. Artif. Intel., vol. 14, 1995, pp. 1137-1143
[39]
N. Srivastava, "Dropout: A simple way to prevent neural networks from overfitting", J. Mach. Learn. Res., vol. 15, pp. 1929-1958, 2014.
[40]
P. Gayathri, "Exploring the potential of VGG-16 architecture for accurate brain tumor detection using deep learning", J. Comput., Mech. Manag., vol. 2, no. 2, pp. 1-7, 2023.
[http://dx.doi.org/10.57159/gadl.jcmm.2.2.23056]
[41]
T. Singh, "Early Stage Lung Cancer Detection Using Deep Learning", In MIT Art, Design and Technology School of Computing International Conference (MITADTSoCiCon). 2024, pp. 1–6.
[http://dx.doi.org/10.1109/MITADTSoCiCon60330.2024.10575345]
[42]
Z. Wang, "Edge AI: On-demand accelerating deep neural network inference via edge computing", IEEE Trans. Neural Netw. Learn. Syst., vol. 31, no. 9, pp. 3305-3316, 2020.
[http://dx.doi.org/10.1109/TNNLS.2020.2966053] [PMID: 31613785]
[43]
N.P. Jouppi, In-datacenter Performance Analysis of a Tensor Processing Unit., Proc. 44th Annu. Int. Symp. Comput. Archit.: Toronto, ON, Canada, 2017, pp. 1-12.
[http://dx.doi.org/10.1145/3079856.3080246]
[44]
O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A.C. Berg, and L. Fei-Fei, "ImageNet large scale visual recognition challenge", Int. J. Comput. Vis., vol. 115, no. 3, pp. 211-252, 2015.
[http://dx.doi.org/10.1007/s11263-015-0816-y]
[45]
T. Lin, "Microsoft COCO: Common objects in context", In Proc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR). 2014, pp. 748–755.
[http://dx.doi.org/10.1007/978-3-319-10602-1_48]
[46]
P.R. Srinivasan, "MRI scan classification using deep learning", J. Med. Imag., vol. 29, no. 2, pp. 1-9, 2018.
[http://dx.doi.org/10.1117/1.JMI.6.2.025001]
[47]
B.V.K. Vijaya, "Hyperspectral image classification with deep learning techniques", In Proc. IEEE Int. Conf. Geosci. Remote Sens. (IGARSS). 2019, pp. 5313–5316.
[http://dx.doi.org/10.1109/IGARSS.2019.8898119]
[48]
A. Dosovitskiy, "An image is worth 16x16 words: Transformers for image recognition at scale", arXiv preprint, arXiv:2010.11929, vol. 1, pp. 1-6, 2020.
[49]
S. Ioffe, and C. Szegedy, "Batch normalization: Accelerating deep network training by reducing internal covariate shift", In Proc. Int. Conf. Mach. Learn. 2015, pp. 448-456.

Rights & Permissions Print Cite