
Performance analysis and comparison of cat and dog image classification based on different models
- 1 Xi’an Jiaotong-Liverpool University
- 2 Xi’an Jiaotong-Liverpool University
* Author to whom correspondence should be addressed.
Abstract
Image classification has widespread applications in computer vision, with significant advancements in performance due to deep learning models. Cat and dog image classification, as a classic problem, has attracted considerable research interest. This study aims to conduct a comprehensive analysis and comparison of deep learning models, including LeNet, ResNet, and VGG, in the context of cat and dog image classification. This paper employed two datasets: traditional cat and dog images and non-traditional, diverse images. Data preprocessing and augmentation were applied, and various model architectures were constructed. Through training and testing, this paper assessed the performance of these models under different conditions. The research findings indicate that ResNet excels in handling various datasets and different dataset sizes, demonstrating outstanding image classification performance. LeNet performs well on traditional datasets but experiences performance degradation when dealing with non-traditional datasets and smaller dataset sizes. VGG performs reasonably well on the original dataset but needs help processing non-traditional datasets. These results provide valuable insights for guiding model selection and optimization in image classification tasks.
Keywords
VGG, ResNet, LeNet, Image Classification
[1]. LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324.
[2]. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778).
[3]. Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
[4]. Szytula A and Leciejewicz J 1989 Handbook on the Physics and Chemistry of Rare Earths vol 12, ed K A Gschneidner Jr and L Erwin (Amsterdam: Elsevier) p 133. O’Shea, K., & Nash, R. (2015). An introduction to convolutional neural networks. arXiv preprint arXiv:1511.08458.
[5]. Agrawal, S. C., Tripathi, R. K., Bhardwaj, N., & Parashar, P. (2023, July). Virtual Drawing: An Air Paint Application. In 2023 2nd International Conference on Edge Computing and Applications (ICECAA) (pp. 971-975). IEEE.
[6]. Yasin, E. T., Ozkan, I. A., & Koklu, M. (2023). Detection of fish freshness using artificial intelligence methods. European Food Research and Technology, 1-12.
[7]. Cai, D. (2017). Fine-grained classification of low-resolution image (Master’s thesis).
[8]. Howard, J., & Gugger, S. (2020). Deep Learning for Coders with fastai and PyTorch. O’Reilly Media.
[9]. Alomar, K., Aysel, H. I., & Cai, X. (2023). Data augmentation in classification and segmentation: A survey and new strategies. Journal of Imaging, 9(2), 46.
[10]. Maurya, L., Lohchab, V., Mahapatra, P. K., & Abonyi, J. (2022). Contrast and brightness balance in image enhancement using Cuckoo Search-optimized image fusion. Journal of King Saud University-Computer and Information Sciences, 34(9), 7247-7258.
Cite this article
Ma,D.;Song,H. (2024). Performance analysis and comparison of cat and dog image classification based on different models. Applied and Computational Engineering,41,197-201.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 2023 International Conference on Machine Learning and Automation
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).