Fish Classification System Using YOLOv3-ResNet18 Model for Mobile Phones
DOI:
https://doi.org/10.21512/commit.v17i1.8107Keywords:
Fish Classification System, YOLOv3- ResNet18 Model, Mobile PhoneAbstract
Every country in the world needs to report its fish production to the Food and Agriculture Organization of the United Nations (FAO) every year. In 2018, Indonesia ranked top five countries in fish production, with 8 million tons globally. Although it ranks top five, the fisheries in Indonesia are mostly dominated by traditional and small industries. Hence, a solution based on computer vision is needed to help detect and classify the fish caught every year. The research presents a method to detect and classify fish on mobile devices using the YOLOv3 model combined with ResNet18 as a backbone. For the experiment, the dataset used is four types of fish gathered from scraping across the Internet and taken from local markets and harbors with a total of 4,000 images. In comparison, two models are used: SSD-VGG and autogenerated model Huawei ExeML. The results show that the YOLOv3-ResNet18 model produces 98.45% accuracy in training and 98.15% in evaluation. The model is also tested on mobile devices and produces a speed of 2,115 ms on Huawei P40 and 3,571 ms on Realme 7. It can be concluded that the research presents a smaller-size model which is suitable for mobile devices while maintaining good accuracy and precision.
Plum Analytics
References
Food and Agriculture Organization of the United Nations, The state of world fisheries and aquaculture 2020: Sustainability in action. Food and Agriculture Organization of the United Nations, 2020.
K. Kusdiantoro, A. Fahrudin, S. H. Wisudo, and B. Juanda, “Kinerja pembangunan perikanan tangkap di Indonesia,” Buletin Ilmiah Marina Sosial Ekonomi Kelautan dan Perikanan, vol. 5, no. 2, pp. 69–84, 2019.
J. Lee and W. Hwang, “Cloud-based facial expression recognition system for customer satisfaction in distribution sectors,” ICIC Express Letters, Part B: Applications, vol. 11, no. 2, pp. 173–179, 2020.
B. W. Yoon, E. Genc, O. F. Ince, and M. E. Yildirim, “Human activity recognition using inter-joint feature fusion with SVD,” ICIC Express Letters, Part B: Applications, vol. 12, no. 3, pp. 215–221, 2021.
R. Yamashita, M. Nishio, R. K. G. Do, and K. Togashi, “Convolutional neural networks: An overview and application in radiology,” Insights into Imaging, vol. 9, pp. 611–629, 2018.
D. I. Patr´ıcio and R. Rieder, “Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review,” Computers and Electronics in Agriculture, vol. 153, pp. 69–81, 2018.
Y. D. Zhang, Z. Dong, X. Chen, W. Jia, S. Du, K. Muhammad, and S. H. Wang, “Image based fruit category classification by 13-layer deep convolutional neural network and data augmentation,” Multimedia Tools and Applications, vol. 78,
pp. 3613–3632, 2019.
R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 580–587.
R. Girshick, “Fast R-CNN,” in IEEE International Conference on Computer Vision (ICCV), Seoul, Korea, Oct. 27–Nov. 2, 2015, pp. 1440–1448.
S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” in Advances in Neural Information Processing Systems, Montreal, Canada, Dec. 7–12, 2015, pp. 91–99.
W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu, and A. C. Berg, “SSD: Single shot multibox detector,” in Computer Vision–ECCV 2016. Amsterdam, The Netherlands,: Springer, Oct. 11–14, 2016, pp. 21–37.
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, realtime object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779–788.
H. Yang, P. Liu, Y. Hu, and J. Fu, “Research on underwater object recognition based on YOLOv3,” Microsystem Technologies, vol. 27, pp. 1837–1844, 2021.
K. Cai, X. Miao, W. Wang, H. Pang, Y. Liu, and J. Song, “A modified YOLOv3 model for fish detection based on MobileNetv1 as backbone,” Aquacultural Engineering, vol. 91, pp. 1–9, 2020.
K. Raza and H. Song, “Fast and accurate fish detection design with improved YOLO-v3 model and transfer learning,” International Journal of Advanced Computer Science and Applications,
vol. 11, no. 2, pp. 7–16, 2020.
A. Wong, M. Famuori, M. J. Shafiee, F. Li, B. Chwyl, and J. Chung, “YOLO nano: A highly compact you only look once convolutional neural network for object detection,” in 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing-NeurIPS Edition (EMC2- NIPS). IEEE, 2019, pp. 22–25.
R. Huang, J. Pedoeem, and C. Chen, “YOLOLITE: A real-time object detection algorithm optimized for non-GPU computers,” in 2018 IEEE International Conference on Big Data (Big Data). Seattle, USA: IEEE, Dec. 10–13, 2018, pp. 2503–2510.
H. Zhao, Y. Zhou, L. Zhang, Y. Peng, X. Hu, H. Peng, and X. Cai, “Mixed YOLOv3-LITE: A lightweight real-time object detection method,” Sensors, vol. 20, no. 7, pp. 1–18, 2020.
P. Adarsh, P. Rathi, and M. Kumar, “YOLO v3- Tiny: Object detection and recognition using one stage improved model,” in 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS). Coimbatore, India: IEEE, March 6–7, 2020, pp. 687–694.
K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, United States, June 26–July 1, 2016, pp. 770–778.
W. Y. Ong, C. W. Too, and K. C. Khor, “Transfer learning on inception ResNet V2 for expiry reminder: A mobile application development,” in International Conference on Mobile Web and Intelligent Information Systems. Virtual (Online): Springer, Aug. 23–25, 2021, pp. 149–160.
Z. Wu, C. Shen, and A. Van Den Hengel, “Wider or deeper: Revisiting the ResNet model for visual recognition,” Pattern Recognition, vol. 90, pp. 119–133, 2019.
I. Martinez-Alpiste, G. Golcarenarenji, Q. Wang, and J. M. Alcaraz-Calero, “Smartphone-based real-time object recognition architecture for portable and constrained systems,” Journal of Real-Time Image Processing, vol. 19, no. 1, pp. 103–115, 2022.
I. Martinez-Alpiste, P. Casaseca-de-la Higuera, J. Alcaraz-Calero, C. Grecos, and Q. Wang, “Benchmarking machine-learning-based object detection on a uav and mobile platform,” in 2019 IEEE Wireless Communications and Networking Conference (WCNC). Marrakesh, Morocco: IEEE, April 15–18, 2019, pp. 1–6.
L. Tob´ıas, A. Ducournau, F. Rousseau, G. Mercier, and R. Fablet, “Convolutional neural networks for object recognition on mobile devices: A case study,” in 2016 23rd International Conference on Pattern Recognition (ICPR). Cancun, Mexico: IEEE, Dec. 4–8, 2016, pp. 3530–3535.
R. Kostoeva, R. Upadhyay, Y. Sapar, and A. Zakhor, “Indoor 3D interactive asset detection using a smartphone,” The International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 42, pp. 811–817, 2019.
J. Redmon and A. Farhadi, “YOLOv3: An incremental improvement,” 2018. [Online]. Available: https://arxiv.org/abs/1804.02767
——, “YOLO9000: Better, faster, stronger,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 7263–7271.
Y. Song, Q. K. Pan, L. Gao, and B. Zhang, “Improved non-maximum suppression for object detection using harmony search algorithm,” Applied Soft Computing, vol. 81, pp. 1–13, 2019.
G. Wu and Y. Li, “Non-maximum suppression for object detection based on the chaotic whale optimization algorithm,” Journal of Visual Communication and Image Representation, vol. 74, pp. 1–8, 2021.
J. Johnson, “CNN benchmarks,” 2017. [Online]. Available: https://github.com/jcjohnson/cnn-benchmarks#readme
P. D. Hung and N. N. Kien, “SSD-Mobilenet implementation for classifying fish species,” in Intelligent Computing and Optimization: Proceedings of the 2nd International Conference on Intelligent Computing and Optimization 2019 (ICO 2019). Springer, 2020, pp. 399–408.
HUAWEI CLOUD, “Introduction to ExeML,” 2023. [Online]. Available: https://support.huaweicloud.com/intl/en-us/exemlug-modelarts/modelarts 21 0001.html
Downloads
Published
Issue
Section
License
Copyright (c) 2023 Suryadiputra Liawatimena, Edi Abdurachman, Agung Trisetyarso, Antoni Wibowo, Muhamad Keenan Ario, Ivan Sebastian Edbert
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms:
a. Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License - Share Alike that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this journal.
b. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal.
c. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work.
USER RIGHTS
All articles published Open Access will be immediately and permanently free for everyone to read and download. We are continuously working with our author communities to select the best choice of license options, currently being defined for this journal as follows: Creative Commons Attribution-Share Alike (CC BY-SA)