Fish Classification System Using YOLOv3-ResNet18 Model for Mobile Phones

Authors

  • Suryadiputra Liawatimena Bina Nusantara University
  • Edi Abdurachman Bina Nusantara University
  • Agung Trisetyarso Bina Nusantara University
  • Antoni Wibowo Bina Nusantara University
  • Muhamad Keenan Ario Bina Nusantara University
  • Ivan Sebastian Edbert Bina Nusantara University

DOI:

https://doi.org/10.21512/commit.v17i1.8107

Keywords:

Fish Classification System, YOLOv3- ResNet18 Model, Mobile Phone

Abstract

Every country in the world needs to report its fish production to the Food and Agriculture Organization of the United Nations (FAO) every year. In 2018, Indonesia ranked top five countries in fish production, with 8 million tons globally. Although it ranks top five, the fisheries in Indonesia are mostly dominated by traditional and small industries. Hence, a solution based on computer vision is needed to help detect and classify the fish caught every year. The research presents a method to detect and classify fish on mobile devices using the YOLOv3 model combined with ResNet18 as a backbone. For the experiment, the dataset used is four types of fish gathered from scraping across the Internet and taken from local markets and harbors with a total of 4,000 images. In comparison, two models are used: SSD-VGG and autogenerated model Huawei ExeML. The results show that the YOLOv3-ResNet18 model produces 98.45% accuracy in training and 98.15% in evaluation. The model is also tested on mobile devices and produces a speed of 2,115 ms on Huawei P40 and 3,571 ms on Realme 7. It can be concluded that the research presents a smaller-size model which is suitable for mobile devices while maintaining good accuracy and precision.

Dimensions

Plum Analytics

Author Biographies

Suryadiputra Liawatimena, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Doctor of Computer Science

Computer Science Department, BINUS Graduate Program – Master of Computer Science

Computer Engineering Department, Faculty of Engineering

Edi Abdurachman, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Doctor of Computer Science

Agung Trisetyarso, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Doctor of Computer Science

Antoni Wibowo, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Doctor of Computer Science

Muhamad Keenan Ario, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Master of Computer Science

Ivan Sebastian Edbert, Bina Nusantara University

Computer Science Department, BINUS Graduate Program – Master of Computer Science

References

Food and Agriculture Organization of the United Nations, The state of world fisheries and aquaculture 2020: Sustainability in action. Food and Agriculture Organization of the United Nations, 2020.

K. Kusdiantoro, A. Fahrudin, S. H. Wisudo, and B. Juanda, “Kinerja pembangunan perikanan tangkap di Indonesia,” Buletin Ilmiah Marina Sosial Ekonomi Kelautan dan Perikanan, vol. 5, no. 2, pp. 69–84, 2019.

J. Lee and W. Hwang, “Cloud-based facial expression recognition system for customer satisfaction in distribution sectors,” ICIC Express Letters, Part B: Applications, vol. 11, no. 2, pp. 173–179, 2020.

B. W. Yoon, E. Genc, O. F. Ince, and M. E. Yildirim, “Human activity recognition using inter-joint feature fusion with SVD,” ICIC Express Letters, Part B: Applications, vol. 12, no. 3, pp. 215–221, 2021.

R. Yamashita, M. Nishio, R. K. G. Do, and K. Togashi, “Convolutional neural networks: An overview and application in radiology,” Insights into Imaging, vol. 9, pp. 611–629, 2018.

D. I. Patr´ıcio and R. Rieder, “Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review,” Computers and Electronics in Agriculture, vol. 153, pp. 69–81, 2018.

Y. D. Zhang, Z. Dong, X. Chen, W. Jia, S. Du, K. Muhammad, and S. H. Wang, “Image based fruit category classification by 13-layer deep convolutional neural network and data augmentation,” Multimedia Tools and Applications, vol. 78,

pp. 3613–3632, 2019.

R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 580–587.

R. Girshick, “Fast R-CNN,” in IEEE International Conference on Computer Vision (ICCV), Seoul, Korea, Oct. 27–Nov. 2, 2015, pp. 1440–1448.

S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” in Advances in Neural Information Processing Systems, Montreal, Canada, Dec. 7–12, 2015, pp. 91–99.

W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu, and A. C. Berg, “SSD: Single shot multibox detector,” in Computer Vision–ECCV 2016. Amsterdam, The Netherlands,: Springer, Oct. 11–14, 2016, pp. 21–37.

J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, realtime object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779–788.

H. Yang, P. Liu, Y. Hu, and J. Fu, “Research on underwater object recognition based on YOLOv3,” Microsystem Technologies, vol. 27, pp. 1837–1844, 2021.

K. Cai, X. Miao, W. Wang, H. Pang, Y. Liu, and J. Song, “A modified YOLOv3 model for fish detection based on MobileNetv1 as backbone,” Aquacultural Engineering, vol. 91, pp. 1–9, 2020.

K. Raza and H. Song, “Fast and accurate fish detection design with improved YOLO-v3 model and transfer learning,” International Journal of Advanced Computer Science and Applications,

vol. 11, no. 2, pp. 7–16, 2020.

A. Wong, M. Famuori, M. J. Shafiee, F. Li, B. Chwyl, and J. Chung, “YOLO nano: A highly compact you only look once convolutional neural network for object detection,” in 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing-NeurIPS Edition (EMC2- NIPS). IEEE, 2019, pp. 22–25.

R. Huang, J. Pedoeem, and C. Chen, “YOLOLITE: A real-time object detection algorithm optimized for non-GPU computers,” in 2018 IEEE International Conference on Big Data (Big Data). Seattle, USA: IEEE, Dec. 10–13, 2018, pp. 2503–2510.

H. Zhao, Y. Zhou, L. Zhang, Y. Peng, X. Hu, H. Peng, and X. Cai, “Mixed YOLOv3-LITE: A lightweight real-time object detection method,” Sensors, vol. 20, no. 7, pp. 1–18, 2020.

P. Adarsh, P. Rathi, and M. Kumar, “YOLO v3- Tiny: Object detection and recognition using one stage improved model,” in 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS). Coimbatore, India: IEEE, March 6–7, 2020, pp. 687–694.

K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, United States, June 26–July 1, 2016, pp. 770–778.

W. Y. Ong, C. W. Too, and K. C. Khor, “Transfer learning on inception ResNet V2 for expiry reminder: A mobile application development,” in International Conference on Mobile Web and Intelligent Information Systems. Virtual (Online): Springer, Aug. 23–25, 2021, pp. 149–160.

Z. Wu, C. Shen, and A. Van Den Hengel, “Wider or deeper: Revisiting the ResNet model for visual recognition,” Pattern Recognition, vol. 90, pp. 119–133, 2019.

I. Martinez-Alpiste, G. Golcarenarenji, Q. Wang, and J. M. Alcaraz-Calero, “Smartphone-based real-time object recognition architecture for portable and constrained systems,” Journal of Real-Time Image Processing, vol. 19, no. 1, pp. 103–115, 2022.

I. Martinez-Alpiste, P. Casaseca-de-la Higuera, J. Alcaraz-Calero, C. Grecos, and Q. Wang, “Benchmarking machine-learning-based object detection on a uav and mobile platform,” in 2019 IEEE Wireless Communications and Networking Conference (WCNC). Marrakesh, Morocco: IEEE, April 15–18, 2019, pp. 1–6.

L. Tob´ıas, A. Ducournau, F. Rousseau, G. Mercier, and R. Fablet, “Convolutional neural networks for object recognition on mobile devices: A case study,” in 2016 23rd International Conference on Pattern Recognition (ICPR). Cancun, Mexico: IEEE, Dec. 4–8, 2016, pp. 3530–3535.

R. Kostoeva, R. Upadhyay, Y. Sapar, and A. Zakhor, “Indoor 3D interactive asset detection using a smartphone,” The International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 42, pp. 811–817, 2019.

J. Redmon and A. Farhadi, “YOLOv3: An incremental improvement,” 2018. [Online]. Available: https://arxiv.org/abs/1804.02767

——, “YOLO9000: Better, faster, stronger,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 7263–7271.

Y. Song, Q. K. Pan, L. Gao, and B. Zhang, “Improved non-maximum suppression for object detection using harmony search algorithm,” Applied Soft Computing, vol. 81, pp. 1–13, 2019.

G. Wu and Y. Li, “Non-maximum suppression for object detection based on the chaotic whale optimization algorithm,” Journal of Visual Communication and Image Representation, vol. 74, pp. 1–8, 2021.

J. Johnson, “CNN benchmarks,” 2017. [Online]. Available: https://github.com/jcjohnson/cnn-benchmarks#readme

P. D. Hung and N. N. Kien, “SSD-Mobilenet implementation for classifying fish species,” in Intelligent Computing and Optimization: Proceedings of the 2nd International Conference on Intelligent Computing and Optimization 2019 (ICO 2019). Springer, 2020, pp. 399–408.

HUAWEI CLOUD, “Introduction to ExeML,” 2023. [Online]. Available: https://support.huaweicloud.com/intl/en-us/exemlug-modelarts/modelarts 21 0001.html

Downloads

Published

2023-03-17
Abstract 414  .
PDF downloaded 541  .