Nudibranch Suborders Classification based on Densely Connected Convolutional Networks
Main Article Content
Abstract
Nudibranchs, often called sea slugs, are a group of soft-bodied marine gastropod mollusks that shed their shells after their larval stage. With their body structure that is very similar between one suborder and another, sometimes it is hard to tell apart the suborder of a nudibranch. In this work, we make an Image Classification model for determining the suborder of a nudibranch using deep learning algorithms DenseNet and EfficientNet. The experiment is conducted using Google Colaboratory environment. For DenseNet, we use 121, 169, and 201 layers; meanwhile, we only use the baseline algorithm for EfficientNet. The dataset for research is randomly taken from marine fauna forums on the internet. DenseNet with 201 layers shows a better generalization than other classifiers (accuracy of DenseNet 121, 169, 201, and baseline EfficientNet, respectively 53%, 41%, 73%, and 47%). The research produces a decent system for classifying the suborder of the Nudibranch. Usage of image recognition or background blurring systems in future research can improve the system's accuracy.
Downloads
Article Details
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Please find the rights and licenses in the Journal of Information Technology and Computer Engineering (JITCE).
1. License
The non-commercial use of the article will be governed by the Creative Commons Attribution license as currently displayed on Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
2. Author(s)’ Warranties
The author(s) warrants that the article is original, written by stated author(s), has not been published before, contains no unlawful statements, does not infringe the rights of others, is subject to copyright that is vested exclusively in the author and free of any third party rights, and that any necessary permissions to quote from other sources have been obtained by the author(s).
3. User Rights
JITCE adopts the spirit of open access and open science, which disseminates articles published as free as possible under the Creative Commons license. JITCE permits users to copy, distribute, display, and perform the work for non-commercial purposes only. Users will also need to attribute authors and JITCE on distributing works in the journal.
4. Rights of Authors
Authors retain the following rights:
- Copyright, and other proprietary rights relating to the article, such as patent rights,
- the right to use the substance of the article in future own works, including lectures and books,
- the right to reproduce the article for own purposes,
- the right to self-archive the article.
- the right to enter into separate, additional contractual arrangements for the non-exclusive distribution of the article's published version (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal (Journal of Information Technology and Computer Engineering).
5. Co-Authorship
If the article was jointly prepared by other authors; upon submitting the article, the author is agreed on this form and warrants that he/she has been authorized by all co-authors on their behalf, and agrees to inform his/her co-authors. JITCE will be freed on any disputes that will occur regarding this issue.
7. Royalties
By submitting the articles, the authors agreed that no fees are payable from JITCE.
8. Miscellaneous
JITCE will publish the article (or have it published) in the journal if the article’s editorial process is successfully completed and JITCE or its sublicensee has become obligated to have the article published. JITCE may adjust the article to a style of punctuation, spelling, capitalization, referencing and usage that it deems appropriate. The author acknowledges that the article may be published so that it will be publicly accessible and such access will be free of charge for the readers.
References
[2] B. Vishal and A. Deepak, “Current Status if Indian Opisthobranch Fauna,” Ecology and Conservation of Tropical Marine Faunal Communities, no. August, pp. 1–481, 2013, doi: 10.1007/978-3-642-38200-0.
[3] L. Hannon, “Nudibranchs.”
[4] F. Alharbi, A. Alharbi, and E. Kamioka, “Animal species classification using machine learning techniques,” MATEC Web of Conferences, vol. 277, p. 02033, 2019, doi: 10.1051/matecconf/201927702033.
[5] A. W. D. U. Shalika and L. Seneviratne, “Animal Classification System Based on Image Processing & Support Vector Machine,” Journal of Computer and Communications, vol. 04, no. 01, pp. 12–21, 2016, doi: 10.4236/jcc.2016.41002.
[6] M. M. Krishna, M. Neelima, M. Harshali, and M. V. G. Rao, “Image classification using Deep learning,” International Journal of Engineering and Technology(UAE), vol. 7, no. August, pp. 614–617, 2018, doi: 10.14419/ijet.v7i2.7.10892.
[7] Y. Hao, “Convolutional Neural Networks for Image Classification and Captioning,” Proceedings - 2021 2nd International Conference on Artificial Intelligence and Computer Engineering, ICAICE 2021, pp. 342–345, 2021, doi: 10.1109/ICAICE54393.2021.00073.
[8] K. O’Shea and R. Nash, “An Introduction to Convolutional Neural Networks,” pp. 1–11, 2015.
[9] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2016-Decem, pp. 770–778, 2016, doi: 10.1109/CVPR.2016.90.
[10] N. Sharma, V. Jain, and A. Mishra, “An Analysis of Convolutional Neural Networks for Image Classification,” Procedia Computer Science, vol. 132, no. Iccids, pp. 377–384, 2018, doi: 10.1016/j.procs.2018.05.198.
[11] D. Sarwinda, R. H. Paradisa, A. Bustamam, and P. Anggia, “Deep Learning in Image Classification using Residual Network ResNet) Variants for Detection of Colorectal Cancer.”
[12] X. Song, K. Chen, and Z. Cao, “ResNet-based Image Classification of Railway Shelling Defect,” Chinese Control Conference, CCC, vol. 2020-July, no. 3, pp. 6589–6593, 2020, doi: 10.23919/CCC50068.2020.9189112.
[13] M. Tan and Q. V. Le, “EfficientNet: Rethinking model scaling for convolutional neural networks,” 36th International Conference on Machine Learning, ICML 2019, vol. 2019-June, pp. 10691–10700, 2019.
[14] “SEASLUG.WORLD.”
[15] W. Jang and E. C. Lee, “Multi-class parrot image classification including subspecies with similar appearance,” Biology, vol. 10, no. 11, pp. 1–14, 2021, doi: 10.3390/biology10111140.
[16] “Nudibranchs (Order Nudibranchia) · iNaturalist.”
[17] G. Huang, Z. Liu, L. van der Maaten, and K. Q. Weinberger, “Densely Connected Convolutional Networks,” American Journal of Veterinary Research, vol. 39, no. 9, pp. 1442–1446, 2017.
[18] H. Gao, Z. Wang, L. Cai, and S. Ji, “ChannelNets: Compact and Efficient Convolutional Neural Networks via Channel-Wise Convolutions,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 8, pp. 2570–2581, 2021, doi: 10.1109/TPAMI.2020.2975796.
[19] M. Sandler, M. Zhu, A. Zhmoginov, and C. V Mar, “MobileNetV2: Inverted Residuals and Linear Bottlenecks”.
[20] J. Hu, L. Shen, S. Albanie, G. Sun, and E. Wu, “Squeeze-and-Excitation Networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 8, pp. 2011–2023, 2020, doi: 10.1109/TPAMI.2019.2913372.
[21] C. Sammut and H. I. Webb, “Encyclopedia of Machine Learning.”