Abstract
Insects are living beings whose utility is critical in life sciences. They enable biologists obtaining knowledge on natural landscapes (for example on their health). Nevertheless, insect identification is time-consuming and requires experienced workforce. To ease this task, we propose to turn it into an image-based pattern recognition problem by recognizing the insect from a photo. In this paper state-of-art deep convolutional architectures are used to tackle this problem. However, a limitation to the use of deep CNNs is the lack of data and the discrepancies in classes cardinality. To deal with such limitations, transfer learning is used to apply knowledge learnt from ImageNet-1000 recognition task to insect image recognition task. A question arises from transfer-learning: is it relevant to retrain the entire network or is it better not to modify some layers weights? The hypothesis behind this question is that there must be part of the network which contains generic (problem-independent) knowledge and the other one contains problem-specific knowledge. Tests have been conducted on two different insect image datasets. VGG-16 models were adapted to be more easily learnt. VGG-16 models were trained (a) from scratch (b) from ImageNet-1000. An advanced study was led on one of the datasets in which the influences on performance of two parameters were investigated: (1) The amount of learning data (2) The number of layers to be finetuned. It was determined VGG-16 last block is enough to be relearnt. We have made the code of our experiment as well as the script for generating an annotated insect dataset from ImageNet publicly available.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Al-Saqer, S.M., Hassan, G.M.: Artificial neural networks based red palm weevil (Rynchophorus Ferrugineous, Olivier) recognition system. Am. J. Agric. Biol. Sci 6, 356–364 (2011)
Bar, Y., Diamant, I., Wolf, L., Greenspan, H.: Deep learning with non-medical training used for chest pathology identification. In: Proceedings of SPIE, Medical Imaging: Computer-Aided Diagnosis, vol. 9414, 94140V–7 (2015)
Belharbi, S., et al.: Spotting L3 slice in CT scans using deep convolutional network and transfer learning. Comput. Biol. Med. 87, 95–103 (2017)
Bengio, Y., Boulanger-Lewandowski, N., Pascanu, R.: Advances in optimizing recurrent networks. CoRR, abs/1212.0901 (2012)
Chollet, F., et al.: Keras (2015). https://github.com/fchollet/keras
Choromanska, A., Henaff, M., Mathieu, M., Arous, G.B., LeCun, Y.: The loss surfaces of multilayer networks. In: Artificial Intelligence and Statistics, pp. 192–204 (2015)
Cireşan, D.C., Meier, U., Schmidhuber, J.: Transfer learning for Latin and Chinese characters with deep neural networks. In: The 2012 International Joint Conference on Neural Networks (IJCNN), pp. 1–6. IEEE (2012)
Dietrich, C.H., Pooley, C.D.: Automated identification of leafhoppers (Homoptera: Cicadellidae: Draeculacephala Ball). Ann. Entomol. Soc. Am. 87(4), 412–423 (1994)
Hafemann, L.G., Sabourin, R., Oliveira, L.S.: Writer-independent feature learning for offline signature verification using deep convolutional neural networks. CoRR, abs/1604.00974 (2016)
Krizhevsky, A., Sutskever, I., Hinton, G.: Imagenet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q. (eds.), NIPS, vol. 25, pp. 1097–1105 (2012)
Lai, M.: Deep learning for medical image segmentation. CoRR, abs/1505.02000 (2015)
Larios, N., et al.: Automated insect identification through concatenated histograms of local appearance features: feature vector generation and region detection for deformable objects. Mach. Vis. Appl. 19(2), 105–123 (2008)
Lin, M., Chen, Q., Yan, S.: Network in network. CoRR, abs/1312.4400 (2013)
Martineau, C., Conte, D., Raveaux, R., Arnault, I., Munier, D., Venturini, G.: A survey on image-based insect classification. Pattern Recognit. 65, 273–284 (2017)
Poznanski, A., Wolf, L.: CNN-N-gram for handwriting word recognition. In: CVPR, pp. 2305–2314 (2016)
Van Straalen, N.M.: Evaluation of bioindicator systems derived from soil arthropod communities. Appl. Soil Ecol. 9(1), 429–437 (1998)
Wang, J., Lin, C., Ji, L., Liang, A.: A new automatic identification system of insect images at the order level. Knowl. Based Syst. 33, 102–110 (2012)
Wen, C., Wu, D., Hu, H., Pan, W.: Pose estimation-dependent identification method for field moth images using deep learning architecture. Biosyst. Eng. 136, 117–128 (2015)
Xie, C., et al.: Automatic classification for field crop insects via multiple-task sparse representation and multiple-kernel learning. Comput. Electron. Agric. 119, 123–132 (2015)
Yosinski, J., Clune, J., Bengio, Y., Lipson, H.: How transferable are features in deep neural networks? In: Advances in Neural Information Processing Systems, pp. 3320–3328 (2014)
Yosinski, J., Clune, J., Nguyen, A.M., Fuchs, T.J., Lipson, H.: Understanding neural networks through deep visualization. CoRR, abs/1506.06579 (2015)
Acknowledgments
This work was supported by a research grant from the Région Centre-Val de Loire, France.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Martineau, C., Raveaux, R., Chatelain, C., Conte, D., Venturini, G. (2018). Effective Training of Convolutional Neural Networks for Insect Image Recognition. In: Blanc-Talon, J., Helbert, D., Philips, W., Popescu, D., Scheunders, P. (eds) Advanced Concepts for Intelligent Vision Systems. ACIVS 2018. Lecture Notes in Computer Science(), vol 11182. Springer, Cham. https://doi.org/10.1007/978-3-030-01449-0_36
Download citation
DOI: https://doi.org/10.1007/978-3-030-01449-0_36
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-01448-3
Online ISBN: 978-3-030-01449-0
eBook Packages: Computer ScienceComputer Science (R0)