Abstract
Precision Agriculture aims to apply selective treatments and tasks at localized areas concerning crop fields. Robotized and autonomous tractors, equipped with perception, decision-making and actuation systems, can apply specific treatments as may be required. Correct plant identification through the perception system, including crops and weeds, is an important issue. Additionally, it is well known that, in autonomous vehicles, safety is a major challenge, where unexpected obstacles in the working area must be conveniently addressed in order to guarantee the security and the continuity of the process. The objective of this study was to design a tri-class Support Vector Machine classifier for identifying plants (crops and weeds), soil and objects in maize fields based on unsupervised learning. For this, a strategy for automatic sample selection was designed to obtain elements of the three involved classes for the training process. In this context, the identification of obstacles for safe navigation makes an important contribution. A comparative analysis of different texture descriptors and local patterns was carried out with the aim of determining the best for characterizing the classes under study; results have shown that the Speeded-Up Robust Features descriptor is the most appropriate to discriminate between plants, soil and objects. The development of an object detection algorithm for agricultural images proved the effectiveness of the tri-class classifier with an accuracy of 94.3%.
Similar content being viewed by others
References
Andrzej, M., & Strzelecki, M. (1998). Texture analysis methods–A review. In Technical university of lodz, institute of electronics, COST B11 report, Brussels, Belgium, pp. 9–11.
Bai, X. D., Cao, Z. G., Wang, Y., Yu, Z. H., Zhang, X. F., & Li, C. N. (2013). Crop segmentation from images by morphology modeling in the CIE L*a*b* color space. Computers and Electronics in Agriculture, 99, 21–34.
Bay, H., Ess, A., Tuytelaars, T., & Van, G. L. (2008). Speeded-up robust features (SURF). Computer Vision and Image Understanding, 10(3), 346–359.
Boor, C. de. (1978). A practical guide to splines (Vol. 27). New York, USA: Springer.
Chang, C. C., & Lin, C. J. (2011). LIBSVM: A library for support vector machines. In ACM Transactions on Intelligent Systems and Technology (Vol. 2, pp. 27:1–27:27). Retrieved November 5, 2016, from http://www.csie.ntu.edu.tw/~cjlin/libsvm.
Dillencourt, M. B., Samet, H., & Tamminen, M. (1992). A general approach to connected-component labeling for arbitrary image representations. Journal of the ACM (JACM), 39(2), 253–280.
Duda, R. O., Hart, P. E., & Stork, D. G. (2000). Pattern classification (2nd ed.). New York, USA: Wiley.
Emmi, L., Gonzalez-de-Soto, M., Pajares, G., & Gonzalez-de-Santos, P. (2014). Integrating sensory/actuation systems in agricultural vehicles. Sensors, 14(3), 4014–4049.
Friedman, J. H. (1996). Another approach to polychotomous classification. Technical report, Department of Statistics, Stanford University, California, USA.
Garrido, M., Perez-Ruiz, M., Valero, C., Gliever, C. J., Hanson, B. D., & Slaughter, D. C. (2014). Active optical sensors for tree stem detection and classification in nurseries. Sensors, 14(6), 10783–10803.
Gonzalez, R. C., & Woods, R. E. (2006). Digital image processing (3rd ed.). Upper Saddle River, NJ, USA: Prentice-Hall Inc.
Guerrero, J. M., Pajares, G., Montalvo, M., Romeo, J., & Guijarro, M. (2012). Support vector machines for crop/weeds identification in maize fields. Expert Systems with Applications, 39(12), 11149–11155.
Guijarro, M., Pajares, G., Riomoros, I., Herrera, P. J., Burgos-Artizzu, X. P., & Ribeiro, A. (2011). Automatic segmentation of relevant textures in agricultural images. Computers and Electronics in Agriculture, 75(1), 75–83.
Guijarro, M., Riomoros, I., Pajares, G., & Zitinski, P. (2015). Discrete wavelets transform for improving greenness image segmentation in agricultural images. Computers and Electronics in Agriculture, 118, 396–407.
Haralick, R. M., & Karthikeyan, S. (1973). Textural features for image classification. IEEE Transactions on Systems, Man, and Cybernetics, 3(6), 610–621.
Haug, S., Michaels, A., Biber, P., & Ostermann, J. (2014) Plant classification system for crop/weed discrimination without segmentation. In IEEE winter conference on applications of computer vision (pp. 1142–1149).
Hlaing, S. H., & Khaing, A. S. (2014). Weed and crop segmentation and classification using area thresholding. International Journal of Research in Engineering and Technology., 3, 375–382.
Hossein, M. (2013). A technical review on navigation systems of agricultural autonomous off-road vehicles. Journal of Terramechanics, 50(3), 211–232.
Hsu, C. W., & Lin, C. J. (2002). A comparison of methods for multiclass support vector machines. IEEE Transactions on Neural Networks, 13(2), 415–425.
Ionescu, R. T., Popescu, A. L., Popescu, M., & Popescu, D. (2015). BiomassID: A biomass type identification system for mobile devices. Computers and Electronics in Agriculture, 113, 244–253.
Jiang, G., Wang, Z., & Liu, H. (2015). Automatic detection of crop rows based on multi-ROIs. Expert Systems with Applications, 42(5), 2429–2441.
Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60(2), 91–110.
Montalvo, M., Guijarro, M., Guerrero, J.M., & Ribeiro, Á., (2016). Identification of Plant Textures in Agricultural Images by Principal Component Analysis. In Proceedings of the 11th international conference HAIS 2016, Seville, Spain, pp. 391–401.
Moorthy, S., Boigelot, B., & Mercatoris, B. C. N. (2015). Effective segmentation of green vegetation for resource-constrained real-time applications. In J. V. Stafford (Ed.), Precision agriculture’15. Proceedings of the 10th European conference on precision agriculture (pp. 259–265). The Netherlands: Wageningen Academic Publishers.
Nissimov, S., Goldberger, J., & Alchanatis, V. (2015). Object detection in a greenhouse environment using the Kinect sensor. Computers and Electronics in Agriculture, 113, 104–115.
Reina, G., & Milella, A. (2012). Towards autonomous agriculture: Automatic ground detection using trinocular stereovision. Sensors., 12(9), 12405–12423.
RHEA (2016). Robot fleets for highly effective agriculture and forestry management. Retrieved November 5, 2016, from http://www.rhea-project.eu/.
Ross, P., English, A., Ball, D., Upcroft, B., Wyeth, G., & Corke, P. (2014). Novelty-based visual object detection in agriculture. In IEEE international conference on robotics and automation (ICRA) (pp. 1699–1705).
Schölkopf, B., & Smola, A. J. (2002). Learning with kernels: Support vector machines, regularization, optimization, and beyond. Cambridge, MA, USA: MIT Press.
Steen, K. A., Christiansen, P., Karstoft, H., & Jørgensen, R. N. (2016). Using deep learning to challenge safety standard for highly autonomous machines in agriculture. Journal of Imaging, 2(1), 6.
SVS-VISTEK. The focal point of machine vision. Retrieved November 5, 2016, from http://www.svs-vistek.com/.
Tang, X. (1998). Texture information in run-length matrices. IEEE Transactions on Image Processing, 7(11), 1602–1609.
Tellaeche, A., Burgos-Artizzu, X.-P., Pajares, G., & Ribeiro, A. (2008). A vision-based method for weeds identification through the Bayesian decision theory. Pattern Recognition, 41, 521–530.
Vedaldi, A., & Fulkerson, B. (2010). Vlfeat: An open and portable library of computer vision algorithms. Proceedings of the 18th ACM international conference on Multimedia (MM ‘10) (pp. 1469–1472). New York, USA: ACM.
Yang, W., Zhao, X., Wang, S., Chen, L., Chen, X., & Lu, S. (2015). A new approach for greenness identification from maize images. In Proceedings of the intelligent computing theories and methodologies: 11th international conference, August 20–23, Fuzhou, China, pp. 339–347. Switzerland: Springer International Publishing.
Ye, M., Cao, Z., Yu, Z., & Bai, X. (2015). Crop feature extraction from images with probabilistic superpixel Markov random field. Computers and Electronics in Agriculture, 114, 247–260.
Acknowledgements
The first author acknowledges The National Council of Science and Technology of Mexico (CONACyT) for the doctoral grant number 210282 to undertake doctoral studies. H. Sossa thanks CONACyT under call: Frontiers of Science (Grant Number 65) for the economic support. The research leading to these results has been funded by the European Unions Seventh Framework Programme (FP7/2007-2013) under Grant Agreement No. 245986. We would like to express our thanks to the reviewers, especially to the Chief Editor, for their valued remarks, their comments and suggestions were fundamental for the improvement of this paper.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Campos, Y., Sossa, H. & Pajares, G. Comparative analysis of texture descriptors in maize fields with plants, soil and object discrimination. Precision Agric 18, 717–735 (2017). https://doi.org/10.1007/s11119-016-9483-4
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11119-016-9483-4