Abstract
In this paper, we propose a new object recognition framework that combines Gabor energy filters, a visual cortex model in which single units alternate with complex units, and color information. Each color image is first converted to the CIELAB color space. Rather using Gabor filters in the first layer of the cortex model, to each color component, a set of Gabor energy filters is applied. Thereafter, the superposition responses of the Gabor energy filter outputs over the color components are normalized by divisive normalization. In the fourth layer, sparse features are calculated using a localized pooling method that allows retention of some geometric information from the prototype patches’ positions. Finally, a set of sparse features are exploited by a linear SVM classifier for object recognition and classification. In the learning stage, a set of prototypes is selected randomly over spatial position, spatial size, and several scales simultaneously, and is extracted by the local maximum over scales and orientations, ignoring weaker training scales and orientations. The results of experiments performed on several datasets show that the use of color information in our framework improves object recognition significantly.
Similar content being viewed by others
References
M. A. Asari, U. U. Sheikh, and E. Supriyanto, “3D shape descriptor for object recognition based on Kinectlike depth image,” Image and Vision Computing, vol. 32, no. 4, pp. 260–269, 2014. [click]
S. J. Lee, M. K. Park, and K. H. Lee, “Full 3D surface reconstruction of partial scan data with noise and different levels of scale,” Journal of Mechanical Science and Technology, vol. 28, no. 8, pp.3171-3180, 2014. [click]
C. S. Jeong, J. H. Park, S. I. Han, and J. S. Kim, “Shape recognition performance analysis and improvement in Sendzimir rolling mills,” Journal of Mechanical Science and Technology, vol. 28, no. 4, pp. 1455–1463, 2014. [click]
D. Zhang, J. Wang, and Y. Yang, “Design 3D garments for scanned human bodies,” Journal of Mechanical Science and Technology, vol. 28, no. 7, pp. 2479–2487, 2014. [click]
Q. N. Vo and G. Lee, “Illumination invariant object tracking with adaptive sparse representation,” International Journal of Control, Automation and Systems, vol. 12, no. 1, pp. 195–201, 2014. [click]
W. Xu and E. J. Lee, “A hybrid method based on dynamic compensatory fuzzy neural network algorithm for face recognition,” International Journal of Control, Automation and Systems, vol. 12, no. 3, pp. 688–696, 2014. [click]
M. Riesenhuber and T. Poggio, “Hierarchical models of object recognition in cortex,” Nature Neuroscience, vol. 2, no. 11, pp. 1019–1025, 1999. [click]
D. H. Hubel and T. N. Wiesel, “Receptive fields of single neuronses in the cat’s striate cortex,” Journal of Physiology-London, vol. 587, no. 12, pp. 2722–2732, 2009.
T. Serre, L. Wolf, and T. Poggio, “Object recognition with features inspired by visual cortex,” IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, pp. 994–1000, 2005. [click]
T. Serre, L. Wolf, S. Bileschi, M. Riesenhuber, and T. Poggio, “Robust object recognition with cortex-like mechanisms,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 29, no. 3, pp. 411–426, 2007. [click]
J. Mutch and D. G. Lowe, “Multiclass object recognition with sparse localized features,” IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, pp. 11–18, 2006. [click]
J. Mutch and D. G. Lowe, “Object class recognition and localization using sparse features with limited receptive fields,” International Journal of Computer Vision, vol. 80, no. 1, pp. 45–57, 2008. [click]
R. Shapley and M. Hawken, “Color in the cortex: Singleand double-opponent cells,” Vision Research, vol. 51, no. 7, pp. 701–717, 2011. [click]
H. Jhuang, E. Garrote, X. Yu, V. Khilnani, T. Poggio, A. D. Steele, and T. Serre, “Automated home-cage behavioural phenotyping of mice,” Nature Communications, vol. 1, AN: 68, 2010. [click]
C. Thériault, N. Thome, and M. Cord, “Extended coding and pooling in the HMAX model,” IEEE Trans. on Image Processing, vol. 22, no. 2, pp. 764–777, 2013. [click]
Y. Huang, K. Huang, D. Tao, T. Tan, and X. Li, “Enhanced biologically inspired model for object recognition,” IEEE Trans. on Systems Man and Cybernetics, part B, vol. 41, no. 6, pp. 1668–1680, 2011. [click]
J. Yang, K. Yu, Y. Gong, and T. Huang, “Linear spatial pyramid matching using sparse coding for image classification,” Proc. of IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, pp. 1794–1801, 2009. [click]
N. Petkov and P. Kruizinga, “Computational models of visual neurons specialized in the detection of periodic and aperiodic oriented visual stimuli: Bar and grating cells,” Biological Cybernetics, vol. 76, no. 2, pp. 83–96, 1997. [click]
Y. S. Heo, K. M. Lee, and S. U. Lee, “Joint depth map and color consistency estimation for stereo images with different illuminations and cameras,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 35, no. 5, pp. 1094–1106, 2013. [click]
A. Turnip, K.-S. Hong, and M.-Y. Jeong, “Real-time feature extraction of P300 component using adaptive nonlinear principal component analysis,” Biomedical Engineering Online, vol. 10, AN: 83, 2011. [click]
T. Banlue, P. Sooraksa, and S. Noppanakeepong, “A practical position-based visual servo design and implementation for automated fault insertion test,” International Journal of Control, Automation and Systems, vol. 12, no. 5, pp. 1090–1101, 2014. [click]
B. Bouchemal and A. Zaatri, “Image-based control for cable-based robots,” International Journal of Control, Automation and Systems, vol. 12, no. 1, pp.118-125, 2014. [click]
K. E. A. van de Sande, T. Gevers, and C. G. M. Snoek, “Evaluating color descriptors for object and scene recognition,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 32, no. 9, pp. 1582–1596, 2010. [click]
T. T. Q. Bui and K.-S. Hong, “Evaluating a color-based active basis model for object recognition,” Computer Vision and Image Understanding, vol. 116, no. 11, pp. 1111–1120, 2012. [click]
J. E. Ha, “An image processing algorithm for the automatic manipulation of tie rod,” International Journal of Control, Automation and Systems, vol. 11, no. 5, pp. 984–990, 2013. [click]
D. R. Martin, C. C. Fowlkes, and J. Malik, “Learning to detect natural image boundaries using local brightness, color, and texture cues,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 26, no. 5, pp. 530–549, 2004. [click]
S. Jalali, C. Tan, J. H. Lim, J. Y. Tham, S. H. Ong, P. Seekings, and E. Taylor, “Visual recognition using a combination of shape and color features,” Proc. of the Annual Meeting of the Cognitive Science Society, pp. 2638–2643, 2013.
S. Palmer, Vision Science: Photons to Phenomenology, MIT Press, Cambridge, MA, 1999.
C. C. Chang and C. J. Lin, “LIBSVM: A library for support vector machines,” ACM Trans. on Intelligent Systems and Technology, vol. 2, no. 3, AN: 27, 2011. [click]
M. Weber, W. Welling, and P. Perona, “Unsupervised learning of models for recognition,” Proc. of the 6th European Conference on Computer Vision, Part I, pp. 18–32, 2000. [click]
R. Fergus, P. Perona, and A. Zisserman, “Object class recognition by unsupervised scale-invariant learning,” Proc. of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 264–271, 2003. [click]
L. Fei-Fei, R. Fergus, and P. Perona, “Learning generative visual models from few training examples: An incremental Bayesian approach tested on 101 object categories,” Proc. of the IEEE Computer Society Conf. on Computer Vision and Pattern Recognition Workshops, pp. 178–187, 2004. [click]
G. Griffin, A. Holub, and P. Perona, Caltech-256 object category dataset. Technical Report #7694, California Institute of Technology, Pasadena, CA, 2007.
J. G. Daugman, “Uncertainty relation for resolution in space, spatial-frequency, and orientation optimized by twodimensional visual cortical filters,” Journal of the Optical Society of America A-Optics Image Science and Vision, vol. 2, no. 7, pp. 1160–1169, 1985. [click]
N. Petkov, “Biologically motivated computationally intensive approaches to image pattern recognition,” Future Generation Computer Systems, vol. 11, no. 4-5, pp. 451–465, 1995. [click]
D. J. Heeger, “Modeling simple-cell direction selectivity with normalized, half-squared, linear operators,” Journal of Neurophysiology, vol. 70, no. 5, pp. 1885–1898, 1993.
A. Turnip and K.-S. Hong “Classifying mental activities from EEG-P300 signals using adaptive neural network,” International Journal of Innovative Computing, Information and Control, vol. 8, no. 9, pp. 6429–6443, 2012.
N. Naseer and K.-S. Hong, “Classification of functional near-infrared spectroscopy signals corresponding to the right-and left-wrist motor imagery for development of a brain-computer interface,” Neuroscience Letters, vol. 553, pp. 84–89, 2013. [click]
N. Naseer, M. J. Hong, and K.-S. Hong, “Online binary decision decoding using functional near-infrared spectroscopy for development of a brain-computer interface,” Experimental Brain Research, vol. 232, no. 2, pp. 555–564, 2014. [click]
G. M. Lim, D. M. Bae, and J. H. Kim, “Fault diagnosis of rotating machine by thermography method on support vector machine,” Journal of Mechanical Science and Technology, vol. 28, no. 8, pp. 2947–2952, 2014. [click]
S. D. Oh, Y. J. Kim, and T. H. Lee, “Fast pattern recognition inspection system (FPRIS) for machine vibration,” Journal of Mechanical Science and Technology, vol. 28, no. 2, pp. 437–444, 2014. [click]
S. Cho and S. Jo, “Incremental online learning of robot behaviors from selected multiple kinesthetic teaching trials,” IEEE Trans. on Systems, Man, and Cybernetics: Systems, vol. 43, no. 3, pp. 730–740, 2013. [click]
H. Lee, K. Kwak, and S. Jo, “Multi-sensor-based online positive learning for drivable region detection,” Electronics Letters, vol. 50, no. 20, pp. 1440–1442, 2014. [click]
B. Choi and S. Jo, “A low-cost EEG system-based hybrid brain-computer interface for humanoid robot navigation and recognition,” PLoS One, vol. 8, no. 9, AN: e74583, 2013. [click]
M. Rehan and K.-S. Hong, “Modeling and automatic feedback control of tremor: Adaptive estimation of deep brain stimulation,” PLoS One, vol. 8, no. 4, AN: e62888, pp. 1–11, 2013. [click]
Y.-S. Kim and K.-S. Hong, “An IMM algorithm with federated information mode-matched filters for AGV,” International Journal of Adaptive Control and Signal Processing, vol. 21, no. 7, pp. 533–555, 2007. [click]
J. E. Ha and W. H. Lee, “Foreground objects detection using multiple difference images,” Optical Engineering, vol. 49, no. 4, AN: e047201, 2010. [click]
M. H. Jeong and W. H. Lee, “Invariants with new indexing functions: Application to welding panel recognition,” International Journal of Control, Automation and Systems, vol. 9, no. 1, pp. 80–85, 2011. [click]
B. M. Suhng and W. H. Lee, “A new link-based single tree building algorithm for shortest path searching in an urban road transportation network,” Journal of Electrical Engineering and Technology, vol. 8, no. 4, pp. 889–898, 2013. [click]
H. Zhou, M. Fei, A. Sadka, Y. Zhang, and X. Li, “Adaptive fusion of particle filtering and spatio-temporal motion energy for human tracking,” Pattern Recognition, vol. 47, no. 11, pp. 3552–3567, 2014. [click]
K.-S. Hong and J. Bentsman, “Direct adaptive control of parabolic systems: Algorithm synthesis, and convergence and stability analysis,” IEEE Trans. on Automatic Control, vol. 39, no. 10, pp. 2018–2033, 1994. [click]
K.-S. Hong, “Asymptotic behavior analysis of a coupled time-varying system: Application to adaptive systems,” IEEE Trans. on Automatic Control, vol. 42, no. 12, pp. 1693–1697, 1997. [click]
Q. C. Nguyen and K.-S. Hong, “Asymptotic stabilization of a nonlinear axially moving string by adaptive boundary control,” Journal of Sound and Vibration, vol. 329, no. 22, pp. 4588–4603, 2010. [click]
M. Rehan and K.-S. Hong, “LMI-based robust adaptive synchronization of FitzHugh-Nagumo neurons with unknown parameters under uncertain external electrical stimulation,” Physics Letters A, vol. 375, no. 15, pp. 1666–1670, 2011. [click]
L. H. Nguyen and K.-S. Hong, “Adaptive synchronization of two coupled chaotic Hindmarsh-Rose neurons by controlling the membrane potential of a slave neuron,” Applied Mathematical Modelling, vol. 37, no. 4, pp. 2460–2468, 2013. [click]
K.-S. Hong, H. C. Sohn, and J. K. Hedrik, “Modified skyhook control of semi-active suspensions: A new model, gain scheduling, and hardware-in-the-loop tuning,” Journal of Dynamic Systems, Measurement and Control, vol. 124, No. 1, pp. 158–167, 2002. [click]
Q. H. Ngo and K.-S. Hong, “Sliding-mode antisway control of an offshore container crane,” IEEE/ASME Transactions on Mechatronics, vol. 17, no. 2, pp. 201–209, 2012. [click]
W. He, S. S. Ge, B. V. E. How, Y. S. Choo, and K.-S. Hong, “Robust adaptive boundary control of a flexible marine riser with vessel dynamics,” Automatica, vol. 47, no. 4, pp. 722–732, 2011. [click]
Y. S. Kim, K.-S. Hong, and S. K. Sul, “Anti-sway control of container cranes: inclinometer, observer, and state feedback,” International Journal of Control, Automation and Systems, vol. 2, no. 4, pp. 435–449, 2004.
K.-S. Hong, K. H. Choi, J. G. Kim, and S. Lee, A PCbased open robot control system: PC-ORC, Robotics and Computer-Integrated Manufacturing, vol. 17, no. 4, pp. 355–365, 2001. [click]
K. J. Yang, K.-S. Hong, and F. Matsuno, “Robust adaptive boundary control of an axially moving string under a spatiotemporally varying tension,” Journal of Sound and Vibration, vol. 273, no. 4, pp. 1007–1029, 2004. [click]
H. Park, D. Chwa, and K.-S. Hong, “A feedback linearization control of container cranes: Varying rope length,” International Journal of Control, Automation and Systems, vol. 5, no. 4, pp. 379–387, 2007.
V. Vapnik, The Nature of Statistical Learning Theory, Springer-Verlag, London, 1995.
Author information
Authors and Affiliations
Corresponding author
Additional information
Recommended by Editor Euntai Kim. This work was supported by the National Research Foundation of Korea funded by the Ministry of Science, ICT and Future Planning, Korea (grant no. NRF-2014-R1A2A1A10049727) and Vietnam Academy of Science and Technology, Vietnam (grant no: VAST01.02/14-15).
T. T. Quyen Bui received the B.S. and M.S. degrees in Instrumentation and Control from Hanoi University of Science and Technology, Vietnam, in 2001 and 2006, respectively, and her Ph.D. degree in Mechanical Engineering from Pusan National University, Korea, in 2013. She has joined the Department of Automation Technology, Institute of Information Technology (IOIT), Vietnam Academy of Science and Technology as a researcher since July 2001. At present, she is the head of the Department of Automation Technology, IOIT. Her research interests include computer vision, image processing, robotic system, and measurement systems.
Thang T. Vu received his B.E. and M.S. degrees in Electronics and Telecommunications from the Hanoi University of Science and Technology in 2002 and 2005, respectively. He received his Ph.D. degree in Information System from the Japan Advanced Institute of Science and Technology in 2008. Currently, he is a senior researcher, leader of the Multimedia Human-Machine Language Technology Department at the Institute of Information Technology, Vietnam Academy of Science and Technology. He is also a lecturer at the University of Science and Technology of Hanoi and a member of the Research Institute of Signal Processing. His research interests include language understanding, computer vision, dialog system, and robotics.
Keum-Shik Hong Please see vol. 13, no. 2, p. 425, April, 2015 of this journal.
Rights and permissions
About this article
Cite this article
Quyen Bui, T.T., Vu, T.T. & Hong, KS. Extraction of sparse features of color images in recognizing objects. Int. J. Control Autom. Syst. 14, 616–627 (2016). https://doi.org/10.1007/s12555-014-0502-9
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12555-014-0502-9