Abstract
Object recognition occurs accurately with human visual neural mechanism despite in different complex background interference. For computer system, it is still a challenging work of object recognition and classification. Recently, many methods for object recognition based on human visual perception mechanism are presented. However, most methods cannot achieve a better recognition accuracy when object images are corrupted by some background interferences. Therefore, it is necessary to propose a method for object recognition of complex scene. Inspired by biomimetic visual mechanism and visual memory, a multi-channel biomimetic visual transformation (MCBVT) is proposed in this paper. MCBVT involves three channels. Firstly, some algorithms including orientation edge detection (OED), local spatial frequency detection (LSFD) and weighted centroid coordinate calculation are adopted for two stage’s visual memory maps creations during the first channel, where some visual memory points are stored in memory map. Secondly, an object hitting map (OHM) is built in the second channel and the OHM is an edge image without background interference. After that, the first stage’s visual memory hitting map is obtained through execute back-tracking second stage’s visual memory map. Furthermore, an OHM is constructed through back-tracking with common memory points in first stage’s visual memory map and first stage’s visual memory hitting map. Thirdly, the OED and LSFD algorithms are conducted to extract a feature map of OHM in the third channel. Consequently, the final feature map is reshaped into a feature vector, which is used for object recognition. Additionally, several image database experiments are implemented, the recognition accuracy for alphanumeric, MPEG-7 and GTSRB database are 93.33%, 91.33 and 90% respectively. Moreover, same object images in different backgrounds share with highly similar feature maps. On the contrary, different object images with complex backgrounds through MCBVT show different feature maps. The experiments reveal a better selectivity and invariance of MCBVT features. In summary, the proposed MCBVT provides a new framework of feature extraction. Background interference of object image is eliminated through the first and second channel, which is a new method for background noise reduction. Meanwhile, the results show that the proposed MCBVT method is better than other feature extraction methods. The contributions of this paper is significant in computational intelligence for the further work.


















Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Rolls ET, Webb TJ (2014) Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems. Front Comput Neurosci 8(85):1–19
Riesenhuber M, Poggio T (1999) Hierarchical models of object recognition in cortex. Nat Neurosci 2(11):1019–1025
Poggio T, Serre T (2013) Models of visual cortex. Scholarpedia. 8(4):3516
Hubel DH (1982) Exploration of the primary visual cortex. Nat. 299(5883):515–524
Yamada Y, Kawabe T, Miyazaki M (2012) Pattern randomness aftereffect. Sci Rep 3(2096):1–8
Graham NV (2011) Beyond multiple pattern analyzers modeled as linear filters (as classical v1 simple cells): useful additions of the last 25 years. Vis Res 51(13):1397–1430
Aboudib A, Gripon V, Coppin G. (2017) A biologically inspired framework for visual information processing and an application on modeling bottom-up visual attention. Cogn Comput. pp.1–20
Sun X, Shang K, Ming D, Tian J, Ma J (2015) A biologically-inspired framework for contour detection using superpixel-based candidates and hierarchical visual cues. Sensors. 15(10):26654–26674
Sountsov P, Santucci DM, Lisman JE (2011) (2011). A biologically plausible transform for visual recognition that is invariant to translation, scale, and rotation. Front Comput Neurosci 5(47):1–7
Liu K, Skibbe H, Schmidt T et al (2014) Rotation-invariant HOG descriptors using Fourier analysis in polar and spherical coordinates. Int J Compute Vis 106(3):342–364
Zhan J, Liang J, Zhang C, Zhao H. Scale invariant texture representation based on frequency decomposition and gradient orientation. Pattern Recogn Lett 2015; 51(C): 57–62
Bigot J, Gamboa F, Vimond M (2009) Estimation of translation, rotation, and scaling between noisy images using the fourier-mellin transform. Siam J Imaging Sci 2(2):614–645
Mennesson J, Saint-Jean C, Mascarilla L (2014) Color fourier-mellin descriptors for image recognition. Pattern Recogn Lett 40(1):27–35
Franklin SW, Rajan SE (2014) Retinal vessel segmentation employing ANN technique by Gabor and moment invariants-based features. Appl Soft Comput 22(9):94–100
Shi Y, Yang X, Guo Y (2014) Translation invariant directional framelet transform combined with gabor filters for image denoising. IEEE Trans Image Process 23(1):44–55
Li H, Liu Z, Huang Y, Shi Y (2015) Quaternion generic fourier descriptor for color object recognition. Pattern Recogn 48(12):3895–3903
Yang J, Yang MH (2013) Top-down visual saliency via joint CRF and dictionary learning. Conference on computer vision and pattern recognition. IEEE. 157(10):2296–2303
Xu Y, Li J, Chen J, Shen G, Gao Y (2017) A novel approach for visual saliency detection and segmentation based on objectness and top-down attention. International Conference on Image, Vision and Computing IEEE pp. 361–365
Liu W, Xue Q, Zhou J (2017) A novel image segmentation algorithm based on visual saliency detection and integrated feature extraction. International conference on communication and electronics systems. IEEE. pp. 1–5
Dieleman S, Willett KW, Dambre J (2105) Rotation-invariant convolutional neural networks for galaxy morphology prediction. Mon Not R Astron Soc 450(2):1441–1459
Cao Y, Chen Y, Khosla D (2015) Spiking deep convolutional neural networks for energy-efficient object recognition. Int J Comput Vis 113(1):54–66
Rolls ET (2012) Invariant visual object recognition: neural and computational bases. Front Comput Neurosci 6(2):35
Rolls ET, Webb TJ (2014) Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems. Front Comput Neurosci 8:85): 1–85):19
Robinson L, Rolls ET (2015) Invariant visual object recognition: biologically plausible approaches. Biol Cybern 109(4–5):505–535
Mutch J, Lowe DG (2008) Object class recognition and localization using sparse features with limited receptive fields. Int J Comput Vis 80(1):45–57
Alameer A, Ghazaei G, Degenaar P, Chambers JA, Nazarpour K (2016) Object recognition with an elastic net-regularized hierarchical max model of the visual cortex. IEEE Signal Proc Lett 23(8):1062–1066
Kheradpisheh SR, Ganjtabesh M, Masquelier T (2016) (2016). Bio-inspired unsupervised learning of visual features leads to robust invariant object recognition. Neurocomputing. 205(C):382–392
Cichy RM, Khosla A, Pantazis D, Torralba A, Oliva, A (2016) Deep neural networks predict hierarchical spatio-temporal cortical dynamics of human visual object recognition
Saifullah M, Balkenius C, Jönsson A (2014) A biologically based model for recognition of 2-d occluded patterns. Cogn Process 15(1):13–28
Meur OL, Baccino T (2013) Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behav Res Methods 45(1):251–266
Lewis-Peacock JA, Drysdale AT, Oberauer K, Postle BR (2012) Neural evidence for a distinction between short-term memory and the focus of attention. J Cogn Neurosci 24(1):61–79
Amit DJ, Fusi S (1994) Learning in Neural Networks with Material Synapses. MIT Press; 6(5): 957–982
Fusi S, Abbott LF (2007) Limits on the memory storage capacity of bounded synapses. Nat Neurosci 10(4):485–493
Pantic L, Torres J, Kappen H, Gielen S (2002) Associative memory with dynamic synapses. Neural Comput 14(12):2903–2923
Otsubo Y, Nagata K, Oizumi M, Okada M (2010) Instabilities in associative memory model with synaptic depression and switching phenomena among attractors. J Phys Soc Jpn 79(8):084002–084002-9
Caroni P, Chowdhury A, Lahr M (2014) Synapse rearrangements upon learning: from divergent-sparse connectivity to dedicated sub-circuits. Trends Neurosci 37(10):604–614
Lehoa VM, Wolfe JM (2015) (2015). The role of memory for visual search in scenes. Ann N Y Acad Sci 1339(1):72–81
Malcolm GL, Henderson JM (2010) Combining top-down processes to guide eye movements during real-world scene search. J Vis 10(2) 4.1
Wolfe JM, Võ ML, Evans KK, Greene MR (2011) Visual search in scenes involves selective and non-selective pathways. Trends Cogn Sci 15(2):77
Ma C, Huang JB, Yang X, Yang MH (2017) (2017). Robust visual tracking via hierarchical convolutional features. IEEE Trans on Pattern Anal & Mach Intell 1-1:99
Li X, Hu W, Shen C, Zhang Z, Dick A, Hengel AVD (2013) A survey of appearance models in visual object tracking. Acm Trans Intell Syst & Technology 4(4):1–48
Smeulders AWM, Chu DM, Cucchiara R, Calderara S, Dehghan A, Shah M (2013) Visual tracking: an experimental survey. IEEE Trans on Pattern Anal & Mach Intell. 36(7):1442–1468
Kristan M, Matas J, Leonardis A, Vojíř T, Pflugfelder R, Fernández G et al (2016) A novel performance evaluation methodology for single-target trackers. IEEE Trans on Pattern Anal & Mach Intell. 38(11):2137–2155
Dacey D, Packer OS, Diller L, Brainard D, Peterson B, Lee B (2000) Center surround receptive field structure of cone bipolar cells in primate retina. Vis Res 40(14):1801–1811
Latecki LJ, Lakamper R, Eckhardt T. (2000) Shape descriptors for non-rigid shapes with a single closed contour. Conference on Computer Vision and Pattern Recognition IEEE; pp. 424–429
Stallkamp J, Schlipsing M, Salmen J, Igel C (2012) Man vs. computer: benchmarking machine learning algorithms for traffic sign recognition. Neural Netw 32(2):323–332
Russell BC, Torralba A, Murphy KP, Freeman WT (2008) Labelme: a database and web-based tool for image annotation. Int J Comput Vis 77(1–3):157–173
Yu L, Zhou K, Yang Y, Chen H (2017) Biomimetic rstn invariant feature extraction method for image recognition and its application. IET Image Process 11(4):227–236
Xie Y, Liu LF, Li CH, Qu YY (2009) Unifying visual saliency with HOG feature learning for traffic sign detection. Intelligent Vehicles Symposium IEEE; pp. 24–29
Senthilkumar R, Gnanamurthy RK (2015) A comparative study of 2DPCA face recognition method with other statistically based face recognition methods. J Inst Eng 97(3):1–6
Lowe DG (2004) Distinctive image features from scale-invariant Keypoints. Int J Comput Vis 60(2):91–110
Bay H, Ess A, Tuytelaars T, Van Gool L (2008) Speeded-up robust features (surf). Comput Vis & Image Und 110(3):346–359
Wang W, Zhang M, Wang D, Jiang Y (2017) (2017). Kernel pca feature extraction and the svm classification algorithm for multiple-status, through-wall, human being detection. Eurasip J Wirel Comm & Networking 2017(1):151
Xu B, Ye Y, Nie L (2012) An improved random forest classifier for image classification. International Conference on Information and Automation IEEE; pp. 795–800
Skurichina M, Duin RPW (2002) Bagging, boosting and the random subspace method for linear classifiers. Pattern Ana Appl 5(2):121–135
Kuncheva LI, Rodriguez JJ, Plumpton CO, LindenDE, Johnston SJ. Random subspace ensembles for FMRI classification. IEEE Trans Med Imaging 2010; 29(2): 531–542
Funding
This research was funded by National Key Research and Development Plan (2018YFB1201602), Major Projects of Science and Technology in Hunan (2017GK1010), Natural Science Foundation of Hunan Province, China(2018JJ2531, 2018JJ2197),Research Foundation of Education Bureau of Hunan Province,China(18A305), in part by the State Key Laboratory of Robotics and System (HIT) (SKLRS-2017-KF-13), and in part by the National Natural Science Foundation of China (61403426).
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Ethical approvals
This article does not contain any studies with human participants or animals performed by any of the authors.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yu, L., Jin, M. & Zhou, K. Multi-channel biomimetic visual transformation for object feature extraction and recognition of complex scenes. Appl Intell 50, 792–811 (2020). https://doi.org/10.1007/s10489-019-01550-0
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-019-01550-0