Volume 17, Issue 3 (11-2020)                   JSDP 2020, 17(3): 141-156 | Back to browse issues page


XML Persian Abstract Print


Download citation:
BibTeX | RIS | EndNote | Medlars | ProCite | Reference Manager | RefWorks
Send citation to:

Ahmadi M A, Dianat R. Introducing a method for extracting features from facial images based on applying transformations to features obtained from convolutional neural networks. JSDP. 2020; 17 (3) :141-156
URL: http://jsdp.rcisp.ac.ir/article-1-837-en.html
University of QOM
Abstract:   (240 Views)
In pattern recognition, features are denoting some measurable characteristics of an observed phenomenon and feature extraction is the procedure of measuring these characteristics. A set of features can be expressed by a feature vector which is used as the input data of a system. An efficient feature extraction method can improve the performance of a machine learning system such as face recognition in the image field.
Most of the feature extraction methods in facial images are categorized as geometric feature extractor methods, linear transformation-based methods and neural network-based methods. Geometric features include some characteristics of the face such as the distance between the eyes, the height of the nose and the width of the mouth. In the second category, a linear transformation is applied to the original data and displaces them to a new space called feature space. In the third category, the last layer in the network, which is used for categorization, is removed, and the penultimate layer output is used as the extracted features. Convolutional Neural Networks (CNNs) are one the most popular neural networks and are used in recognizing and verifying the face images, and also, extracting features.
The aim of this paper is to present a new feature extraction method. The idea behind the method can be applied to any feature extraction problem. In the proposed method, the test feature vector is accompanied with the training feature vectors in each class. Afterward, a proper transform is applied on feature vectors of each class (including the added test feature vector) and a specific part of the transformed data is considered. Selection of the transform type and the other processing, such as considering the specific part of the transformed data, is in such a way that the feature vectors in the actual class are encountered with less disturbing than the other ones. To meet this goal, two transformations, Fourier and Wavelet, have been used in the proposed method. In this regard, it is more appropriate to use transformations that concentrate the energy at low frequencies. The proposed idea, intuitively, can lead to improve the true positive (TP) rate.
As a realization, we use the idea in CNN-based face recognition problems as a post-processing step and final features are used in identification. The experimental results show up to 3.4% improvement over LFW dataset.
Full-Text [PDF 4358 kb]   (66 Downloads)    
Type of Study: Research | Subject: Paper
Received: 2018/04/10 | Accepted: 2019/09/2 | Published: 2020/12/5 | ePublished: 2020/12/5

References
1. [1] A. Krizhevsky, I. Sutskever, and G. E. Hinton, "Imagenet classification with deep convolutional neural networks," in Advances in neural information processing systems, pp. 1097-1105, 2012.
2. [2] R. Brunelli and T. Poggio, "Face recognition: Features versus templates," IEEE transactions on pattern analysis and machine intelligence, vol. 15, no. 10, pp. 1042-1052, 1993. [DOI:10.1109/34.254061]
3. [3] R. J. Baron, "Mechanism of human facial recognition," International Journal of Man Machine Studies, vol. 15, pp. 137-178, 1981. [DOI:10.1016/S0020-7373(81)80001-6]
4. [4] D. Ghimire, J. Lee, Z.-N. Li, and S. Jeong, "Recognition of facial expressions based on salient geometric features and support vector machines," Multimedia Tools and Applications, vol. 76, no. 6, pp. 7921-7946, 2017 [DOI:10.1007/s11042-016-3428-9]
5. [5] L. Wiskott, N. Krüger, N. Kuiger, and C. Von Der Malsburg, "Face recognition by elastic bunch graph matching," IEEE Transactions on pattern analysis and machine intelligence, vol. 19, no. 7, pp. 775-779, 1997. [DOI:10.1109/34.598235]
6. [6] J. yves Bouguet, "Pyramidal implementation of the lucas kanade feature tracker," Intel Corporation, Microprocessor Research Labs, 2000.
7. [7] Y. Freund and R. E. Schapire, "A desicion-theoretic generalization of on-line learning and an application to boosting," in European conference on computational learning theory, 1995: Springer, pp. 23-37. [DOI:10.1007/3-540-59119-2_166]
8. [8] G.-B. Huang, H. Zhou, X. Ding, and R. Zhang, "Extreme learning machine for regression and multiclass classification," IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 42, no. 2, pp. 513-529, 2012. [DOI:10.1109/TSMCB.2011.2168604] [PMID]
9. [9] W. Ouarda, H. Trichili, A. M. Alimi, and B. Solaiman, "Face recognition based on geometric features using Support Vector Machines," in Soft Computing and Pattern Recognition (SoCPaR), 2014 6th International Conference of, 2014: IEEE, pp. 89-95. [DOI:10.1109/SOCPAR.2014.7007987]
10. [10] M. A. Turk and A. P. Pentland, "Face recognition using eigenfaces," in Computer Vision and Pattern Recognition, 1991. Proceedings CVPR'91., IEEE Computer Society Conference on, 1991: IEEE, pp. 586-591.
11. [11] H. Hotelling, "Analysis of a complex of statistical variables into principal components," Journal of educational psychology, vol. 24, no. 6, p. 417, 1933. [DOI:10.1037/h0071325]
12. [12] P. N. Belhumeur, J. P. Hespanha, and D. J. Kriegman, "Eigenfaces vs. fisherfaces: Recognition using class specific linear projection," IEEE Transactions on pattern analysis and machine intelligence, vol. 19, no. 7, pp. 711-720, 1997. [DOI:10.1109/34.598228]
13. [13] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, "Robust face recognition via sparse representation," IEEE transactions on pattern analysis and machine intelligence, vol. 31, no. 2, pp. 210-227, 2009. [DOI:10.1109/TPAMI.2008.79] [PMID]
14. [14] A. Wagner, J. Wright, A. Ganesh, Z. Zhou, H. Mobahi, and Y. Ma, "Toward a practical face recognition system: Robust alignment and illumination by sparse representation," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 2, pp. 372-386, 2012. [DOI:10.1109/TPAMI.2011.112] [PMID]
15. [15] Y. Sun, X. Wang, and X. Tang, "Deep learning face representation from predicting 10,000 classes," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1891-1898. [DOI:10.1109/CVPR.2014.244]
16. [16] Y. Sun, Y. Chen, X. Wang, and X. Tang, "Deep learning face representation by joint identification-verification," in Advances in neural information processing systems, 2014, pp. 1988-1996.
17. [17] Y. Sun, X. Wang, and X. Tang, "Deeply learned face representations are sparse, selective, and robust," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 2892-2900. [DOI:10.1109/CVPR.2015.7298907]
18. [18] Y. Sun, D. Liang, X. Wang, and X. Tang, "Deepid3: Face recognition with very deep neural networks," arXiv preprint arXiv:1502.00873, 2015.
19. [19] Y. Taigman, M. Yang, M. A. Ranzato, and L. Wolf, "Deepface: Closing the gap to human-level performance in face verification," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 1701-1708. [DOI:10.1109/CVPR.2014.220]
20. [20] O. M. Parkhi, A. Vedaldi, and A. Zisserman, "Deep Face Recognition," in BMVC, 2015, vol. 1, no. 3, p. 6. [DOI:10.5244/C.29.41]
21. [21] K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition," arXiv preprint arXiv:1409.1556, 2014.
22. [22] F. S. Samaria and A. C. Harter, "Parameterisation of a stochastic model for human face identification," in Proceedings of 1994 IEEE workshop on applications of computer vision, 1994: IEEE, pp. 138-142.
23. [23] S. Milborrow, J. Morkel, and F. Nicolls, "The MUCT landmarked face database," Pattern Recognition Association of South Africa, vol. 201, no. 0, 2010.
24. [24] G. B. Huang, M. Mattar, T. Berg, and E. Learned-Miller, "Labeled faces in the wild: A database forstudying face recognition in unconstrained environments," in Workshop on faces in'Real-Life'Images: detection, alignment, and recognition, 2008.
25. [25] A. Vedaldi and K. Lenc, "Matconvnet: Convolutional neural networks for matlab," in Proceedings of the 23rd ACM international conference on Multimedia, 2015: ACM, pp. 689-692. [DOI:10.1145/2733373.2807412]
26. [26] GTDLBench. The Database of Faces (AT&T) [Online]. Available: https://git-disl.github.io/GTDLBench/datasets/att_face_dataset/.
27. [27] P. Grother and M. Ngan, "Performance of face identification algorithms," NIST Inter-agency Internal Report8009, 2014.
28. [28] R. A. Fisher, "The use of multiple measurements in taxonomic problems," Annals of eugenics, vol. 7, no. 2, pp. 179-188, 1936. [DOI:10.1111/j.1469-1809.1936.tb02137.x]
29. [29] S. Mika, G. Ratsch, J. Weston, B. Scholkopf, and K.-R. Mullers, "Fisher discriminant analysis with kernels," in Neural networks for signal processing IX: Proceedings of the 1999 IEEE signal processing society workshop (cat. no. 98th8468), 1999: Ieee, pp. 41-48.

Add your comments about this article : Your username or Email:
CAPTCHA

Send email to the article author


© 2015 All Rights Reserved | Signal and Data Processing