Effective Facial Feature Representation Based on Directional Micro-Patterns

Show simple item record

dc.contributor.author Ahmed, Faisal
dc.date.accessioned 2021-08-03T05:22:09Z
dc.date.available 2021-08-03T05:22:09Z
dc.date.issued 2012-09-30
dc.identifier.citation [1] Z. Zhang, “Feature-based facial expression recognition: Sensitivity analy- sis and experiment with a multi-layer perceptron,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 13, no. 6, pp. 893–911, 1999. [2] M. Z. Uddin, J. J. Lee, and T. S. Kim, “An enhanced independent component-based human expression recognition from video,” IEEE Trans- action on Consumer Electronics, vol. 55, no. 4, pp. 2216–2224, 2009. [3] Z. Guo, L. Zhang, and D. Zhang, “Rotation invariant texture classification using lbp variance(lbpv) with global matching,” Pattern Recognition, vol. 43, pp. 706–719, 2010. [4] T. Jabid, M. Kabir, and O. Chae, “Local directional pattern (ldp) for face recognition,” International Journal of Innovative Computing, Information and Control, vol. 8, no. 4, pp. 2423–2437, 2012. [5] X. Tan and B. Triggs, “Enhanced local texture feature sets for face recog- nition under difficult lighting conditions,” in IEEE International Workshop on Analysis and Modeling of Faces and Gestures, LNCS 4778, 2007, pp. 168–182. [6] T. Jabid, M. H. Kabir, and O. Chae, “Robust facial expression recognition based on local directional pattern,” ETRI Journal, vol. 32, no. 5, pp. 784– 794, 2010. 46 [7] M. C. Hwang, L. T. Ha, N. H. Kim, and C. S. Park, “Person identifica- tion system for future digital tv with intelligence,” IEEE Transaction on Consumer Electronics, vol. 53, no. 1, pp. 218–226, 2007. [8] P. Corcoran, C. Iancu, F. Callaly, and A. Cucos, “Biometric access control for digital media streams in home networks,” IEEE Transaction on Consumer Electronics, vol. 53, no. 3, pp. 917–925, 2007. [9] D. S. Kim, I. J. Jeon, S. Y. Lee, P. K. Rhee, and D. J. Chung, “Embedded face recognition based on fast genetic algorithm for intelligent digital pho- tography,” IEEE Transaction on Consumer Electronics, vol. 52, no. 3, pp. 726–734, 2006. [10] F. Zuo and P. H. N. de With, “Real-time embedded face recognition for smart home,” IEEE Transaction on Consumer Electronics, vol. 51, no. 1, pp. 183–190, 2005. [11] C. Shan, S. Gong, and P. W. McOwan, “Facial expression recognition based on local binary patterns: A comprehensive study,” Image and Vision Com- puting, vol. 27, no. 6, pp. 803–816, 2009. [12] Y. Rodriguez, “Face detection and verification using local binary patterns,” Ph.D. dissertation, Ecole Polytechnique Federale De Lausanne, 2006. [13] K. K. et al., “Facial feature extraction based on private energy map in dct domain,” ETRI Journal, vol. 29, no. 2, pp. 243–245, 2007. [14] K. C. et al., “A probabilistic network for facial feature verification,” ETRI Journal, vol. 25, no. 2, pp. 140–143, 2007. [15] P. Ekman and W. Friesen, “Facial action coding system: A technique for measurement of facial movement,” Palo Alto: Consulting Psychologists Press, 1978. [16] G. D. Guo and C. R. Dyer, “Simultaneous feature selection and classifier training via linear programming: A case study for face expression recogni- tion,” in IEEE Conference on Computer Vision and Pattern Recognition, 2003, pp. 346–352. [17] M. Valstar, I. Patras, and M. Pantic, “Facial action unit detection using probabilistic actively learned support vector machines on tracked facial point data,” in IEEE CVPR Workshop, vol. 3, 2005, pp. 76–84. [18] M. Valstar and M. Pantic, “Fully automatic facial action unit detection and temporal analysis,” in IEEE CVPR Workshop, 2006, p. 149. [19] L. Wiskott, J. Fellous, N. Kruger, and C. von der Malsburg, “Face recog- nition by elastic bunch graph matching,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 775–779, 1997. [20] R. Brunelli and T. Poggio, “Hyperbf networks for gender classification,” in DARPA Image Understanding Workship, 1992, pp. 311–314. [21] H. Abdi, D. Valentin, B. Edelman, and A. O’Toole, “More about the dif- ference between men and women: Evidence from linear neural network and principal component approach,” Neural Computing, vol. 7, no. 6, pp. 1160– 1164, 1995. [22] C. Padgett and G. Cottrell, “Representation face images for emotion clas- sification,” Advances in Neural Information Processing Systems, vol. 9, pp. 894–900, 1997. [23] M. S. Bartlett, J. R. Movellan, and T. J. Sejnowski, “Face recognition by independent component analysis,” IEEE Transaction on Neural Networks, vol. 13, no. 6, pp. 1450–1464, 2002. [24] C. C. Fa and F. Y. Shin, “Recognizing facial action units using indepen- dent component analysis and support vector machine,” Pattern Recognition, vol. 39, no. 9, pp. 1795–1798, 2006. [25] M. J. Lyons, J. Budynek, and S. Akamatsu, “Automatic classification of single facial images,” IEEE Transaction on Pattern Analysis and Machince Intelligence, vol. 21, no. 12, pp. 1357–1362, 1999. [26] Y. Tian, “Evaluation of face resolution for expression analysis,” in IEEE Workshop on Face Processing in Video, 2004. [27] T. Ahonen, A. Hadid, and M. Pietikainen, “Face recognition with local bi- nary pattern,” in European Conference on Computer Vision, vol. LNCS 3021, 2004, pp. 469–481. [28] M. Turk and A. Pentland, “Face recognition using eigenfaces,” in IEEE Conference on Computer Vision and Pattern Recognition, 1991, pp. 586– 591. [29] K. Etemad and R. Chellappa, “Dicriminant analysis for recognition of human face images,” Journal of the Optical Society of America, vol. 14, pp. 1724– 1733, 1997. [30] J. Yang, D. Zhang, A. Frangi, and J. Yang, “Two-dimensional pca: A new approach to appearance-based face representation and recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 1, pp. 131–137, 2004. [31] P. Penev and J. Atick, “Local feature analysis: A general statistical theory for object representation,” Network: Computation in Neural Systems, vol. 7, pp. 477–500, 1996. [32] M. Lades, J. Vorbruggen, J. Buhmann, J. Lange, C. von der Malsburg, R. Wurtz, and W. Konen, “Distortion invariant object recognition in the dynamic link architecture,” IEEE Transactions on Computers, vol. 42, no. 3, pp. 300–311, 1993. [33] S. Zhao, Y. Gao, and B. Zhang, “Sobel-lbp,” in IEEE International Confer- ence on Image Processing, 2008, pp. 2144–2147. [34] G. Donato, M. S. Bartlett, J. C. Hagar, P. Ekman, and T. J. Sejnowski, “Classifying facial actions,” IEEE Transaction on Pattern Analysis and Machince Intelligence, vol. 21, no. 10, pp. 974–989, 1999. [35] T. Ahonen, A. Hadid, and M. Pietikainen, “Face description with local bi- nary patterns: Application to face recognition,” IEEE Transaction on Pat- tern Analysis and Machince Intelligence, vol. 28, no. 12, pp. 2037–2041, 2006. [36] G. Zhao and M. Pietikainen, “Boosted multi-resolution spatiotemporal de- scriptors for facial expression recognition,” Pattern Recognition Letters, vol. 30, no. 12, pp. 1117–1127, 2009. [37] A. Shobeirinejad and Y. Gao, “Gender classification using interlaced deriva- tive patterns,” in International Conference on Pattern Recognition, 2010, pp. 1509–1512. [38] T. Ojala, M. Pietikainen, and T. Maenpaa, “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” IEEE Transaction on Pattern Analysis and Machince Intelligence, vol. 24, no. 7, pp. 971–987, 2002. [39] N. Sun, W. Zheng, C. Sun, C. Zou, and L. Zhao, “Gender classification based on boosting local binary pattern,” in International Symposium on Neural Networks, no. 2, 2006, pp. 194–201. [40] H. Lian and B. Lu, “Multi-view gender classification using local binary pat- terns and support vector machines,” in International Symposium on Neural Networks, no. 2, 2006, pp. 202–209. [41] H. Kabir, T. Jabid, and O. Chae, “Local directional pattern variance (ldpv): A robust feature descriptor for facial expression recognition,” International Arab Journal of Information Technology, vol. 9, pp. 382–391, 2012. [42] C. W. Hsu and C. J. Lin, “A comparison on methods for multiclass support vector machines,” IEEE Transaction on Neural Networks, vol. 13, no. 2, pp. 415–425, 2002. [43] S. Gundimada and V. K. Asari, “Facial recognition using multisensor im- ages based on localized kernel eigen spaces,” IEEE Transaction on Image Processing, vol. 18, no. 6, pp. 1314–1325, 2009. [44] S. E. Umbaugh, Digital Image Processing and Analysis. CRC Press, 2011. [45] T. Kanade, J. Cohn, and Y. Tian, “Comprehensive database for facial ex- pression analysis,” in IEEE International Conference on Automated Face and Gesture Recognition, 2000, pp. 46–53. [46] P. Phillips, H. Wechler, J. Huang, and P. Rauss, “The feret database and evaluation procedure for face recognition algorithms,” Image and, vol. 16, no. 10, pp. 295–306, 1998. [47] J. Bassili, “Emotion recognition: The role of facial movement and the relative importance of upper and lower areas of the face,” Journal of Personality and Social Psychology, vol. 37, no. 11, pp. 2049–2058, 1979. en_US
dc.identifier.uri http://hdl.handle.net/123456789/816
dc.description Supervised by Md. Hasanul Kabir, PhD, Department of Computer Science and Engineering (CSE), Islamic University of Technology (IUT), Gazipur 1704, Bangladesh. en_US
dc.description.abstract Automatic recognition and analysis of human face allow many interesting appli- cations in biometrics, human-computer interaction, and security industry, such as person identification, age estimation, gender classification, and facial expres- sion analysis. Deriving an efficient and effective feature representation is the fundamental component for any successful facial recognition system. However, the inherent variability of facial images caused by different factors like variations in illumination, pose, facial expression, alignment, and occlusions make classifi- cation a challenging task. Therefore, the aim of the ongoing research in facial recognition is to increase the robustness of the underlying feature representation against these factors. In this thesis, we work toward developing a simple, yet effective appearance- based feature descriptor for representing human facial image. A new local texture pattern, namely the directional ternary pattern (DTP) has been introduced that can effectively capture the texture properties of a local neighborhood and exhibits robustness against illumination variations and random noise. The proposed DTP operator encodes the texture information of a local neighborhood by labeling the edge response values in all eight directions around the center point using three different levels. Our encoding scheme employs a threshold in order to differentiate between uniform and high-textured face regions, and thus, ensures the generation of ternary micro-patterns consistent with the local texture property. The location and occurrence information of the DTP micro-patterns within the facial image is then represented with a spatial histogram, which functions as the facial feature descriptor. i Abstract ii The performance of the proposed method has been evaluated in three differ- ent applications, which are i) facial expression analysis, ii) face recognition, and iii) gender classification from facial image. Different publicly-available bench- mark image datasets were used for the experiments. We have also compared the performance of the proposed method with some widely-used local pattern-based feature descriptors. Experimental results demonstrate that, the DTP descriptor is more robust in extracting facial information and provides higher classification rate compared to some existing feature representation techniques. en_US
dc.language.iso en en_US
dc.publisher Department of Computer Science and Engineering, Islamic University of Technology, Board Bazar, Gazipur, Bangladesh en_US
dc.title Effective Facial Feature Representation Based on Directional Micro-Patterns en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search IUT Repository


Advanced Search

Browse

My Account

Statistics