1. Introduction
Over the last decades, face recognition has emerged as an active research area in computer vision with numerous potential applications including video-mediated communication biometrics, human-computer interaction, surveillance and content-based access of images and video databases. Moreover, systems that rely on face recognition have gained great importance ever since terrorist threats imposed weakness among the implementd security systems.
Various face recognition algorithms have been developed and used as an application of access control and surveillance. On the analysis of face recognition system, some of the important studies on face recognition systems are discussed. Principal Component Analysis (PCA) is also known as Eigen face method, in PCA method the images are projected onto the facial value so called eigenspace [1,2]. PCA approach reduces the dimension of the data by means of basic data compression method [3] and reveals the most effective low dimensional structure of facial patterns [4]. Local Feature Analysis (LFA) method of recognition is based on the analysis of the face in terms of local features e.g. eye, nose etc. by what is referred LFA kernels. References [5-7] is based on learning of the faces in an “Example Set” by the machine in the “Training Phase” and carrying out recognition in the “Generalization Phase”. The support vectors consist of a small subset of training data extracted by the algorithm given in [8]. This has resulted in the development of successful algorithms and the introduction of commercial products.
However, it remains still an unsolved problem which presents only successful results under controlled situations where the limited conditions like illumination, pose, facial expressions, and especially in the case of partial occlusion, the performance of recognition is abruptly degraded under these uncontrolled situations. Face recognition under controlled environments has been in scene for the past decades but recognition under uncontrolled situations like illumination, expression and partial occlusion is a recent issue. Partial face occlusion is one of the most challenging problems in face recognition. A face recognition system can confront occluded faces in real world applications very due to use of accessories, such as scarf or sunglasses, hands on the face, the objects that persons carry, and external sources that partially occlude the camera view, samples of patially occluded face image from AR face database are shown in Fig. 5. Therefore, the face recognition system has to be robust to occlusion in order to guarantee reliable real-world operation [9].
Numerous researches have been conducted in order to address this challenge [9-15]. In reference [9], the face image is first divided into k local regions and for each region an eigenspace is constructed. If a region is occluded, it is automatically detected. Moreover, weighting of the local regions are also proposed in order to provide robustness against expression variations. M. Aleix proposed a similar approach in [10], where a self-organizing map (SOM) is used to model the subspace. Face attributed relational graph (ARG) represented the face in [12], this representation contains a set of nodes and binary relations between these nodes. In [13] robustness against occlusion is provided by combining the subspace methods that aim at best reconstruction, such as principal component analysis, with the subspace methods that aim at discrimination, such as linear discriminat analysis. A sparse signal representation is used to analyze partially occluded face images in [14]. Another representation based approach is proposed in [15]. Different from the studies [9-14] in which occluded images are only included in the testing set, they are included both in the training and testing sets in [15]. Overall, a major drawback of these methods is that they have a high computational cost in terms of training and testing the data.
To address the challenge of partial occluded face recognition, a novel method based on linear regression-based classification (LRC) algorithm is proposed in this paper. First, all images are downsampled and divided into several blocks, and then we exploit the evaluator of each block to determine the clear blocks of the test face image by using linear regression technique. Next, the remained uncontaminated blocks are utilized to partial occluded face recognition issue. Furthermore, an improved Distance-based Evidence Fusion approach is proposed to decide in favor of the class with average value of corresponding minimum distance.
The rest of this paper is organized as follow: In Section 2, we review the previous work in face recognition and partially occluded face recognition, especially. Section 3 focuses on determining the clear blocks of the test face image using B-LRC and recognizing face by using i-DEF algorithm. Section 4 presents experimental evaluation of classification accuracy using proposed method compared to other methods. Section 5 gives the conclusion for this study.
2. Related Work
Generally speaking, face recognition methods whether under controlled conditions or uncontrolled conditions can be classified into three groups i.e., feature based method that deal with features like eyes, mouth, nose and establish a geometrical correspondence between them [16]. The second category is appearance-based method that focus on the holistic features of face images by considering the whole face region. Another one deals with the hybrid local and global features of face images to be used for recongnition purpose.
Nevertheless, It is difficult for appearance-based face recognition techniques based on the conventional component analysis techniques, such as principal component analysis (PCA) [1, 2], linear discriminate analysis (LDA) [4], and independent component analysis (ICA) [23] due to partial occlusions caused by sunglass, scarf, hair or other factors, Basic methods used up to now for handling occlusion belongs to one of the following methods: fractal based method, feature based method and part based method.
For fractal based method, Marsico et al. presented Partitioned Iterated Function System (PIFS) based face recognition algorithm in [17]. This method computes the self-similarities among the image and establishes a relationship between the squares of grid region. But PIFS algorithm is sensitive to occlusions, so to avoid this individual component of face, parts are handled locally like eys, mouth and nose. Futher distortions are removed by ad-hoc distance measure.
Global and local features play an important role in occluded face recognition. The conventional methods used global features from one kernel [18, 19]. These features can be affected by noise or occlusions which reduces the robustness of these methods. The recognition techniques have also been used effectively using local features [15, 18, 20, 21, 22]. The partial occlusion affects the local features but the recognition methods can be made robust if these local features are merged together intelligently.
Recently, sparse representation technique has drawn wide attention in solving computer vision issues [24]. Wright et al. proposed the sparse representation-based classification (SRC) method for robust face recognition [25]. SRC uses the entire training set to represent a test sample with the coding vector subject to sparse constraint. The core of SRC is a l1-minimization problem (minimizing the l1-norm problem), which is approximately equal to the related l0-minimization problem (minimizing the l0-norm problem) under some conditions. Although SRC can achieve very promising recognition results, the computational cost of solving the l1-minimization problem is quite high. Especially, when dealing with random corruptions and occlusions, an augmented sample matrix (ASM) is used to represent the test sample, which is referred to as the extended SRC (eSRC) method. ASM is very large compared with the original sample matrix, which greatly increases the scale of the optimization problem of eSRC and results in very heavy computational load. Another important relevant research called Linear regression-based classification (LRC) is proposed for robust face recognition [26]. LRC represents a test image as a linear combination of training images of class-specific samples. Therefore, face recognition can be cast as a problem of linear regression. Compared with SRC, LRC is much more time-saving to implement because it only uses least-squares estimation in computation while SRC needs iterative procedures to solve the l1- minimization. Howerver, for LRC which essentially can be categorized as a nearest subspace [26], it estimates a given probe against class-specific models in the least-squares sense, and insufficient training samples will also make the estimation incorrect.
3. The Methodology of Partial Occluded Face Recognition
In this section, we first briefly describe the LRC algorithm and analyze that the error components derived from LRC are more suitable for identifying occlusion parts than original test face image, which motivateds the work of this paper in section 3.2. The main steps of our method are as follows:
1. Downsampled and divide the face image into M blocks uniformly;
2. Construct the linear model to represent the input image based on LRC;
3. Exploit the evaluator of each block to determine the clear blocks of the test face image using B-LRC;
4. An improved Distance-based Evidence Fusion (i-DEF) method is proposed to final face recognition issue on the selected blocks.
3.1 Introduction of Linear Regression-based classification
Let there be N number of distinguished classes with pi number of training images from the i-th class, i = 1,2,…, N. Each gray-scale training images is of an order a × b and is represented as , i = 1,2,…, N and m = 1,2,..., pi. Each gallery image is downsampled to an order c × d and transformed to vector through column concatenation such that → , where q = cd, cd < ab. Each image vector is normalized so that maximum pixel value is 1. Using the concept that patterns from the same class lie on a linear subspace [27], a class-specific model [26] Xi by stcking the q-dimensional image vectors as Xi = , i = 1,2,…, N. Each vector spans a subspace of IRq also called the column space of Xi. Therefore, at the training level, each class i is represented by a vector subspace, Xi, which is also called the regressor or predictor for class i. Let z be an unlabeled test image and our problem is to classify z as one of the class i = 1,2,…, N. The author transformed and normalized the gray-scale image z to an image vector y ∈ Rq×1 as discussed for the gallery. If y belongs to the i-th class, it should be represented as a linear combination of the training images from the same class (lying in the same subspace).
Where αi ∈ IRpi×1 is the vector of parameters. Given that q ≥ pi, the system of equations in (1) is well conditioned and αi can be estimated using least-squares estimation [26].
The estimated vector of parameters, , along with the predictors Xi are used to predict the response vector for each class i,
Where the predicted vector ∈ IRq×1 is the projection of y onto the i-th subspace. In another word, is the closest vector, in the i-th subspace, to the observation vector y in the euclidean sense. H is called hat matrix since it maps y into . The distance measure between the predicted response vector and the original response vector y is calculated as
And rule in faver of the class with minimum distance, i.e.,
3.2 Detection of Occluded Region Based on Block-LRC (B-LRC)
A partial occluded part in face images usually occurs due to sunglasses, scarves, mustaches and so on. Such variations in facial appearance are commonly encountered in uncontrolled situations and may cause big trouble to the face recognition system. The challenge of partial occluded face recognition could be efficiently dealt with using the block (modular) representation approach [26]. Contiguous occlusion can safely be assumed local in nature in a sense that it corrupts only a portion of conterminous pixels of the image, the amount of contamination being unknown.
This is to say, the occlusion is not scattered over the images so that if we divide the face images into several blocks called sub-images, some of them remain clear and support distinctive information. Thus, local approaches try to extract meaningful partial facial features that can eliminate or compensate the difficulties brought by occlusion variations. Some previous algorithms [25, 26] processed each block individually and made the final decision by fusing each separate recognition result using majority voting or other principle. However, a major pitiful with majority voting is that it treats noisy and clean partitions equally. In this way, it is likely to be erroneous no matter how significant the clean partition may be in the context of facial features. To address this issue, a modular processing that a simple but effective approach is presented for handling the occlusion and corruption problem in this paper. The idea is to segment the face image into M non-overlap blocks uniformly (M=8 in this study [25,26]). Then an occluded portion detection method based on block-linear regression classification (B-LRC) is proposed for removing the occluded blocks.
For partial occluded face recognition scenario, we assume there are N classes and pi training samples in the i-th class, i = 1,2,…, N. We transform images into vector (k=1,2,…,M, j=1,2,…,pi), the vector of the k-th block from the i-th class can be described as
Based on the analysis of LRC, we still assume that the sub-images of the same block belonging to a subject lie on a certain subspace of the corresponding portion. Thus, given an input face image y belonging to the i-th class, the following equation is approximately reasonable.
To a specified block (sub-image) yk of the input test image
As the fundamental assumption of LRC that samples from a specific object class lie on a linear combination of training images of class-specific samples, namely, LRC represents a test image as a linear combination of training images of class-specific. In other words, if the k-th block of the test image yk is contaminated by occlusion or corruption, it will not be represented as a linear combination of training images of class-specific. To evaluate the block based on the above analysis, the distance measure between the predicted response vector and the original response vector y is calculated as
where
For estimating which porion is contaminated by occlusion, an evaluator is defined as
Here indicates a corresponding minimum distance, and
In the case of a block without any occlusion, yk, can be represented very well by one linear combination from its corresponding class of training sample while models of other classes can not offer accurate representations. Therefore, the true class of the test sample could express the block image much better than the average of all the class, which indicates that this block has a big value of E(yk). While there are some blocks contaminated by occlusion, no class, generally, can accurately express this non-facial subimage or all classes offer approximately equally poor representation of this block, which means this block has a small E(yk). In summary, a block with big E(yk) means it is very likely that block of the test sample is clear, while small E(yk) indicates this block is occluded.
Thus, the r(r ≤ M) effective blocks with r largest E(yk), which clear correspondingly or without any occlusion are selected as
So far we have detected which block should be removed and the remain r blocks (r=4 blocks are selected in this paper based on experience) would be utilized for face recognition in next section.
3.3 Face Recognition Using Improved Distance-based Evidence Fusion (i-DEF)
In reference [25], authors made use of the specific nature of distance classification to develop a fairly simple but efficient fusion strategy which implicitly deemphasizes corrupted subimages, significantly improving the overall classification accuracy. They proposed using the distance metric as evidence of their belief in the “goodness” of intermediate decisions taken on the subimages, which called “Distance-based Evidence Fusion” (DEF).
Inspired by aforementioned work, an improved Distance-based Evidence Fusion (i-DEF) method is proposed to final face recognition issue in this study. Unlike DEF method, we have approximately eliminate the effect of occlusion by removing the contaminated portions based on analysis above in order that the recognition can well proceed on the remaining parts of the image. So only the DEF on the remaining blocks should be calculated rather than all of that in the image.
Furthermore, as analyzed in reference [25], l2-norm reflects the overall difference between two vectors, and it is sensitive to large errors. On the other hand, we consider that if the occluded portion in the test image yk is correctly removed, it can be well-approximated as a linear combination of of the real subject i, and the error between yk and should be sparse. This error may affect any part of the image and may be arbitrarily large in magnitude. It is worth noting that l1-norm is typically used to characterize the sparsity of vectors [25]. Therefore, we attempt to make use of l1-norm instead of l2-norm for classification. For verifying this suppose, we compare l1-norm with l2-norm in Section 4.2.
So the l1-norm distance measure between the r estimated and the original response vector is computed as
Now, for the k-th partition, an intermediate decision called j(k) is reached with average value of corresponding minimum distance calculated as
Therefore, we now have r decisions j(k) with r corresponding distances dj(k) and we decide in favor of the class with minimum average distance
The overall procedure of considerations above can be summarized as Algorithm.
Algorithm. Block-Linear Regression Classification with improved DEF (i-DEF).
Inputs: Class models ∈ IRq×pi, and a test image yk ∈ IRq×1 (i = 1,2,…, N , k = 1,2,…, M, j = 1,2,…, pi).
1. is evaluated aginst each class model,
2. is computed for each
3. Distance calculation between original and predicted response variables di(yk) =
4. Evaluator is computed as
5. The r(r ≤ M) effective blocks with r largest E(yk) are selected
6. l1-norm distance measure:
7. Compute average value of corresponding minimum distance dj(k)
8. end for
Output: Decision = arg min(i)dj(k)
4. Experimental Results and Analysis
The partial occluded face recognition system we designed in this study is running on the hardware environment of Intel (R) Core (TM) i5 (3.40 GHz) with 12.00GB RAM, and the software environment of Windows 7 and Matlab R2012b.
Extensive experiments are carried out to illustrate the efficiency and robustness of the proposed approach in this study. All images are divided into 8 blocks (4*2) [25, 26] and L1LS [28] approach is used for solving the l1-minimization problem for proposed i-DEF.
4.1 Complexity Analysis
The computational cost of the algorithm is an essential problem for face recognition application in real life. The core of SRC is a l1-minimization problem, which has shown good potential in handling occlusions that are spatially correlated, e.g., random pixel corruption. However, it is not robust to contiguous occlusion, e.g., sunglasses and scarf [26]. Although eSRC [25] has satisfactory performance to identify occluded subjects, it also has an obvious disadvantage as some useful information is lost caused by occlusion, the highest possible resolution version of images is used to exploit more information. However, this makes input image be a very large matrix. For instance, assume that the size of face images is 200*150 and 1000 images are used for training samples, so X ∈ R(a×b)×(q+n) (a × b = 200 × 150, n = 1000) is larger than X ∈ R(a×b)×n (SRC). Since l1-minimization solver [29] with the empirical computational complexity of SRC is O((a×b)2nε), where ε ≈ 1.5. When dealing with eSRC, its complexity is O((a×b)2(a×b+n)ε). It is very clear that eSRC has much higher computational cost than SRC so that it is restricted in reality for time-consuming even using modern high-end devices.
Compared with SRC and eSRC algorithm, our proposed method not only does not add much computational cost, but also can handle the problem of contiguous occlusion. As mentioned in section 3.1, face images are first downsampled to q-dimension (q = c×d, c×d ≪ a×b). In our method, yk is transformed into a new subspace by H(k) and H(k) will not be changed with the probe images. The off-line computational complexity for all H(k) is O(NMn2). To a probe image, the on-line cost for determining the clear blocks is O(NMn). Therefore, the whole procedure has the approximate complexity with O(NMn + (c×d)2nε). The computational cost of classification by proposed method will decrease correspondingly due to removing the contaminated blocks.
4.2 Partial Face Recognition with Random Block Occlusion
For this experiment, we test the effectiveness and robustness of the proposed method against random block occlusion. Extensive experiments are carried out using the Extend Yale B database [30]. The database consists of 2414 frontal face images of 38 individuals under various lighting conditions. The database was divided in five subsets; subset 1 consisting of 66 images (seven images per subject) under nominal lighting conditions was used as the gallery, while all others were used for validation (see Fig. 1). Subsets 2 and 3, each consisting of 12 images per subject, characterize slight-to-moderate luminance variations, while subset 4 (14 images per person) and subset 5 (19 images per person) depict severe light variations.
Fig. 1.Five Subjects of Extended Yale B database, each row illustrates sample from Subsets
All experiments for the proposed method are conducted with images downsampled to an order 20 × 20. Subset 1is selected for training, and Subject 2 and 3 for testing, respectively. We simulate various levels of random contiguous occlusion, from 10% to 50%, by replacing a square block of each test image with a “baboon” image. The location of occlusion is randomly selected for Subset 2 and 3 (see Fig. 2).
Fig. 2.Test images under varying level of contiguous occlusion with baboon image, from 0% to 50%
Fig. 3 represents the recognition accuracy with random block occlusion under Subset 2 for testing based on different algorithms, including Eigenfaces+NN, Fisherfaces+NN, SOM, LRC, SRC, eSRC and proposed method. The proposed method comprehensively outperforms the other methods for all levels of occlusion. Up to 30% occlusion in this Subset, our method performs correctly identifying of 94.39%, 1.83 percent more than SRC and approximately equal to eSRC. In the case of 40% occlusion, the proposed method receives 87.73% correct rate, it is better than any these method. Even at 50% occlusion, none of these methods achieves higher than 60% recognition rate, while the proposed method achieves 65.4%.
Fig. 3.Comparison result of random block occlusion on subset 2 of Extended Yale B database
Fig. 4 shows the recognition rate of all compared algorithms using subset 3 for testing, the test images are under more various illumination conditions apart from random occlusion. The proposed method significantly outperforms the other seven methods for all levels of occlusion. The recognition rates of other methods drop to 90%, while our proposed method achieves 93.79% recognition rate and much better than SRC and eSRC. Even in the case of half face occluded, proposed method still achieves 65.4% recognition rate. Based on these two experiments, our proposed method is verified for effectiveness and a good tolerance on random occlusion problem. On the other hand, the proposed method with l2-norm might achieve good performance till 20 percent occlusion, the recognition rate drops drastically from 30 percent occlusion. Therefore, we will use the proposed method with l1-norm for extensive experiments later.
Fig. 4.Comparison result of random block occlusion on subset 3 of Extended Yale B database
4.3 Partial Face Recognition with Real Face Disguises
In this section, we test proposed method on AR database [31], which consists of over 4000-color image of 126 subjects (70 males and 56 females) with size of 576 × 768. The database characterizes divergence from ideal conditions by incorporating various (neutral, smile, anger and scream), luminance alterations (left light on, right light on and all side light on), and occlusions modes (sunglass and scarf). There are 26 different images per person, recorded in two different sessions separated by two weeks, each session consists of 13 images. In each session, there are 3 images with different illumination conditions, 4 images with different expressions and 6 images with different facial disguises (3 images wearing sunglasses and 3 images wearing scarf, respectively). It has been used by researcher as a testbed to evaluate and benchmark face recognition algorithms.
As the problem of face identification in the presence of contiguous occlusion is arguably one of the most challenging paradigms in the context of robust face recognition. Commonly used objects, such as cap, sunglasses and scarves, tend to obstruct facial features, causing recognition errors. Therefore, we focus on sunglasses and scarves occlusion in this experiment for verifying the effectiveness of the proposed method. The AR database consists of two modes of contiguous occlusion, images with a pair of sunglasses and scarf, see Fig. 5. We only used a subset of the original database, which includes randomly selected 100 individuals (50 males and 50 females). For each subject, 8 images without occlusion are used as the training samples for recognizing the images with sunglasses and scarf. All images are cropped and downsampled to 40 × 50 pixels according to the eye positions (see Fig. 5 (b)).
Fig. 5.Samples of AR database from one subject. (a) Original images from AR database, (b) Cropped images for training (first two rows) and testing (third row)
Table 1 depicts a detailed comparison of the proposed method with the other six representative and state-of-the-art algorithms, consisting of Eigenfaces+NN, Fisherfaces+NN, SOM, LRC, SRC and eSRC. For the scenario of sunglass occlusion, this occludes roughly 25 percent of the whole face image. Our proposed method achieves a successful recognition accuracy of 98.5 percent in a 200D feature space, which outperforms eSRC by a margin of 10 percent approximately. In addition, the other scenario considers images with the subject wearing a scarf, which occludes roughly 40% of the face image. It is easy to find that recognition accuracy drops substantially except our proposed method. The proposed method achieves the recognition rate of 78.9%, outperforming LRC, SRC and eSRC by 36.3%, 17.4% and 12.1%, respectively. It indicates that our proposed method is reasonable and effective. By removing the occluded blocks exactly, the remained blocks can represent useful information for recognition issue.
Table 1.Recognition results (%) for sunglasses and scarf occlusion
For further verifying the effect of feature dimension to recognition accuracy, we downsample the face images to 10 × 10 , 20 × 25 , 25 × 30 , 35 × 40 and 40 × 50 , respectively, which correspond to feature dimension of 100, 500, 750, 1400 and 2000. The comparison results are shown as Fig. 6 and Fig. 7.
Fig. 6.Recognition rates of different methods on a subset of the AR database with the feature dimensions varying from 100 to 2000 for sunglass occlusion
Fig. 7.Recognition rates of different methods on a subset of the AR database with the feature dimensions varying from 100 to 2000 for scarf occlusion
From Fig. 6 and Fig. 7, global feature-based methods, such as eigenfaces+NN and fisherfaces+NN method cannot achieve a good performance on partial occluded face recognition issue under low dimension features. When images are downsampled to low dimension (such as 100D and 500D), the significant information of face is lost as well. eSRC algorithm has a better performance as it uses the raw image as input vectors, significant information is exploited fully rather than r “clear” blocks in the proposed method. However, the proposed method can handle this problem and achieve better performance with much faster than SRC and eSRC by using low dimension images. Even the images are downsample to 750D feature, the proposed method still outperforms SRC and eSRC by a margin of 8 percent and 4 percent approximately in the case of scarf occlusion. In other words, the proposed method might be suitable to corruption problem as well.
For verifying the performance of i-DEF, a comparison experiment is also conducted by using DEF and i-DEF, respectively, for partially occluded face recognition issue. We also used a subset of the original database of AR, which includes randomly selected 100 individuals (50 males and 50 females). For each subject, 8 images without occlusion are used as the training samples for recognizing the images with sunglasses and scarf. All images are cropped and downsampled to 40 × 50 pixels according to the eye positions like Fig. 5 (b). Moreover, the occlusion detection results (r=4) are utilized for DEF and iDEF, simultaneously, to test the recognition rate of these two algorithms. The experimental results on AR database are expressed as Table 2. From Table 2, the proposed method outperforms the B-LRC with DEF algorithm by a margin of 1.8 percent and 14.6 percent in case of occlusion by sunglasses and scarf, respectively. It is easy to find that recognition accuracy drops rapidly by using DEF when scarf occlusion occurred, the proposed method could handle this issue better by using i-DEF. Because if the occluded portion in the test image y is correctly removed, it can be well approximated as a linear combination of of the same subject, and the error between yk and reconstruction vector should be sparse. This error may affect any part of the image and may be arbitrarily large in magnitude. l1-norm is typically used to characterize the sparsity of vectors. On the other hand, l2-norm reflects the overall difference between two vectors, and it is sensitive to large errors, as analyzed in the literature [25].
Table 2.Recognition results (%) by using DEF and i-DEF
Comparisons results indicate the effectiveness of removing the contaminated portions and the recognition algorithm could well proceed on the remaining parts of the image. The utilization of l1-norm instead of l2-norm for classification is reasonable based on analysis above. Combing with the analysis in section 4.1, the proposed method is with low computational cost and high efficiency. It is reasonable to believe that the proposed method can handle robust face recognition in practical application due to its advantage corresponding to other methods.
5. Conclusion
In this paper, a novel method based on linear regression-based classification (LRC) algorithm is proposed in this paper. First, all images are downsampled and divided into several blocks, and then we exploit the evaluator of each block to determine the clear blocks of the test face image by using linear regression technique. Next, the remained uncontaminated blocks are utilized to partial occluded face recognition issue. Furthermore, an improved Distance-based Evidence Fusion approach is proposed to decide in favor of the class with average value of corresponding minimum distance.
The experimental results based on proposed method that performed on AR face database and extended Yale B database show the effectiveness and robustness. Based on these experimental results, we can conclude that the proposed method has better performance on the issue of partial occluded face recognition than the conventional algorithms such as eigenface+NN, fisherfaces+NN and the state-of-the-art method such as LRC, SRC and eSRC. Moreover, the computational cost of classification by proposed method will decrease correspondingly due to removing the contaminated blocks, and the i-DEF can achieve satisfactory capacity of classification.
Additionally, the issue of frontal face recognition is concerned in this paper; we would like to expand our face recognition system for more comprehensive situations in future work, to enhance its practicability for real world.
Cited by
- Misclassified Samples based Hierarchical Cascaded Classifier for Video Face Recognition vol.11, pp.2, 2015, https://doi.org/10.3837/tiis.2017.02.009