A Novel Algorithm for Face Recognition From Very Low Resolution Images

  • Senthilsingh, C. (Dept. of Electronics, Anna University) ;
  • Manikandan, M. (Dept. of Electronics, Anna University)
  • Received : 2014.07.13
  • Accepted : 2014.10.01
  • Published : 2015.03.01


Face Recognition assumes much significance in the context of security based application. Normally, high resolution images offer more details about the image and recognizing a face from a reasonably high resolution image would be easier when compared to recognizing images from very low resolution images. This paper addresses the problem of recognizing faces from a very low resolution image whose size is as low as $8{\times}8$. With the use of CCTV(Closed Circuit Television) and with other surveillance camera-based application for security purposes, the need to overcome the shortcomings with very low resolution images has been on the rise. The present day face recognition algorithms could not provide adequate performance when employed to recognize images from VLR images. Existing methods use super-resolution (SR) methods and Relation Based Super Resolution methods to construct from very low resolution images. This paper uses a learning based super resolution method to extract and construct images from very low resolution images. Experimental results show that the proposed SR algorithm based on relationship learning outperforms the existing algorithms in public face databases.


Face recognition;Face super-resolution (SR);Relationship learning;Very low resolution(VLR)

1. Introduction

The face recognition research is not in its infancy and it has been more than 36 years since researchers have started solving the problem of recognizing faces. Day by day, new application that involves face recognition are being deployed. As the research has been pursued on this field for a very long time, the researchers have started believing that saturation has reached in this field. Immediately it is realized that with security based application like surveillance camera and CCTV, people realized that the need to find the solution has gathered momentum. In order to maximize camera viewing and covering an area, wide area cameras are used. The relative area covered by the face image of the person is very small compared to the total area of the image and in most of the cases, especially when the person is at long distance, the pixel resolution of the image would be even less that 8 × 8. Working on such a very low resolution (VLR) face image is called a VLR face problem.

Minimum size required to reconstruct an image from a low resolution is found out to be between 32× 32 and 64× 64 [17]. It further states that any image which is found to be less than 32 × 32 will have less detail, such as the recognition performance be reduced [9]. Fig. 1(a), (b) gives a complete scenario of the stated problem.

Fig. 1.Courtesy: Typical frame from a surveillance-CCTV camera. The persons seen are stealing from the shop: (a) Surveillance video; (b) Face region.

Fig. 1 (b) clearly elucidates the amount of details lost when the face region of the image is zoomed. This VLR problem can be solved if we could supply the missing details in the image, Therefore, super-resolution (SR) algorithms has been employed. [3, 4, 6, 7]. Super resolution employs two methods of reconstructing the image from the VLR data base. The reconstruction of images based on super resolution, because of the self-similarity of the face images, learning-based SR algorithms are always used to enhance the resolution of the face images.

In the literature, different kinds of super resolution algorithms [1, 5, 8-10, 15, 16, 18, 19] have been developed. Many of the stated algorithms are based on two approaches: maximum a posteriori (MAP) and example-based approaches. Error for the reconstructed image is measured using data constraint [15].

In a surveillance system, the resolution of face image is normally low in the video. Hence the details of facial features which are available in the respective high resolution image are very important to compare and identify the face from the low resolution images. Hence, a LR image is obtained from HR images by smoothing and down-sampling. It is expected that it must be close to the input LR image. Therefore, this constraint called data constraint is introduced. Here, the images are compared by calculating thee distance in the low-resolution (LR) image space for SR processing. For example, Baker and Kanade’s method [1] copies the high-resolution ( HR) details from HR training images, which are the best match in the corresponding LR training images, as the missing details in the input query LR image.

There are other constraints to be looked after to get a good match in the case of very low resolution images. These constrainnts are stated clearly in [15]. (a) Global constraint. The result must have common characteristics of a human face, e.g. eyes, mouth and nose, symmetry, etc. The facial features should be coherent. (b) Local constraint. The result must have specific characteristics of this face image, with photorealistic local features. The global face similarity constraint and local constraints are important to avoid noise and to have match with ordinarry facial features.

For this method to be successful, thee distance metric calculated in the VLR image space must reflect the actual face similarity in the HR image space. However, the studies based on empirical results [17] show that the face similarity measure in the VLR image space cannot reflect the actual face similarity of HR face images. As such, many SR algorithms in the current literature may not be successful to solve the VLR problem.

Later, VLR problem was given a new approach namely relationship-learning-based SR (RLSR) approach [20]. The RLSR algorithm has significantly reduced the effect of the VLR problem. RLSR algorithm is a type of algorithm which defines a mapping (called relationship) from the VLR image space to the HR image space for HR image reconstruction. Humans can identify the images and machines may also need to identify the images. Each requires different approaches. For humans to identify the image, the image quality needs to be good. In order to achieve that, RLSR algorithm has introduced a new data constraint and in case of machine based recognition, a discriminative constraint is used as to enable the machines to identify the images. It is based on discriminative features.

In Fig. 2(a) the circles represent images. Existing algorithms evaluate the error in the LR image space by an old data constraint; the proposed method makes use of a new data constraint that does measurement in the HR image space. In Fig. 2(b) clearly gives an overview of the RLSR algorithm. This algorithm makes use of the information that may be available during the training phase. SR and LSR methods normally recover the images directly from the VLR image(s) and the training images. But, the RLSR algorithm does the following:

Fig. 2. (a)Existing methods.

Fig.2. (b)Propposed method

First, a relationship VLR and HR image spaces is established. Secondly, the outcome is applied to recover the high resolution images from the low resolution images.

In this paper, a few changes have been incorporated to improve the performance of the RLSR algorithm and better results are obtained.

In order to establish the relationship in HR image space, regression analysis is done. The error measurement in HR image space is found using the new data constraint and is introduced with the help of regression model. Ordinary least squares is used. This method minimizes the sum of squared vertical distances between the observ ed responses in the dataset and responses predicted by the linear approximation. As such the proposed method recovers images with more details and handles the very low resolution problem better.

K-means clustering algorithm is proposed to avoid the over fitting effect of the Ordinary least squares method. This new clustering algorithm maximizes the linearity of the data in each cluster. This clustering ensures data linearity in each cluster. Therefore, the linearity−clustering-baseed relationship learning method can handle the complex nonlinear case.

In order to categorize and to enhance the differentiability of the face image, a categorizing constraint is integrated in the optimization function with the help of class label information


2. Related Works

Many super resolution methods have been introduced in the literature SR [1, 9-13, 15, 16, 18 -20] to meet the growing demand and challenges of the face recognition applications. Almost all the stated SR algorithms are machine learning based. The machine based approach basically does the following: Given a set of training LR-HR image pairs and an input query LR image, the SR algorithm reconstructs an HR image from and traing data

These algorithms normally follows two approaches, namely, MAP and example-based. The reconstructed image in both the method is verified with the help of data constraint Both approaches employ the same data constraint to verify the reconstructed HR images. But the inherent shortcomings that exist in the data constrain creates many problems and hence, it may not work with the VLR problem. In order to overcome this difficulty, the association relation based approach learns a relationship between the input VLR image space and HR images which is explained in the Fig.2(b). Unlike most of the existing methods recovering the images directly from the VLR image(s) and the traing images, our method first determines relationship between VLR and HR image spaces and then applies on VLR images to recover the HR ones. Very low resolution face recognition problem by W. W. Zou and P. C. Yuen [20], formed where the VLR problem was introduced and a new ALSR algorithm was proposed to alleviate it. Super-resolution from a single image [7] proposes a unified framework for combining the two general approaches can be applied to obtain super resolution from as little as a single image (with no database).

Learning the relationship between high and low resolution images in kernel space for face super resolution [20] proposes a new nonlinear face super resolution algorithm to address recognition of low resolution face image with nonlinear variations. The proposed method learns the nonlinear relationship between low resolution face image and high resolution face image in (nonlinear) kernel feature space. Super resolution imaging: A survey of current techniques [3] explains an insight into the state-of-the-art Super Resolution techniques and illustrated comparative experiments on real data that portrayed the robustness and utilization of all the methods.

2.1 Super resolution based on MAP approach

In this approach, the HR image is constructed via finding the HR image by maximizing conditional probability. The Probability of Ih probided Il has occurred P (Ih | Il)

Generally P(Il | Ih) is modeled by the Gaussian distribution, so that maximizing P(Il | Ih) is equivalent to minimizing the data constraint [15] as follows: Ih is the input high resolution image and the very low resolution image Il is constructed by means of the Down sampling operator D. Each Low Resolution pixel is a linear combination of HR pixels as shown in Fig. 3.

Fig. 3.SR based on MAP Approach

Hence it is imperative that in order to minimize the probability, it is enough to minimize the data constraint || DIh - Il ||

Different SR methods uses different techniques like to calculate the Ih

2.2 Example based approach

In this method, the prior knowledge is used to reconstruct a HR image. It constructs tthe HR image by a linear combination of the HR training images.

Mathematically, the linear combination of LR images can be written as

where w represents the weights and other Ihi represents the High resolution images from training pairs. We would like minimize the Euclidean distance between the reconstructed image and trained HR images. is to be minimized. But we know that is obtained by down sampling the input High resolution image shown in Fig. 4. Therefore, is to be minimized.

Fig. 4.Example based Approach

The greatest drawback of the method is that the large magnification factor causes overly smooth HR images which ultimately results in loss of fine details. Some of the popular example based methods are: Hallucination (Simon Baker and Takeo Kanade), Example Based, Single Image Super Resolution.


3. Proposed Method

An overview of the proposed method is shown in Fig. 5. The shortcomings those arise in the previous methods are mainly due to their inability to derive an association between the HR image space and the low resolution image space. Hence, this paper first attempts at deriving a relationship between HR and VLR image spaces. An association is learned in the training phase and is used to reconstruct the HR images in testing phase. The training phase consists of two steps, namely, modified centroid K means clustering and relationship learning. In the first step, a clustering algorithm is propposed as a pre-processing step.

Fig. 5.Block diagram of the association based learning algorithm

3.1 Association learning based super resolution

Let the size of the image be m × n. Then we will have exactly mn pixels.

The super resolution is the process of obtaining single image from multiple images of the same dimension by means of a linear combination of the given set of images, The super resolution image obtained from given set of high resolution image {h1, h2, h3, … hi,… hn} resolution images of the given person with different exposures is given by the following equation:

Let h1, h2, h3, … hi,… hn be the n different high Let a1, a2, a3,…. ai, …an be the corresponding averages of luminance values of the given n images. Then,

Total luminance value of the pixel =

The average value of the pixel is calculated to calculate the appropriate value of the weigh to be given to each high resolution images hi I = 1, 2, 3,….. n

Based on the average value, the weights w1, w2, …………. Wi,…. Wn are calculated.

It is well known facts that down sampling and stretch are linear operators.

Let D be theddown sampling operator that acts on each set of high resolution images to produce corresponding set of very low resolution operator. The down sampled images are given by

The super low resolution image obtained from given set of low resolution imags { l1, l2, l3, …… li,… ln } is given by

The same weight is used here to obtain the SR image for the low resolution image also.

From (5)


Since D is a linear operator, we can rewrite (4) as

Since Dhn = ln, the high resolution can be obtained by stretching the corresponding low resolution image.

From (6) we can understand that Super resolution of VLR images can be obtained from the SR image of all high resolution images by just down sampling it.

Let dhi = Shi – Hi, I = 1,.2,………n be the difference image between the super resolution image and HR image hi.


For a single individual, we would be having a set of images, which would be conpriseing of all its low and high resolution images. We consider this set as one cluster and (Si, Sh) would be its center.

Let R denotes the association relationship map from VLR – HLR training images.

For each cluster, we denote the training image pairs by

For each new input VLR – HLR mage for that group, the new found association map is applied to the VLR image and the corresponding high super resolution images is found and is compared with the available HLR image and mean square error method is used to reduce the reconstruction error.

The reconstruction error is calculated,

And it is to be made sure that it is minimum and which would help us learn the relationship

3.2.Linearity clustering

Improved k clustering is employed here, by which the linear relationship, that is expected to be there between the input and corresponding HLR is ensured in the clustered training image pairs

3.2.1 Modified centroid k means clustering algorithm

A cluster is a collection of pixels which are similar between themselves. Clustering the image is the process of grouping the pixels according to some characteristics such as pixel intensity, distance criterion etc. Clustering is the most important unsupervised learning problem. Here the algorithm classifies object automatically based on the criteria given by the user K-means clustering algorithm is an exclusive clustering algorithm in the sense that if a certain pixel belongs to a definite cluster then it could not be included in another cluster.

In the K-means clustering algorithm, initially we have to define the number of clusters k. Initial cluster centers have to be found. In the standard K- means clustering algorithm, initial cluster centers are found in random but here, in our work, a new method for defining initial cluster centers is proposed. Difference between the maximum and minimum pixel of an input image is divided by the number of clusters k which gives an interval. Initial centers for the clusters are chosen using this interval such that each cluster center varies with a fixed range among themselves. After finding initial cluster centers the distance between each input image pixel to each cluster centers are calculated. The distance is calculated by finding the absolute difference between a pixel and a cluster center. The pixel is moved to particular cluster which has shortest distance among all. Then the centroid is re-estimated. It is done by finding the average of the pixels in the clusters got. Again each pixel is compared against each centroid. The process continuous until the center converges. The main aim is to minimize an objective function.

Modified Centroid K Means Clustering Algorithm Steps: 1. Let y1, …, yM are L data points in the input image or input dataset, k be the number of clusters, and w1,…, wK be k cluster centers. 2. Find the maximum and minimum pixel or data point from the input image or input dataset. 3. Their difference is found and it is divided by K clusters which gives an interval. 4. In a loop determined by number of clusters using the interval got from step 3, find the initial centroids of the k clusters such that each cluster center (centroid) varies with a fixed range amongst them. 5. Distance between each pixel and each cluster center is found. 6. The distance function is given by J=| xi - cj| For i=1,…, N and for j=1,…, k Where | xi - cj| , the absolute difference of the distance between a data point xi and the cluster center cj is an indicator of the distance of the N data points from their respective cluster centers. 7. Distribute the data points x among the k clusters using the relation y∈ Cj if |x - cj|<|x - ci| for i=1, 2,…, k, i ≠ j, where Cj denotes the set of data points whose cluster center is cj. 8. Updated cluster center is given as, , for i=1,…, k, where mi is the number of objects in the dataset Ci, where Ci is the ith cluster and ci is the centre of cluster Ci. 9. Repeat from Step 5 to Step 8 till convergence is met (old and new centroids remain same).

Algorithm Association Learning relationship mapping R from training data

Require: training image pairs terminating condition ∈, initial relationship R(0), step α0 1: n ← 1 2: d(0) ← (𐐀ℰ(R(0))/𐐀R)T 3: while ||d(n – 1)|| >∈ do 4: find the optimal ᾶ ∈ [ 0 α0 ] that minimizes f(α) = ℰ(R(n − 1) − αd(n − 1)) 5: update : R(n) ← R(n − 1) − ᾶd(n − 1) 6: n ← n + 1 7: end while 8: return R(n) as the learned relationship mapping


4. Experiments

4.1 Performance analysis of VLR algorithm

Face Recognition appears to be relatively a trivial task; however, the task contains a lot of challenges. In this paper, the algorithm developed for VLR problems are designed to address those challenges and come out with successful results. Since Extracting image from VLR and its success largely depends on learning process.

For human based visual detection and machine based face detection, this paper has suggested two different methods and the experiments are conducted to evaluate the performance of the algorithms. The following popular methods that have been used in the literature are taken to compare the performance of the proposed methods.

1. Hallucinating face (HLF), 2. Eigen transformation-based face SR (ETF), 3. Kernel prior face SR (KPF), 4. Position face hallucination (PFH), 5. Relation based face recognition [RLSR]

The study of the outcomes of the proposed method confirms that it surpasses the existing methods in terms of visual quality and recognition performance.

4.1.1 Experiment for human based face detection

Super resolution is the process of constructing the high resolution image from many very low resolution images. As explained before, the association relation that could exist between the high resolution image space and very low resolution image space is used to construct the HR image from the very low resolution query image (obtained from videos of CCTV etc). Then, quality assessment of the reconstructed HR image is done. Normally, the visual quality assessment is categorized into two: first assessment is based on human and is called subjective visual quality; second one is objective visual quality, which is based on any statistical measure. Here, in this paper both methods are employed to assess the visual quality of the reconstructed image; first method employs the opinion of the humans and second one uses mean squared error.

Database and Settings: Public face databases taken and our own created database containing images of hundred people with different expressions and exposures are used for the experiments. From the database, a subset that consists of 500 frontal view face images (one image per individual) is used. From Indian database, a subset of frontal view images of 38 persons with 64 different illuminations is used. All images are manually aligned by the position of the eyes and normalized to the resolutions of 64× 48 (HR) and 8 × 8(VLR).

Experiments are conducted on 2 databases, one public database and another locally created. The images in the two databases are divided into two disjoint sets. Division is carried out based on their class labels. The 200 images from Indian databases are randomly selected for training and classifying and the rest of the data are used for testing purpose. Similarly a data size of 50 images from our own database is kept for training and the remaining 50 images are used for testing purposes.

The high resolution images and the corresponding low resolution images are clustered using k means algorithms. For each input image with different exposures and different resolutions, the corresponding LVR images are created and the low resolution images are super resoluted to form a HR image and in the process, an association is learned. Thus each cluster will represent a particular kind of association and that association is very much used later to identity where the input low resolution images are to be associated in the testing phase

Thus, the VLR testing image will be classified into one of the clusters, and then, the reconstructed image is obtained by applying associated learning. Ma, J. Zhang, and C. Qi, suggested one way of improving the quality of the image in which each image is divided into small patches, and super resolution is done patch by patch and is overlapped with its neighboring pictures to avoid the blocky effects.

Results: The very low resolution images from our own database and Indian database are taken and the proposed algorithm is applied on these images and the performance of the algorithm is compared with other prominent algorithms in the literature and the results are shown in the Figs. 6, 7. Fig. 6 shows very low resolution input images of several persons taken from our database and their corresponding output images by different methods and the proposed method. Fig. 6(a) and 6(h) show the input very low resolution query image and its original HR image, respectively. Fig. 6(b)-(f) displays the results using the various methods (b) PFH. (c) HLF. (d) ETF. (e) KPF. (f) RLSR. (g) Proposed ALSR.

Fig. 6.Super Resolution results on the OUR-OWN database. Very low resolution Images of different persons: (a) Input VLR image; (b) PFH; (c) HLF [1] ; (d) ETF; (e) KPF [2] ; (f) RLSR; (g) Proposed ALSR; (h) Original HR image.

Fig. 7.Super Resolution results on the Indian database with different positions: (a) Input VLR image; (b) PFH; (c) HLF [1]; (d) ETF; (e) KPF [2]; (f) RLSR; (g) Proposed ALSR; (h) Original HR image.

Most of the existing face recognition methods are incapable to unfailingly identify persons or worse it will recognize the incorrect person. Face Images of same person with different expressions, postures, and exposures have to be promptly recognized. It is highly ideal to get the images which are center, portrait type environment for the system to identify the faces correctly. Getting such images are highly improbable. This paper has introduced methods that overcome the above said difficulties and experimental results shown in Figs. 6, 7 greatly corroborate our claim.

Fig. 6 shows how our algorithm produces better results for those very low resolution images of sizes as low as 8× 8 of different persons and reconstructs the corresponding high resolution images. Reconstructing the face images is quite challenging and is one factor that highly affects recognition performance of any algorithm. The Fig. 6(a) shows the original VLR images. Figs. 6(b) to 6(f) represent the reconstruction results of the existing methods (HLF) [1] (ETF), (KPF) [2], (PFH) [18], [RLSR] respectively which are taken for comparison. Fig. 6(g) shows the reconstructed HR image from the corresponding VLR image.

Fig. 7 shows how our algorithm produces better results for those very low resolution images of sizes as low as 8 × 8 of with different expressions and reconstructs the corresponding high resolution images.

From the Figs. 6, 7, it is well understood that for a very low resolution image of size 8 × 8, KPF method gives relatively blurred image. With the size of VLR images with low dimensions as 8 × 8 the algorithm fails to produce many high frequency details. The visual quality of the reconstructed image by ETF method is reasonably good. But, high resolution images produced by this algorithms lack the look and feel of the original High resolution images as the some important high frequency details are lost. The reconstructed images by the HLF method are better when compared with KPF method but it still has some blurring effect. The reconstructed images produced by ALSR method are better in quality and appearance when compared with all other four methods but it lacks clarity when compared with the original image. The method proposed in this paper can easily recover the images with face variations and lighting variations. When compared other existing methods which are taken for comparison, this algorithm produces better results and reconstructed images are close to the original HR images. Moreover, the times taken by PFH and RLSR algorithms to reconstruct the original images are 43 s and 30 s respectively. But, our proposed method is implemented in Matlab and FPGA, and it takes 24 s when implemented in matlab and 18 s when implemented in FPGA.

Moreover, Table 1 shows the Root Mean Square Error (RMSE) values of different super resolution methods for our own database and Indian database. The recorded values clearly show that our algorithm has low RMSE and hence it performs better when compared to every other SR methods under consideration. Fig. 8 shows the real time implemention of face recognition in FPGA.

Table 1.Root Mean Square Error(RMSE)

Fig. 8.Real time setup for face recognition in FPGA

4.2 Experiment 2: Recognition performance

This part of this paper describes the superiority of the proposed algorithm in terms of its performance. The performance is measured by measuring the recognition accuracy. The recognition accuracy is evaluated by means of false acceptance rate (FAR). The FAR means that at what rate the system accepts incorrect input as a positive match. For every ten thousand inputs, an error of I false recognition is accepted. Due care need to be taken to ensure that our system has very low false rejection rate also. For this part of the experiment, the images from our own database, Indian database, GTAV and LFW (labeled Faces in the wild) database were taken. The reason why various databases used in this experiment are: variations in face positions, variations in expressions, variations in illuminations, variation in clothing, variations in ethnicity, and variations in background. The experimental results are organized as follows: Part A reports the performance of Distinguishing-Super-Resolution method on our own database and Indian database using face recognition algorithms, namely, PCA and eigen faces algorithm, a combination of PCA and LDA algorithm, face recognition based on SVM machines is shown in Fig. 9; Part B investigates the performance of Distinguishing-Super- Resolution-method on query images with different (low) resolutions is shown in Fig. 10; Part C further evaluates the proposed method on Extended Yale Face database, CMUPIE and surveillance camera face (SCface) databases in Figs. 11, 12.

Fig. 9.Recognition results on the created Database with PCA and INN face recognition Engine: (a) CMC; (b) ROC

Fig. 10.Recognition results on the Indian Database with PCA and INN face recognition Engine: (a) CMC; (b) ROC

Fig. 11.Recognition results on the created Database with PCA and SVM face recognition Engine: (a) CMC; (b) ROC

Fig. 12.Recognition results with the Kernel L DA + INN face recognition engine: (a) CMC; (b) ROC

Databases and Settings: The databases used for this experiment are our developed database, Indian data base, CAS-PEAL face database, Extended Yale Face database. Extended Yale Face Database B is mainly used for images with different poses and different illumination conditions. It contains 16128 images. For CAS-PEAL database, 20 images with a frontal view, 20 images with different pose angles, 20 images with different facial expressions like neutral, eye closing, frown, smile, surprise, and mouth open are used. The database contains nearly 1040 face images and a subset of the database is used for this experiment. The same is repeated in the case of other databases also. Only the frontal view images from the CMU-PIE database are used. A part of the image set is used as the training set and another portion of the image set is used in the training phase. The test results are very encouraging for those frontal images and for those images with face variations, differing poses, smiles, etc,

Results and Analysis: Five existing algorithms have been taken to compare the performance of the proposed method. The experiment is carried out using the reconstructed HR images evaluate the performance of the proposed method on a face recognition system; we carry out the experiments using the reconstructed HR images with existing face recognition engines. The VLR and original HR images are used as benchmark performance, and five existing SR algorithms namely PFH, ETF, KPF, HLF and RLSR are used for comparison.

Part A: describes the performance of our discriminative method on the Indian database: In this experiment In Fig. 9-18, face recognition algorithms have been employed to evaluate the performance of the algorithm. PCA + 1NN, PCA SVM, kernel PCA+1NN, kernel LDA + 1NN

Fig. 13.Results (ROC curves) of the proposed method on images with different resolutions. The resolution of the probe image for LR is (a) 76, (b) 14 12, and (c) 28 24.

Fig. 14.Recognition results (CMC curves) on the CMU-PIE database using different face recognition algorithms: (a) PCA 1NN; (b) Kernel PCA1NN; (c) PCASVM

Fig. 15.Recognition results (CMC curves) on the SCface database using different face recognition alg: (a) PCA 1NN; (b) Kernel PCA1NN; (c)PCASVM

Fig. 16.Recognition results (ROC curves) on the CMU-PIE db using different algo: (a) PCA1NN; (b) Kernel PCA1NN; (c) PCA SVM.

Fig. 17.Recognition results (ROC curves) on the CMU-PIE database using different face recognition algorithms: (a) PCA-1NN; (b) Kernel PCA-1NN; (c) PCA SVM

Fig. 18.Recognition results (ROC curves) on the SCface database using different face recognition algorithms: (a) PCA1NN; (b) Kernel PCA1NN; (c) PCA-SVM.

In PCA+1NN face recognition algorithm, PCA is used to extract the eigenfaces and will be used acurve which helps analyze the performance of the system and CMC-Cumulative match Characteristic as the feature for the 1NN classifier. 95% of the variance is used in the training phase of PCA. Fig 10a clearly reveals that recognition accuracy has come down markedly with VLR image when compared to the HR image. Marginal improvement is noticed when HLF and ETF algorithms are employed. RLSR performs better. The performance has dropped when other two existing algorithms are employed. Our method improves the performance and is evident from the figure and has improved it to a tune of 16%. Fig. 11 shows the performance when kernel PCA is used as the feature extractor. The other things remain the same and no noticeable difference is found. Further experiments are conducted with distinguishable face recognition algorithms and kernel LDA 1NN [14] and PCA SVM, as the face recognition algorithms. Kernel functions for Kernel LDA and State vector machine are Gaussian radial basis function and Inner product respectively. The results are shown in Fig. 12 and Fig. 13. The performance has seen a drop of 20% for VLR method has shown 6%.

A receiver operating characteristic (ROC) curve is curve is a curve which highly points the increase in the identification rate of the system. ROC Curve and the CMC are drawn and they depict how our method outperforms other method.


5. Conclusion

He very low resolution face recognition problem has been one of the most important face recognition problem and solution to such a problem has assumed much significance, for, in the changing requirements, the application based on this are wide and varied. Hence, this paper has defined and discussed the very low resolution problem, and proposed one good algorithm which used ordinary least square method of linear regression model and a novel association relationship based SR algorithm. Based on this idea, for godd visual quality applications, a new data constraint that measures the error in the HR image space was developed, and RLSR was proposed. For machine-based face recognition applications, a discriminative constraint was designed and integrated with the new data constraint, and DSR was proposed. Experimental results show that the proposed method outperforms the existing SR algorithms in terms of visual quality and recognition performance.


  1. J. S. Park and S. W. Lee, “An example-based face hallucination method for single-frame , low-resolution facial images,” IEEE Trans. Image Process., vol. 17, no. 10, pp. 1806-1816, Oct. 2008.
  2. W. Liu, D. Lin, and X. Tang, “Hallucinating faces: Tensorpatch super-resolution and coupled residue compensation,” in Proc. IEEE Int. Conf.Comput. Vision Pattern Recognit.,, 2005, pp. 478-484.
  3. Y. M. Lui, D. Bolme, B. A. Draper, J. R. Beveridge, G. Givens, and P.J. Phillips, “A meta-analysis of face recognition covariates,” in Proc.Int. Conf. Biometrics: Theory, Appl. Syst., 2009, pp. 1-8.
  4. X. Ma, J. Zhang, and C. Qi, “Hallucinating face by position-patch,” Pattern Recognit., vol. 43, no. 6, pp. 2224-2236, Jun. 2010.
  5. W. W. Zou and P. C. Yuen, “Face superresolution,” in Emerging Topics in Computer Vision and its Applications, C. Chen, Ed. Singapore: World Scientific, 2011, ch. 2.
  6. P. H. Hennings-Yeomans, S. Baker, and B. Kumar, “Simultaneous super-reeolution and feature extraction for recognition of low-resolution faces,” in Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit., 2008, pp.1-8.
  7. C. Liu, H. Y. Shum, and W. T. Freeman, “Face hallucination: Theoryand practice,” Int. J. Comput. Vis., vol. 75, no. 1, pp. 115-134, Oct. 2007.
  8. D. Glasner, S. Bagon, and M. Irani, “Super-resolution from a single image,” in Proc. IEEE Int. Conf. Comput. Vis., 2009, pp. 349-356.
  9. M. Grgic, K. Delac, and S. Grgic, “SCface-surveillance cameras face database,” Multimedia Tools Appl. J., vol. 51, no. 3,, pp. 863-879,, Feb. 2011.
  10. K. Jia and S. Gong, “Generalized face super-resolution,” IEEE Trans. Image Process., vol. 17, no. 6, pp. 873-886, Jun. 2008.
  11. K. Kim and Y. Kwon, “Single-image super-resolution using sparse regression and natural image prior,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 6, pp. 1127-1133, Jun. 2010.
  12. B. Li, H. Chang, S. Shan, X. Chen, and W. Gao, “Hallucinating facial images and features,” in Proc. IEEE Int. Conf. Pattern Recognit., 2008,
  13. H. Li, T. Jiang, and K. Zhang, “Efficient and robust feature extraction by maximum margin criterion,” IEEE Trans. Neural Netw., vol. 17, no .1, pp. 157-1665, Jan.2006.
  14. C. Liu, “Capitalize on dimensionality increasing techniques for improving face recognition grand challenge performance,” IEEE TransPattern Anal. Mach. Intell., vol. 28, no. 5, pp. 725-737, May 2006.
  15. A. Georghiades, P. Belhumeur, annd D. Kriegman, “From few to many: Illumination cone models for face recognition under variable lighting and pose,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 6 , pp. 643-660, Jun. 2001.
  16. S. Baker and T. Kanade, “Limits onn super-resoluution and how to break them,” IEEE Traans. Pattern AAnal. Mach. Inttell., vol. 24, no. 9, pp.1677-1183, Sep. 2002.
  17. A. Chakrabarti, A. N. Rajagopalan, and R. Chellappa, “Super-resolution of face images uusing kernel PPCA-based prior,” IEEE Trans. Multimedia, vol. 9, no. 4, pp. 888-892, Jun. 2007.
  18. G. Cristóbal, E. Gil, F.Šroubek, J. Flusser, C. Miravet, and F. Rodrıa-cute; guez, uperresolution imaging: A survey of current techniques,” in Proc. Adv. Signal Process. Algorithms, rchitectures, Implementations XVIII, 2008, vol. 7074, pp. 0C1-0C18.
  19. W. Freeman, E. Pasztor, and O. Carmichael, “Learning low-level vision,” Int. J. Comput. Vis., vol. 40, no. 1, pp. 25-47, Oct. 2000.
  20. W. Gao, B. Cao, S. Shan, X. Chen, D. Zhou, X. Zhang, and D. Zhao, “The CAS-PEAL large- scale Chinese face database and baseline evaluations,” IEEE Trans. Syst., Man, Cybern A, Syst., Humans, vol. 38, no. 1, pp. 149-161, Jan. 2008.