2
samples. Thus it is not a surprise that, despite of the impressive
results of regression classifiers and various extensions, many
works [11], [13]–[15] show doubts about their validity for
image classification.
To overcome the first constraint of regression classifiers,
query-adapted technique, also called distance weights learning
[16], provides an idea to separate outlier samples from the
dictionary atoms. A weights vector is obtained by computing
the distances from query sample to the whole training sam-
ples [17]. Accordingly, many developed weighted regression
algorithms benefit from this simple idea, including weighted
SRC (WSRC) [18], weighted CRC (WCRC) [19], and locality
group sensitive sparse representation (LGSR) [20]. Tang et
al. argued that locality weighted regularized term of LGSR
disrupts its group structure of sparse solution. Considering
that each class plays a different role in regressing the query
samples, they presented a weighted GSC (WGSC) [21] al-
gorithm with consideration of the influence of the similarity
between query samples and classes. However, the distribution
structure of the training samples may not coincide with the
natural class structure for a FR problem due to the influence
of corruption. As a result, in seeking a weights vector, these
classifiers may undesirably remove some samples which are
in fact needed to represent the query image. Timofte et al.
[12] adopted fixed point theorem to fully exploit the weights
information from dictionary atoms, without employing query-
adapted technique. They claimed that their proposed method
has higher computational efficiency than WCRC and WSRC,
while it keeps the same recognition performance. However,
their method requires carefully controlled training images with
both quality and quantity, which is difficult to achieve in
practice.
There are some attempts to alleviate the second constraint
of regression classifiers. RSRC [5] introduces an identity
matrix as a dictionary to code the outlier features (e.g.,
pixels with corruption or occlusion). Naseem et al. [22] and
Zhang et al. [23] respectively extended their LRC and CRC
to the robust version, robust linear regression classification
(RLRC) and robust collaborative representation classification
(RCRC), using the Huber and Laplacian estimator to deal
with severe random pixel noise and illumination changes.
To unify the existing robust sparse regression models: the
additive model for error correction and multiplicative model
for error detection, He et al. [24], [25] created a half-quadratic
framework by defining different half-quadratic functions based
on the maximum correntropy criterion. Borrowing the idea of
matrix based representation from NMR, Luo et al. [26] and
Chen et al. [27] respectively introduced matrix variate slash
and elliptically contoured distribution to image representation
for better noise resistant. In addition, Yang et al. sought for
a maximum a posterior solution and proposed a regularized
robust coding (RRC) model for FR [28], which is robust to
various types of feature outliers (e.g. corruption and facial
expression). Qian et al. [29] further extended RRC to robust
general regression and representation model (RGRR) by using
of the prior information of the training set. RGRR works well
when the query samples share the same probability distribution
with the training samples. However, it involves an independent
training stage that is unnecessary for the traditional regression
based classification approaches. To sum up, although much
progress has been made, robust FR is still an open issue due
to the complex variation of corruption.
In this work, we manage to solve the two elementary
limitations of regression based classification methods. We
propose an iterative re-constrained group sparse classification
(IRGSC) to increase the robustness of FR in dealing with
severe occlusion, complex corruption, real disguises and large
expression variation. The main contributions of this paper are
outlined as follows:
1) A general framework is presented for regression-based
classification. It unifies previous l
1
, l
2
or l
2,1
regularized norm
into a general formulation and learns the feature weights
and distance weights simultaneously to achieve the optimal
representation coefficients. We derive a new and efficient
algorithm to iteratively and adaptively update the weights
vector. Especially for the feature weights, we present a closed
solution seamlessly connected to the model with only one
univocal parameter, while the existing approaches all rely on
different distribution functions for corresponding noises.
2) We extend the convex group norm to a concave surrogate
function for a tighter approximation of the l
2,0
-norm. Then the
weighted l
2,p
-norm penalty is enforced on the coefficients for
purpose of imposing both distance locality and group sparsity,
where p is released from a fixed value and is flexible to various
training size and feature dimension. By introducing the feature
weights vector to compute the reconstruction residuals and
distance measurement, the proposed approach uses selected
features to reflect the true distribution structure. Compared
with WGSC, the sparse solution of IRGSC maintains locality
at a feature level and contains more discriminative information.
3) In our implementation, the IRGSC minimization problem
is transformed into an iteratively re-constrained group sparse
coding problem with a reasonably designed weight learning
strategy for robust FR. In theory, we prove that IRGSC mono-
tonically decreases the objective value and any coefficients
sequence is a stationary point. Our extensive experiments in
benchmark face databases show that IRGSC achieves much
better performance than existing regression based FR classi-
fiers, especially when there are complicated variations, such
as severe occlusions and corruptions, etc.
The rest of this paper is organized as follows: we introduce
a general formulation of regression based classifiers in Section
2. Section 3 introduces IRGSC classifier for face recognition.
In Section 4, we present the optimization algorithm of IRGSC.
Section 5 analyses the complexity and convergence of the
proposed method. In Section 6, we conduct experiments on 3
public face databases and compare our results with the state-
of-the-art methods. Finally, Section 7 concludes the paper.
II. A GENERAL FRAMEWORK FOR REGRESSION BASED
CLASSIFIER
For various classification task, different regression based
approaches have been proposed due to the variation of the
motivations, however, their purposes are often similar in
the sense that they aim to derive a series of representation