Semi-supervised feature selection via hierarchical regression for web ...

3 downloads 9551 Views 700KB Size Report
May 27, 2014 - Feature selection is an important step for large-scale image data analysis, ... feature selection, especially for the large web image datasets.
Multimedia Systems (2016) 22:41–49 DOI 10.1007/s00530-014-0390-0

SPECIAL ISSUE PAPER

Semi-supervised feature selection via hierarchical regression for web image classification Xiaonan Song • Jianguang Zhang • Yahong Han Jianmin Jiang



Published online: 27 May 2014  Springer-Verlag Berlin Heidelberg 2014

Abstract Feature selection is an important step for largescale image data analysis, which has been proved to be difficult due to large size in both dimensions and samples. Feature selection firstly eliminates redundant and irrelevant features and then chooses a subset of features that performs as efficient as the complete set. Generally, supervised feature selection yields better performance than unsupervised feature selection because of the utilization of labeled information. However, labeled data samples are always expensive to obtain, which constraints the performance of supervised feature selection, especially for the large web image datasets. In this paper, we propose a semi-supervised X. Song School of Software Engineering and Technology, Tianjin University, Tianjin, China e-mail: [email protected]

feature selection algorithm that is based on a hierarchical regression model. Our contribution can be highlighted as: (1) Our algorithm utilizes a statistical approach to exploit both labeled and unlabeled data, which preserves the manifold structure of each feature type. (2) The predicted label matrix of the training data and the feature selection matrix are learned simultaneously, making the two aspects mutually benefited. Extensive experiments are performed on three large-scale image datasets. Experimental results demonstrate the better performance of our algorithm, compared with the state-of-the-art algorithms. Keywords Feature selection  Multi-class classification  Semi-supervised learning

1 Introduction J. Zhang  Y. Han (&) School of Computer Science and Technology, Tianjin University, Tianjin, China e-mail: [email protected] J. Zhang e-mail: [email protected] J. Zhang Department of Mathematics and Computer Science, Hengshui University, Hengshui, China Y. Han Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, China J. Jiang University of Surrey, Guildford, UK e-mail: [email protected] J. Jiang School of Computer Science and Software Engineering, Shenzhen University, Shenzhen, China

With rapid advances of digital camera and various image processing tools, a huge amount of images are available online through various image sharing websites such as Flickr and Facebook. To efficiently browse and retrieve the web images, numerous approaches have been proposed, such as image classification [1], video annotation [2], and image retrieval [3, 4]. Most of these existing visual computing tasks use image features that are often noisy and redundant. Hence, it is of great value to propose effective feature selection methods for a more precise and compact representation of images. The huge image databases also pose significant challenges in existing applications in terms of scalability. Virtually images are usually represented by different features. Unfortunately, the combination of these features results in high-dimensional vector data (e.g., hundreds– thousands), leading to computational and memory-related complications. Reducing the number of redundant features

123

42

saves the running time of a learning algorithm and yields a more general concept. Feature selection is one effective means that aims to choose a subset of features for improving prediction accuracy or decreasing the size of the structure without significantly decreasing prediction accuracy of the classifier built using only the selected features [5]. Supervised feature selection evaluates feature relevance by measuring the correlation between the feature and class labels. Because of the utilization of labeled information, supervised feature selection usually yields better and more reliable performance than unsupervised feature selection [6]. However, supervised feature selection requires sufficient labeled data samples which are expensive to obtain because of the excessive cost in human labor. That is the reason why in many real applications, we have huge unlabeled data and small labeled samples. The desire to develop feature selection methods that are capable of exploiting both labeled and unlabeled data motivates us to introduce semi-supervised learning [7, 8] into the feature selection process. Recent researches of semi-supervised learning [9–11] focus on the utilization of unlabeled data. Generally, exploiting unlabeled data is based on the manifold assumption that most data examples lie on a low-dimensional manifold in the input space. The well-known algorithms in this category include label propagation, harmonic function, graph cuts, spectral graph transducer, and manifold regularization. [8] can be referred to as a comprehensive study of semi-supervised learning techniques. Different from most of the existing semi-supervised algorithms [12, 13], Yang et al. [7] propose a new semi-supervised learning method that preserves the manifold structure of each feature type and utilizes a statistical approach to better exploit the manifold sturcture of the training data. Experimental results demonstrate that their method is more robust than simply using the pairwise distances of the training data. In this paper, we reserve the existing advantages of the semisupervised learning model proposed in [7] and benefit the semi-supervised feature selection from it. Traditional feature selection algorithms, e.g., Fisher Score [14], generally evaluate the importance of each feature individually and discard important features that are less informative by themselves but are informative when combined with other features. In contrast, a recently popular approach, sparsity-based feature selection, selects features jointly across all data points [6, 15–19]. ‘1 -SVM is well known to perform feature selection using the ‘1 -norm regularization that gives sparse solution [20]. A combination of both ‘1 -norm and ‘2 -norm is proposed to form a structured regularization in [21]. In [6, 15, 22], they all use ‘2;1 -norm on regularization to evaluate features jointly. The regularization function in their works has twofold role in avoiding over-fitting and inducing sparsity, making it particularly suitable for feature selection.

123

X. Song et al.

In this paper, we propose a new semi-supervised feature selection algorithm, namely feature selection via hierarchical regression (FSHR). Different from most of the existing semi-supervised feature selection algorithms [22, 23], we preserve the manifold structure of each feature type during the training phase. Local classifiers and global classifiers are learnt simultaneously. ‘2;1 -norm is added to the regularization of global classifiers, making many rows of optimal global classifiers shrink to zeros. Once the optimal global classifiers are obtained, we sort the features accordingly in descending order and select the top-ranked ones. The selected discriminative features are proved to be effective to classify web images, when combined with a multi-class SVM classifier in the experiments. In comparison with the existing popular feature selection method [14, 15, 22, 23], our proposed take advantages of both semisupervised learning of [7] and ‘2;1 -norm-based sparsity, which can be highlighted as: (1) FSHR embeds the feature selection process with the semi-supervised learning which preserves the manifold structure of each feature type. A statistical approach is used to better exploit the manifold structure of the training data, resulting in a more faithful and robust learning result. (2) We propose to simultaneously learn the predicted label matrix of the training data and the feature selection matrices, making the two aspects mutually promotive.

2 The objective function In this section, we give the objective function of the proposed algorithm. We begin with the preliminaries. t is the number of labeled training images, and n is the number of the training images, t  n. Given an integer g  v, we denote xig as the g-th feature of the i-th image and Xg ¼ ½x1g ; . . .; xng  2 Rdg n , where dg is the dimension of the g-th feature. Denote Y ¼ ½Y1 ; Y2 ; . . .Yn T 2 f0; 1gnc as the label information. Yij is 1 if xi belongs to the j-th class, and 0 otherwise. Let F ¼ ½F1 ; F2 ; . . .Fn T 2 Rnc , where Fi 2 Rc is the predicted label vector of xi . We define fg ¼ ½fg1 ; fg2 ; . . .fgn T 2 Rnc as the predicted matrix of training images derived from the g-th feature. In the rest of this paper, k  kF denotes the Frobenius norm. 1m 2 Rm is a column vector with all its elements being 1, and Hm ¼ I  m1 1m 1Tm is the centering matrix, where I is the identity matrix. For an arbitrary matrix A 2 Rrp , its ‘2;1 -norm is ffi P qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi Pp 2 . Tr() is the trace A defined as kAk2;1 ¼ ri¼1 j¼1 ij operator. As indicated by previous semi-supervised learning approaches [13, 24, 25], exploiting the ‘‘manifold

Hierarchical regression for web image classification

43

structure’’ of the data is a promising family of techniques that are generally based upon an assumption that similar unlabeled examples should be given the same classification. To utilize unlabeled data effectively, we construct a i local set, denoted as N g , consisting of xig and its k-nearest neighbors according to the distance derived from the g-th feature. We use the statistical approach proposed in [7] to exploit the manifold structure of the input images for semisupervised learning. Similar to [7], we assume there is a i local classifier that classifies all the training images in N g . We use least square regression for classification as 2  2  v X n X  X   i T j  i i min ðw Þ x þ b  f ;  g g gj  þkwg  g fg ;wig big

F

g¼1 i¼1 xjg 2N i g

ð1Þ

F

2R and 2 R are the local classifier and where bias term of xig w.r.t. the g-th feature. The predicted label matrix F of the training images should be consistent with each evidence fg and the ground truth if exists. Therefore, P we minimize minF;fg vg¼1 kF  fg k2F and impose a hard constraint on F that Fi ¼ Yi , if xi is a labeled training image. In this way, each training image learns a soft label on each kind of feature, which is gained by considering both labeled images and the manifold structure of input images. We formulate the semi-supervised learning parts as  2  2  v X n X  X  i T j  i  i min ðw Þ x þ b  f  gj  þkwg  g g g i i wig

fg ;wg ;bg

dg c

big

c

F

g¼1 i¼1 xjg 2N i g

F

v  X  F  fg 2 ; þ l2 F g¼1

s:t:Fi ¼ Yi ; if xi is a labeled training image:

ð2Þ

To select features, we utilize the ‘2;1 -norm to control classifier’s capacity and also ensure it is sparse in rows, making it particularly suitable for feature selection. We propose to simultaneously learn the predicted label matrix fg from the g-th feature, and the v feature selection matrices Wg . Specifically, the objective function is shown as follows: 2  2  v X n X  X  i T j  i  i ðw Þ x þ b  f min  g g gj  þkwg  g fg ;wig

big ;F

F

g¼1 i¼1 xjg 2N i g

þ l1 þ l2

F

v X n  2    X   T ðWg Þ xig þ Bg  fgi  þcWg 2;1

Xgi ¼ ½xig ; xig1 ; . . .; xigk  2 Rdg ðkþ1Þ . The objective function can be rewritten as 2  2  v X n  X  i T i    min ðXg Þ wg þ 1Kþ1 big  fgi  þkwig  F;fg ;wig

Wg ;big ;Bg

g¼1 v    X  F  fg 2 þ Tr ðF  YÞT UðF  YÞ ; þ l2 F

ð4Þ

g¼1

where fgi ¼ ½fgi ; fgi1 ; . . .; fgik T 2 Rkc is the predicted local label matrix of Xgi . 3 Solution to FSHR The objective problem of (4) can be solved as follows. By setting the derivative of (4) w.r.t. wig and big to the 0, we have  1  i T ð5Þ ðfg Þ 1kþ1  ðwig ÞT Xgi 1kþ1 ; big ¼ kþ1  1 ð6Þ Xgi Hkþ1 fgi : wig ¼ Xgi Hkþ1 ðXgi ÞT þ kI Similarly, by setting the derivative of (4) w.r.t. Wg and Bg to be zero, we have 1 ð7Þ Bg ¼ ðfgT 1n  WgT Xg 1n Þ; n Wg ¼ ðXg Hn XgT þ cDg Þ1 Xg Hn fg ;

ð8Þ

where Dg is a diagonal matrix with each element on the diagonal, i.e., Diig , defined as 1  Diig ¼  ð9Þ  i : 2wg  2

Substituting big ; wig ; Bg ; and Wg in (4) by (5), (6), (7) and (8), respectively, we arrive at v X n v   X X Tr ðfgi ÞT Lig fgi þ l1 Tr ðfgT Ag fg Þ min g¼1 i¼1 v  X

g¼1

 F  fg 2 þ TrððF  YÞT UðF  YÞÞ; F

ð10Þ

g¼1

F

g¼1

s:t:Fi ¼ Yi ; if xi is a labeled training data;

ð3Þ

where Wg 2 R and Bg 2 R are the global classifier and bias term w.r.t. the g-th feature. Denote U as a diagonal dg c

F

v  X     ðXg ÞT Wg þ 1n Bg  fg 2 þcWg  F 2;1

þ l2

 F  fg 2 ;

F

g¼1 i¼1

þ l1

F;fg

F

g¼1 i¼1 v  X

matrix. If xi is a labeled datum, Uii ¼ 1, and Uii ¼ 0 otherwise. Let N ig ¼ fxig ; xig1 ; . . .; xigk g where xig1 ; . . .; xigk are the k-nearest neighbors of xig according to the g-th feature.

c

where  1 Lig ¼ Hkþ1  Hkþ1 ðXgi ÞT  Xgi Hkþ1 ðXgi ÞT þ kI Xgi Hkþ1 ð11Þ

123

44

X. Song et al. v X

and Ag ¼ H n 

Hn XgT



Xg Hn XgT

þ cDg

1

Xg Hn :

ð12Þ

For the ease of representation, we define the selection matrix Spg 2 Rnðkþ1Þ as follows:    1; if xi is the j-th element in N p ; g g Spg ¼ ð13Þ ij 0; otherwise:

l2 ðF  fg Þ þ UðF  YÞ ¼ 0:

ð22Þ

g¼1

Substituting fg in (22) by (21), we have F¼

vl2 I þ U 

l22

v X

!1 ðLg þ l1 Ag þ l2 IÞ

1

UY:

g¼1

ð23Þ

Therefore, we have fgi ¼ ðSig ÞT fg :

ð14Þ

Then, we have n X

n   X   Tr ðfgi ÞT Lig fgi ¼ Tr fgT Sig Lig ðSig ÞT fg

i¼1

i¼1

¼ Tr

! n   X T fgT Sig Lig ðSig Þ fg :

ð15Þ

i¼1

Denote Lg ¼

Pn  i¼1

 Sig Lig ðSig ÞT , Then, (10) can be written

as min F;fg

v X

Tr ðfgT Lg fg Þ þ l1

g¼1

v X

Tr ðfgT Ag fg Þ

g¼1

v  X 

F  fg 2 þ Tr ðF  YÞT UðF  YÞ ; þ l2 F

ð16Þ

g¼1

which is equivalent to the following min F;fg

v X

  Tr fgT ðLg þ l1 Ag Þfg

g¼1

þ l2

v  X 

F  fg 2 þ Tr ðF  YÞT UðF  YÞ : F

ð17Þ

g¼1

Let us define 2 1 Lg 6  Qg ¼ 4 0

0 Lng

3 7 5

ð18Þ

and Sg ¼ ½S1g ; . . .; Sng :

ð19Þ

2

Then, we have Lg ¼ Sg Qg STg :

ð20Þ

By setting the derivative of (17) w.r.t fg to be 0, we have

1 ð21Þ fg ¼ l2 Lg þ l1 Ag þ l2 I F: By setting the derivative of (17) w.r.t F to be 0, we have

123

Based on the above mathematical deduction, the procedure can be described in Algorithm 1. For each feature, once Wg     is obtained, we sort the features according to wig  in descending order and select the top-ranked ones. The complexity of the proposed algorithm is discussed as follows. In each iteration, the most time-consuming steps are Step 8, Step 10, and Step 11. In Step 8, the time complexity of computing Lg defined in (20) is about Oðn3 Þ. In Step 10 and Step 11, the complexity of calculating the inverse of a few matrices is Oðn3 Þ.

Hierarchical regression for web image classification

4 Experiments

45



4.1 Datasets descriptions • We choose three image datasets, i.e., MFlickr [26], MSRAMM [27], and NUS-WIDE [28] in our experiments. As our method is irrelevant to multi-labeled classification, we remove multi-labeled samples so that each sample only belongs to one class. However, the labels of the samples in the three datasets are not balanced. For instance, in NUSWIDE dataset, there is no positive image for one class while there are 31,617 positive images for another class after removing multi-labeled samples. Therefore, we preprocess the dataset so that the numbers of positive images of different classes are balanced. The following is a brief description of the three datasets and their preprocessing. MFlickr This dataset contains 25,000 samples in 33 classes from the Flickr website. Classes whose samples are \2,000 or more than 5,000 are excluded. Images that are not labeled or multi-labeled are removed. After preprocessing, we obtain a dataset with 9,439 images in eight classes. We extract the three types of features for each image, namely 499 dimensional Bag of Words, 128 dimensional Color Coherence, and 256 dimensional Color Histogram. MSRA-MM We use the concept set of the original MSRA-MM 2.0 dataset, which includes 50,000 images related to 100 concepts. After the same preprocessing of MFlickr, there are 14,445 images in six classes remaining. Three feature types used in [22, 29], namely Color Correlogram, Edge Direction Histogram, and Wavelet Texture are also used in our experiments to represent the images. NUS-WIDE The dataset includes 269,648 images and 81 concepts that can be used for evaluation. Concepts with more than 10,000 or \5,000 images are excluded, and images associated with no concept or more than two concepts are removed. We use the remaining 45,227 images associated with 10 concepts in our experiments. The images are also represented by Color Correlogram, Edge Direction Histogram, and Wavelet Texture. 4.2 Experiment setup We compare our method proposed in this paper with the following feature selection algorithms. •



Fisher Score [14] depends on fully labeled training data to select features and calculates a Fisher/correlation score for each feature. Feature selection via joint l2;1 -norms minimization (FSNM) [15] employs joint l2;1 -norm minimization on both loss function and regularization to realize feature selection across all data points.

Feature selection via spectral analysis (FSSA) [23] is a semi-supervised feature selection method using spectral regression. Structural feature selection with sparsity (SFSS) [22] jointly selects the most relevant features from all the data points using a sparsity-based model and exploiting both labeled and unlabeled data to learn the manifold structure.

To generate the training data, we randomly sample 1,000 data for each dataset. This sampling process is repeated five times, and average results are reported. We set the percentage of labeled data to be 1, 5, 10, 25, 50, and 100 %, respectively, and make sure the labeled data cover all the semantic concept of each dataset. The rest images of each dataset work as the testing set to evaluate the multiclassification performance. There are five parameters in our algorithm. However, as the performance is not sensitive to the local regularization parameter k [30], we did not tune this parameter and fix it as one. In our experiment, the parameter k specifies the k nearest neighbors used to compute the graph Laplacian matrix, and it is used in the same way as FSSA and SFSS do. We set k to five in our algorithm for all three datasets, as well as in FSSA and SFSS. Except these parameters, we tune all the other parameters (if any) by a grid-search strategy from f106 ; 103 ; 1; 103 ; 106 g to fairly compare different feature selection algorithms. For all the algorithms, we report the best results obtained from different parameters. Generally, we can see that in terms of average accuracy, our method works best and is followed by SFSS, FSNM, FSSA, and Fisher Score in sequence. SFSS is the semi-supervised feature selection algorithm published in the ACM-MM conference, 2011. Our method and SFSS both utilize the unlabeled data. When \25 % of the data are labeled, our method is competitive with SFSS. In our experiment, each feature selection algorithm is first performed to select features. Then, the support vector machine (SVM) classifier with the linear kernel is employed to these datasets, using fivefold cross-validation SVM classifier. Accuracy is widely used as evaluation metric, so we use accuracy in this paper. 4.3 Experimental results and discussions For each dataset and different percentages of labeled data, we average the accuracy for different numbers of selected features ({10, 20,…, 100}). Figure 1 shows the classification accuracy compared to other feature selection algorithms when different percentages of data are labeled. The detailed differences of the averaged accuracy are listed in Tables 1, 2, and 3, when 1, 5, and 10 % of the training data are labeled.

123

46

X. Song et al. MSRA−MM data

NUS−WIDE data 45

40

40

40

35

35

35

30 25 Fisher−score FSNM FSSA SFSS ours

20 15 10

1 5 10

25

50

30 25 Fisher−score FSNM FSSA SFSS ours

20 15

100

Percentage of labeled training data (%)

Accuracy (%)

45

Accuracy (%)

Accuracy (%)

MFlickr data 45

10

1 5 10

25

50

Percentage of labeled training data (%)

30 25 Fisher−score FSNM FSSA SFSS ours

20 15

100

10

1 5 10

25

50

100

Percentage of labeled training data (%)

Fig. 1 The classification accuracy compared with other feature selection algorithms when different percentages of data are labeled

Table 1 Performance comparison (accuracy  standard deviation) when 1 % training images are labeled

Table 2 Performance comparison (accuracy  standard deviation) when 5 % training images are labeled

Table 3 Performance comparison (accuracy  standard deviation) when 10 % training images are labeled

MFlickr

NUS-WIDE

Ours

0.1864 ± 0.0669

0.2426 ± 0.0288

0.1811 ± 0.0349

F score

0.1444 ± 0.0371

0.2009 ± 0.0380

0.1603 ± 0.0120

FSNM

0.1794 ± 0.0676

0.2332 ± 0.0430

0.1735 ± 0.0306

FSSA

0.1566 ± 0.0576

0.2149 ± 0.0373

0.1641 ± 0.0098

SFSS

0.1844 ± 0.0644

0.2341 ± 0.0329

0.1738 ± 0.0289

MFlickr

MSRA-MM

NUS-WIDE

Ours

0.2577 ± 0.0283

0.2949 ± 0.0185

0.2890 ± 0.0213

F score

0.2152 ± 0.0256

0.2651 ± 0.0238

0.2590 ± 0.0251

FSNM

0.2533 ± 0.0332

0.2804 ± 0.0190

0.2811 ± 0.0259

FSSA

0.2432 ± 0.0315

0.2649 ± 0.0186

0.2596 ± 0.0196

SFSS

0.2577 ± 0.0286

0.2837 ± 0.0176

0.2845 ± 0.0262

MFlickr

MSRA-MM

NUS-WIDE

Ours

0.2621 ± 0.0031

0.3263 ± 0.0181

0.3213 ± 0.0114

F score

0.2364 ± 0.0154

0.2934 ± 0.0204

0.2966 ± 0.0183

FSNM

0.2620 ± 0.0259

0.3258 ± 0.0214

0.3120 ± 0.0109

FSSA

0.2610 ± 0.0027

0.2950 ± 0.0212

0.2888 ± 0.0128

SFSS

0.2620 ± 0.0027

0.3167 ± 0.0242

0.3163 ± 0.0150

We have the following observations and analysis from the experimental results: (1) Generally, we can see that in terms of average accuracy, our method works best and is followed by SFSS, FSNM, FSSA, and Fisher Score in sequence. Our method and SFSS both utilize the unlabeled data. This observation suggests that the unlabeled information is important for feature selection. (2) When 10 % or less of the training data are labeled, our method outperforms Fisher Score and FSSA obviously. Our method is competitive with SFSS and FSNM or better, but the performance improvement is not significant. On the MFlickr dataset and the NUS-WIDE dataset, our method achieves

123

MSRA-MM

the same performance with the better of SFSS and FSNM. On the MSRA-MM dataset, our method outperforms SFSS and FSNM slightly. (3) Finally, when more than 10 % of the data are labeled, our method consistently outperforms other methods on the NUS-WIDE dataset and the MSRAMM dataset. However, the relative difference between our method and other feature selection methods becomes smaller. On the MFlickr dataset, our method achieves similar performance with FSSA, FSNM, and SFSS, but outperforms Fisher Score. Figure 2 shows the classification accuracy comparisons of all five feature selection methods with different numbers

Hierarchical regression for web image classification MFlickr data, with 5% labeled training data

30

20

25

25

20

15

Fisher−score FSNM FSSA SFSS ours

15

20

30

40

50

60

70

80

90

10 10

100

20

Number of Selected Features

Fisher−score FSNM FSSA SFSS ours

Accuracy (%)

Accuracy (%)

50

60

70

25

40

50

60

70

80

10 10

100

20

90

30

30

Fisher−score FSNM FSSA SFSS ours 20

30

40

50

60

70

80

40

50

60

70

80

90

100

MSRA−MM data, with 10% labeled training data 35

15 10

100

30

Number of Selected Features

35

20

30

90

25

20

20

80

MSRA−MM data, with 5% labeled training data

35

15 10

40

Fisher−score FSNM FSSA SFSS ours

Number of Selected Features

MSRA−MM data, with 1% labeled training data

30

30

20

15

Accuracy (%)

10 10

30

90

25 Fisher−score FSNM FSSA SFSS ours

20

100

15 10

20

30

40

50

60

70

80

90

100

Number of Selected Features

Number of Selected Features

Number of Selected Features

NUS−WIDE data, with 1% labeled training data

NUS−WIDE data, with 5% labeled training data

NUS−WIDE data, with 10% labeled training data

35

Accuracy (%)

30

Accuracy (%)

Fisher−score FSNM FSSA SFSS ours

25

20

15

10 10

35

35

30

30

25

20 Fisher−score FSNM FSSA SFSS ours

15

20

30

40

50

60

70

80

Number of Selected Features

90

100

Accuracy (%)

Accuracy (%)

25

Accuracy (%)

Fisher−score FSNM FSSA SFSS ours

MFlickr data, with 10% labeled training data

30

Accuracy (%)

MFlickr data, with 1% labeled training data

47

10 10

20

30

40

50

60

70

80

Number of Selected Features

90

25

20 Fisher−score FSNM FSSA SFSS ours

15

100

10 10

20

30

40

50

60

70

80

90

100

Number of Selected Features

Fig. 2 The classification accuracy comparisons of all five feature selection methods with different numbers of selected feature and accuracy, when 1, 5, and 10 % of the training data are labeled

of selected features and accuracy, when 1, 5, and 10 % of the training data are labeled. From Fig. 2, we can see that when labeled images are 1 %, the performance of FSHR is constraint by the very few labeled images, limiting the accuracy almost constant among different numbers of features on the three datasets. However, the first few features can indeed improve the classification accuracy, and the proposed method works better than other approaches. In MFlickr dataset, when the labeled images are more than 5 %, FSHR is constantly comparative with FSNM, FSSA, and SFSS. In MSRA-MM and NUS-WIDE datasets, when the labeled images are more than 5 %, the general trend shows that classification accuracy increases fast at beginning and then

stays steady with the increasing number of features. Moreover, method works consistently better than the other feature selection algorithms on the top 100 features. To evaluate the accuracy of our algorithm w.r.t the three parameters c; l1 ; and l2 , we perform an experiment on the parameter sensitivity where 10 % training data are labeled. We average the accuracy for different datasets and different numbers of selected features ({10, 20, . . ., 100}). Our method gains the best performance when c ¼ 1; l1 ¼ 106 ; and l2 ¼ 106 for the three datasets in average. From Fig. 3, we can see that our algorithm is less sensitive to c; l1 ; and l2 when c is not \1.

123

48

X. Song et al.

(a) 28 26 24 22

28 26 24 22

20

20

10^−6 10^−3

10^−6 10^−3

1 10^3 10^6 μ 2

1 10^−3 10^−6 μ 1

10^3

10^6

30

Accuracy (%)

30

Accuracy (%)

30

Accuracy (%)

(c)

(b)

26 24 22 20

1 10^3 10^6 μ 2

Fig. 3 Averaged accuracy of three datasets for different numbers of selected features with 10 % training data labeled. a Performance variation w.r.t. l1 and l2 when c is fixed as 1. b Performance

1 10^−3 10^−6

γ

In this paper, we have proposed a new semi-supervised feature selection algorithm FSHR. In comparison with most of feature selection algorithms reported in the literature, our essential contribution is to simultaneously learn the statistically predicted label matrix and the ‘2;1 -normbased feature selection matrix, making the two aspect mutually promotive. Via embedding the feature selection process with the existing state-of-the-art semi-supervised learning algorithm published lately, the manifold structure of each feature type is preserved, resulting in a more faithful learning result. Experiments support that the proposed algorithm achieves better classification results on three large image datasets, in comparison with the popular supervised and semi-supervised feature selection methods. Although the semi-supervised feature selection algorithm has shown great potential for effective web image classification, exploiting the unlabeled data might have negative effects if the manifold assumption does not hold. Also, the performance is related to the number of selected features. Too many features selected may bring in redundant or noisy features, while too few features may not provide enough information for classification. Therefore, the future work includes designing: (1) algorithms that are able to predict if the unlabeled data will contribute and (2) methods that decide the optimal number of features that should be selected.

2.

3.

4.

5. 6.

7.

8.

9.

10.

11.

12. Acknowledgments This paper was partially supported by the National Program on the Key Basic Research Project (under Grant 2013CB329301), NSFC (under Grant 61202166), and Doctoral Fund of Ministry of Education of China (under Grant 20120032120042).

13.

14.

References 15. 1. Zha, Z.-J., Hua, X.-S., Mei, T., Wang, J., Qi, G.-J., Wang, Z.: Joint multi-label multi-instance learning for image classification.

10^3

10^6

10^−6 10^−3 1 10^3 10^6 μ 1

1 10^−3 10^−6

10^3

10^6

γ

variation w.r.t. c and l2 when l1 is fixed as 106 . c Performance variation w.r.t. c and l1 when l2 is fixed as 106

5 Conclusion

123

28

In: IEEE Conference on Computer Vision and Pattern Recognition, 2008. IEEE, (2008) Zha, Z.-J., Wang, M., Zheng, Y.-T., Yang, Y., Hong, R., Chua, T.-S.: Interactive video indexing with statistical active learning. IEEE Trans. Multimed. 14, 17–27 (2012) Zha, Z.-J., Yang, L., Mei, T., Wang, M., Wang, Z.: Visual query suggestion. In: Proceedings of the 17th ACM international conference on Multimedia, pp. 15–24. ACM, (2009) Zha, Z.-J., Yang, L., Wang, Z., Chua, T.-S., Hua, X.-S.: Visual query suggestion: towards capturing user intent in internet image search. ACM Trans. Multimed. Comput. Commun. Appl. (TOMCCAP) 6(13), 1–19 (2010) Koller, D., Sahami, M.: Toward optimal feature selection. Technical Report 1996–77, Stanford InfoLab, February (1996) Han, Y., Yang, Y., Zhou, X.: Co-regularized ensemble for feature selection. In: Proceedings of the Twenty-Third international joint conference on Artificial Intelligence (2013) Yang, Y., Song, J., Huang, Z., Ma, Z., Sebe, N.: Multi-feature fusion via hierarchical regression for multimedia analysis. IEEE Trans. Multimed. 15, 572–581 (2012) Zhu, X.: Semi-supervised learning literature survey. Technical Report 1530, Computer Sciences, University of WisconsinMadison (2005) Zhang, T., Changsheng, X., Hanqing, L.: A generic framework for video annotation via semi-supervised learning. IEEE Trans. Multimed. 14, 1206–1219 (2012) Zha, Z.-J., Mei, T., Wang, J., Wang, Z., Hua, X.-S.: Graph-based semi-supervised learning with multiple labels. J. Vis. Commun. Image Represent. 20(2), 97–103 (2009). Special issue on Emerging Techniques for Multimedia Content Sharing, Search and Understanding Zhu, J., Hoi, S.C.H., Lyu, M.R., Yan, S.: Near-duplicate keyframe retrieval by semi-supervised learning and nonrigid image matching. ACM Trans. Multimed. Comput. Commun. Appl. (TOMCCAP) 7(1), 4:1–4:24 (2011) Belkin, M., Niyogi, P., Sindhwani, V.: Manifold regularization: a geometric framework for learning from labeled and unlabeled examples. J. Mach. Learn. Res. 7, 2399–2434 (2006) Nie, F., Dong, X.: Flexible manifold embedding: a framework for semi-supervised and unsupervised dimension reduction. IEEE Trans. Image Process. 19(7), 1921–1932 (2010) Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. Wiley-Interscience, New York (2001) Nie, F., Huang, H., Cai, X., Ding, C.H.: Efficient and robust feature selection via joint l2;1 -norms minimization. Adv. Neural Inf. Process. Syst., pp. 1813–1821 (2010)

Hierarchical regression for web image classification 16. Yang, Y., Shen, H.T., Ma, Z., Huang, Z., Zhou, X.: l 2, 1-norm regularized discriminative feature selection for unsupervised learning. In: Proceedings of the Twenty-Second international joint conference on Artificial Intelligence, pp. 1589–1594. AAAI Press, (2011) 17. Zhao, Z., Wang, L., Liu, H.: Efficient spectral feature selection with minimum redundancy. In: Proceedings of the AAAI Conference on Artificial Intelligence, (2010) 18. Bao, B.-K., Liu, G., Yan, S.: Inductive robust principal component analysis. IEEE Trans. Image Process. 21(8), 3794–3800 (2012) 19. Bao, B.-K., Zhu, G., Shen, J., Yan, S.: Robust image analysis with sparse representation on quantized visual features. IEEE Trans. Image Process. 22(3), 860–871 (2013) 20. Bradley, P. S., Mangasarian, O.L.: Feature selection via concave minimization and support vector machines. In: International Conference on Machine Learning (ICML), vol. 98, pp. 82–90, (1998) 21. Sun, L., Liu, J., Chen, J., Ye, J.: Efficient recovery of jointly sparse vectors. Adv. Neural Inf. Process. Syst., pp. 1812–1820, (2009) 22. Ma, Z., Yang, Y., Nie, F., Uijlings, J., Sebe, N.: Exploiting the entire feature space with sparsity for automatic image annotation. In: Proceedings of the 19th ACM international conference on Multimedia, pp. 283–292. ACM, (2011) 23. Zhao, Z., Liu, H.: Semi-supervised feature selection via spectral analysis. In: SIAM International Conference on Data Mining, (2007)

49 24. Zhu, X., Ghahramani, Z., Lafferty, J. et al.: Semi-supervised learning using gaussian fields and harmonic functions. In: ICML, vol. 3, pp. 912–919, (2003) 25. Zenglin, X., Irwin King, M.R.-T., Lyu, R.J.: Discriminative semisupervised feature selection via manifold regularization. IEEE Trans. Neural Netw. 21(7), 1033–1047 (2010) 26. Huiskes, M.J., Lew, M.S.: The mir flickr retrieval evaluation. In: Proceedings of the 1st ACM international conference on Multimedia information retrieval, pp. 39–43. ACM, (2008) 27. Li, H., Wang, M., Hua, X.-S.: MSRA-MM 2.0: a large-scale web multimedia dataset. In: IEEE International Conference on Data Mining Workshops, pp. 164–169. IEEE, (2009) 28. Chua, T.-S., Tang, J., Hong, R., Li, H., Luo, Z., Zheng, Y.: NUSWIDE: a real-world web image database from National University of Singapore. In: Proceedings of the ACM International Conference on Image and Video Retrieval, number 48 in CIVR ’09, pp. 1–9. ACM, (2009) 29. Wu, F., Yuan, Y., Zhuang, Y.: Heterogeneous feature selection by group lasso with logistic regression. In: Proceedings of the international conference on Multimedia, pp. 983–986, (2010) 30. Yang, Y., Xu, D., Nie, F., Luo, J., Zhuang, Y.: Ranking with local regression and global alignment for cross media retrieval. In: Proceedings of the 17th ACM international conference on Multimedia, pp. 175–184. ACM, (2009)

123

Suggest Documents