Indexed by:
Abstract:
Data fusion is used to integrate features from heterogenous data sources into a consistent and accurate representation for certain learning tasks. As an effective technique for data fusion, unsupervised multimodal feature representation aims to learn discriminative features, indicating the improvement of classification and clustering performance of learning algorithms. However, it is a challenging issue since varying modality favors different structural learning. In this paper, we propose an efficient feature learning method to represent multimodal images as a sparse multigraph structure embedding problem. First, an effective algorithm is proposed to learn a sparse multigraph construction from multimodal data, where each modality corresponds to one regularized graph structure. Second, incorporating the learned multigraph structure, the feature learning problem for multimodal images is formulated as a form of matrix factorization. An efficient corresponding algorithm is developed to optimize the problem and its convergence is also proved. Finally, the proposed method is compared with several state-of-the-art single-modal and multimodal feature learning techniques in eight publicly available face image datasets. Comprehensive experimental results demonstrate that the proposed method outperforms the existing ones in terms of clustering performance for all tested datasets.
Keyword:
Reprint 's Address:
Version:
Source :
IEEE TRANSACTIONS ON MULTIMEDIA
ISSN: 1520-9210
Year: 2017
Issue: 7
Volume: 19
Page: 1454-1466
3 . 9 7 7
JCR@2017
8 . 4 0 0
JCR@2023
ESI Discipline: COMPUTER SCIENCE;
ESI HC Threshold:187
JCR Journal Grade:1
CAS Journal Grade:1
Cited Count:
WoS CC Cited Count: 66
SCOPUS Cited Count: 70
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 0
Affiliated Colleges: