中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Self-taught dimensionality reduction on the high-dimensional small-sized data

文献类型:期刊论文

作者Zhu, Xiaofeng1; Huang, Zi1; Yang, Yang1; Shen, Heng Tao1; Xu, Changsheng2; Luo, Jiebo3
刊名PATTERN RECOGNITION
出版日期2013
卷号46期号:1页码:215-229
关键词Dimensionality reduction Self-taught learning Joint sparse coding Manifold learning Unsupervised learning
英文摘要To build an effective dimensionality reduction model usually requires sufficient data. Otherwise, traditional dimensionality reduction methods might be less effective. However, sufficient data cannot always be guaranteed in real applications. In this paper we focus on performing unsupervised dimensionality reduction on the high-dimensional and small-sized data, in which the dimensionality of target data is high and the number of target data is small. To handle the problem, we propose a novel Self-taught Dimensionality Reduction (STDR) approach, which is able to transfer external knowledge (or information) from freely available external (or auxiliary) data to the high-dimensional and small-sized target data. The proposed STDR consists of three steps: First, the bases are learnt from sufficient external data, which might come from the same "type" or "modality" of target data. The bases are the common part between external data and target data, i.e., the external knowledge (or information). Second, target data are reconstructed by the learnt bases by proposing a novel joint graph sparse coding model, which not only provides robust reconstruction ability but also preserves the local structures amongst target data in the original space. This process transfers the external knowledge (i.e., the learnt bases) to target data. Moreover, the proposed solver to the proposed model is theoretically guaranteed that the objective function of the proposed model converges to the global optimum. After this, target data are mapped into the learnt basis space, and are sparsely represented by the bases, i.e., represented by parts of the bases. Third, the sparse features (that is, the rows with zero (or small) values) of the new representations of target data are deleted for achieving the effectiveness and the efficiency. That is, this step performs feature selection on the new representations of target data. Finally, experimental results at various types of datasets show the proposed STDR outperforms the state-of-the-art algorithms in terms of k-means clustering performance. (C) 2012 Elsevier Ltd. All rights reserved.
WOS标题词Science & Technology ; Technology
类目[WOS]Computer Science, Artificial Intelligence ; Engineering, Electrical & Electronic
研究领域[WOS]Computer Science ; Engineering
关键词[WOS]SUPPORT VECTOR MACHINES ; GEOMETRIC FRAMEWORK ; FEATURE-SELECTION ; ALGORITHM
收录类别SCI
语种英语
WOS记录号WOS:000309785000019
源URL[http://ir.ia.ac.cn/handle/173211/2878]  
专题自动化研究所_模式识别国家重点实验室_多媒体计算与图形学团队
作者单位1.Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia
2.Chinese Acad Sci, Inst Automat, Beijing 100864, Peoples R China
3.Univ Rochester, Dept Comp Sci, Rochester, NY 14627 USA
推荐引用方式
GB/T 7714
Zhu, Xiaofeng,Huang, Zi,Yang, Yang,et al. Self-taught dimensionality reduction on the high-dimensional small-sized data[J]. PATTERN RECOGNITION,2013,46(1):215-229.
APA Zhu, Xiaofeng,Huang, Zi,Yang, Yang,Shen, Heng Tao,Xu, Changsheng,&Luo, Jiebo.(2013).Self-taught dimensionality reduction on the high-dimensional small-sized data.PATTERN RECOGNITION,46(1),215-229.
MLA Zhu, Xiaofeng,et al."Self-taught dimensionality reduction on the high-dimensional small-sized data".PATTERN RECOGNITION 46.1(2013):215-229.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。