Self-taught dimensionality reduction on the high-dimensional small-sized data
文献类型:期刊论文
作者 | Zhu, Xiaofeng1; Huang, Zi1; Yang, Yang1; Shen, Heng Tao1; Xu, Changsheng2![]() |
刊名 | PATTERN RECOGNITION
![]() |
出版日期 | 2013 |
卷号 | 46期号:1页码:215-229 |
关键词 | Dimensionality reduction Self-taught learning Joint sparse coding Manifold learning Unsupervised learning |
英文摘要 | To build an effective dimensionality reduction model usually requires sufficient data. Otherwise, traditional dimensionality reduction methods might be less effective. However, sufficient data cannot always be guaranteed in real applications. In this paper we focus on performing unsupervised dimensionality reduction on the high-dimensional and small-sized data, in which the dimensionality of target data is high and the number of target data is small. To handle the problem, we propose a novel Self-taught Dimensionality Reduction (STDR) approach, which is able to transfer external knowledge (or information) from freely available external (or auxiliary) data to the high-dimensional and small-sized target data. The proposed STDR consists of three steps: First, the bases are learnt from sufficient external data, which might come from the same "type" or "modality" of target data. The bases are the common part between external data and target data, i.e., the external knowledge (or information). Second, target data are reconstructed by the learnt bases by proposing a novel joint graph sparse coding model, which not only provides robust reconstruction ability but also preserves the local structures amongst target data in the original space. This process transfers the external knowledge (i.e., the learnt bases) to target data. Moreover, the proposed solver to the proposed model is theoretically guaranteed that the objective function of the proposed model converges to the global optimum. After this, target data are mapped into the learnt basis space, and are sparsely represented by the bases, i.e., represented by parts of the bases. Third, the sparse features (that is, the rows with zero (or small) values) of the new representations of target data are deleted for achieving the effectiveness and the efficiency. That is, this step performs feature selection on the new representations of target data. Finally, experimental results at various types of datasets show the proposed STDR outperforms the state-of-the-art algorithms in terms of k-means clustering performance. (C) 2012 Elsevier Ltd. All rights reserved. |
WOS标题词 | Science & Technology ; Technology |
类目[WOS] | Computer Science, Artificial Intelligence ; Engineering, Electrical & Electronic |
研究领域[WOS] | Computer Science ; Engineering |
关键词[WOS] | SUPPORT VECTOR MACHINES ; GEOMETRIC FRAMEWORK ; FEATURE-SELECTION ; ALGORITHM |
收录类别 | SCI |
语种 | 英语 |
WOS记录号 | WOS:000309785000019 |
源URL | [http://ir.ia.ac.cn/handle/173211/2878] ![]() |
专题 | 自动化研究所_模式识别国家重点实验室_多媒体计算与图形学团队 |
作者单位 | 1.Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia 2.Chinese Acad Sci, Inst Automat, Beijing 100864, Peoples R China 3.Univ Rochester, Dept Comp Sci, Rochester, NY 14627 USA |
推荐引用方式 GB/T 7714 | Zhu, Xiaofeng,Huang, Zi,Yang, Yang,et al. Self-taught dimensionality reduction on the high-dimensional small-sized data[J]. PATTERN RECOGNITION,2013,46(1):215-229. |
APA | Zhu, Xiaofeng,Huang, Zi,Yang, Yang,Shen, Heng Tao,Xu, Changsheng,&Luo, Jiebo.(2013).Self-taught dimensionality reduction on the high-dimensional small-sized data.PATTERN RECOGNITION,46(1),215-229. |
MLA | Zhu, Xiaofeng,et al."Self-taught dimensionality reduction on the high-dimensional small-sized data".PATTERN RECOGNITION 46.1(2013):215-229. |
入库方式: OAI收割
来源:自动化研究所
浏览0
下载0
收藏0
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。