Simultaneous multiple low-dimensional subspace dimensionality reduction and classification Online publication date: Mon, 20-May-2019
by Lijun J. Dou; Rui Yan; Qiaolin L. Ye
International Journal of Computational Science and Engineering (IJCSE), Vol. 19, No. 1, 2019
Abstract: Fisher linear discriminant (FLD) for supervised learning has recently emerged as a computationally powerful tool for extracting features for a variety of pattern classification problems. However, it works poorly with multimodal data. Local Fisher linear discriminant (LFLD) is proposed to reduce the dimensionality of multimodal data. Through experiments tried out on the multimodal but binary datasets created from several multi-class datasets, it has been shown to be better than FLD in favour of performance. However, LFLD has a serious limitation, which is that it is limited to the use on small-scale datasets. In order to address the above disadvantages, in this paper, we develop a multiple low-dimensionality dimensionality reduction technique (MSDR) of performing the dimensionality reduction (DR) of input data. In contrast to FLD and LFLD finding an optimal low-dimensional subspace, the new algorithm attempts to seek multiple optimal low-dimensional subspaces that best make the data sharing the same labels more compact. Inheriting the advantages of NC, MSDR reduces the dimensionality of data and directly performs classification tasks without needing to train the model. Experiments of comparing MSDR with the existing traditional approaches tried out on UCI, show the effectiveness and efficiency of MSDR.
Online publication date: Mon, 20-May-2019
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Computational Science and Engineering (IJCSE):
Login with your Inderscience username and password:
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.
If you still need assistance, please email email@example.com